⚡
Zephyr
HuggingFace's distilled LLM trained with direct preference optimization
Code & Development
Zephyr is a series of language models from HuggingFace trained using Direct Preference Optimization (DPO) and distilled supervised fine-tuning (dSFT) on AI feedback. Zephyr-7B-beta was a breakthrough model demonstrating that DPO training on AI-generated preference data could produce models outperforming much larger instruction-tuned models. HuggingFace releases Zephyr as a research artifact demonstrating alignment training techniques.
Key Features
- ✓DPO training
- ✓7B parameter
- ✓AI feedback alignment
- ✓HuggingFace native
- ✓Apache 2.0
- ✓Research focused
#llm#huggingface#dpo#alignment#open-source
Quick Info
- Category
- Code & Development
- Pricing
- Free
More Code & Development Tools
GitHub Copilot
Code & DevelopmentThe AI pair programmer trusted by millions of developers
Cursor
Code & DevelopmentThe code editor built around AI from the ground up
Tabnine
Code & DevelopmentPrivacy-first AI code completion
Codeium
Code & DevelopmentFree AI coding assistant with no usage limits