Skip to main content
🧪

Mixtral

Mistral's sparse mixture-of-experts LLM with GPT-4 level performance

Code & Development
Mixtral logo

Mixtral

Mistral's sparse mixture-of-experts LLM with GPT-4 level performance

Mixtral 8x7B is a sparse mixture-of-experts language model developed by Mistral AI that achieves performance comparable to GPT-4 while using only a fraction of the compute at inference time. The model contains 8 expert networks, with only 2 active for any given token, giving it 46.7B parameters in total but behaving like a 12.9B parameter model for inference. Mixtral outperforms Llama 2 70B and matches or exceeds GPT-3.5 on most benchmarks while running faster. The model is openly released under the Apache 2.0 license, making it free for commercial use and self-hosting. Mixtral set a new benchmark for open-weight model efficiency when released in December 2023.

Key Features

  • Mixture-of-experts architecture
  • GPT-4 level performance
  • Open Apache 2.0 license
  • Fast inference
  • Self-hostable
  • Strong code generation
#llm#open-source#mistral#mixture-of-experts#self-hosted

Get Started

Visit Mixtral
🟢
Free
Completely free to use

Quick Info

Category
Code & Development
Pricing
Free

More Code & Development Tools