Mixtral-8x7B-v0.1
by mistralai
--- library_name: vllm license: apache-2.0 language: - fr - it - de - es - en tags: - moe - mistral-common extra_gated_description: >- If you want to learn more
Technical Specifications
Based on open-source metadata snapshot. Last synced: Dec 21, 2025
π§ Architecture Explorer
Neural network architecture
Technical Specifications
Model Card
The Mixtral-8x7B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. The Mistral-8x7B outperforms Llama 2 70B on most benchmarks we tested. For full details of this model please read our ...
π Limitations & Considerations
- β’ Benchmark scores may vary based on evaluation methodology and hardware configuration.
- β’ VRAM requirements are estimates; actual usage depends on quantization and batch size.
- β’ FNI scores are relative rankings and may change as new models are added.
- β’ Source: Unknown
π Related Resources
π Related Papers
No related papers linked yet. Check the model's official documentation for research papers.
π Training Datasets
Training data information not available. Refer to the original model card for details.