Efficient open model with top-tier performance and fast inference
50K+
A fast and powerful 7B parameter model excelling in reasoning, code, and math. Mistral 7B is a powerful 7.3B parameter language model that outperforms Llama 2 13B across a wide range of benchmarks, including reasoning, reading comprehension, and code generation. Despite its smaller size, it delivers performance comparable to much larger models, making it efficient and versatile.
Mistral 7B is designed to provide high-quality responses across a wide range of general-purpose NLP tasks while remaining efficient in resource usage. Also, this model is fine-tuned to follow instructions, allowing it to perform tasks and answer questions naturally. The base model doesn’t have this capability.
| Attribute | Details |
|---|---|
| Provider | Mistral AI |
| Architecture | Llama |
| Cutoff date | December 2023ⁱ |
| Languages | English (primarily) |
| Tool calling | ❌ |
| Input modalities | Text |
| Output modalities | Text |
| License | Apache 2.0 |
i: Estimated
| Model variant | Parameters | Quantization | Context window | VRAM¹ | Size |
|---|---|---|---|---|---|
ai/mistral:latestai/mistral:7B-Q4_K_M | 7B | IQ2_XXS/Q4_K_M | 33K tokens | 4.85 GiB | 4.07 GB |
ai/mistral:7B-Q4_0 | 7B | Q4_0 | 33K tokens | 4.61 GiB | 3.83 GB |
ai/mistral:7B-Q4_K_M | 7B | IQ2_XXS/Q4_K_M | 33K tokens | 4.85 GiB | 4.07 GB |
ai/mistral:7B-F16 | 7B | F16 | 33K tokens | 14.10 GiB | 13.50 GB |
¹: VRAM estimated based on model characteristics.
latest→7B-Q4_K_M
First, pull the model:
docker model pull ai/mistral
Then run the model:
docker model run ai/mistral
For more information on Docker Model Runner, explore the documentation.
[INST] and [/INST] tags. The first instruction must start with a beginning-of-sentence token, while any following instructions should not. The assistant's response will automatically end with an end-of-sentence token.| Capability | Benchmark | Mistral 7B |
|---|---|---|
| Natural Language Understanding | MMLU | 60.1% |
| HellaSwag | 81.3% | |
| WinoGrande | 75.3% | |
| PIQA | 83.0% | |
| Arc-e | 80.0% | |
| Arc-c | 55.5% | |
| Knowledge Retrieval | NQ | 28.8% |
| TriviaQA | 69.9% | |
| Code Generation & Debugging | HumanEval | 30.5% |
| MBPP | 47.5% | |
| Mathematical Reasoning | MATH | 13.1% |
| GSM8K | 52.1% |
Content type
Model
Digest
sha256:74890f8fb…
Size
3.8 GB
Last updated
about 1 year ago
docker model pull ai/mistral:7B-Q4_0Pulls:
712
Last week