Solid LLaMA 3 update, reliable for coding, chat, and Q&A tasks
100K+
Llama 3.2 introduced lightweight 1B and 3B models at bfloat16 (BF16) precision, later adding quantized versions. The quantized models are significantly faster, with a much lower memory footprint and reduced power consumption, while maintaining nearly the same accuracy as their BF16 counterparts.
Llama 3.2 instruct models are designed for:
| Attribute | Details |
|---|---|
| Provider | Meta |
| Architecture | Llama |
| Cutoff date | December 2023 |
| Languages | English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai |
| Tool calling | ✅ |
| Input modalities | Text |
| Output modalities | Text, Code |
| License | Llama 3.2 Community License |
| Model variant | Parameters | Quantization | Context window | VRAM¹ | Size |
|---|---|---|---|---|---|
ai/llama3.2:latestai/llama3.2:3B-Q4_K_M | 3B | IQ2_XXS/Q4_K_M | 131K tokens | 2.77 GiB | 1.87 GB |
ai/llama3.2:1B-Q4_0 | 1B | Q4_0 | 131K tokens | 1.35 GiB | 727.75 MB |
ai/llama3.2:1B-Q8_0 | 1B | Q8_0 | 131K tokens | 1.87 GiB | 1.22 GB |
ai/llama3.2:1B-F16 | 1B | F16 | 131K tokens | 2.95 GiB | 2.30 GB |
ai/llama3.2:3B-Q4_0 | 3B | Q4_0 | 131K tokens | 2.68 GiB | 1.78 GB |
ai/llama3.2:3B-Q4_K_M | 3B | IQ2_XXS/Q4_K_M | 131K tokens | 2.77 GiB | 1.87 GB |
ai/llama3.2:3B-F16 | 3B | F16 | 131K tokens | 6.89 GiB | 5.98 GB |
¹: VRAM estimated based on model characteristics.
latest→3B-Q4_K_M
First, pull the model:
docker model pull ai/llama3.2
Then run the model:
docker model run ai/llama3.2
For more information on Docker Model Runner, explore the documentation.
| Capability | Benchmark | Llama 3.2 1B |
|---|---|---|
| General | MMLU | 49.3 |
| Re-writing | Open-rewrite eval | 41.6 |
| Summarization | TLDR9+ (test) | 16.8 |
| Instruct. following | IFEval | 59.5 |
| Math | GSM8K (CoT) | 44.4 |
| MATH (CoT) | 30.6 | |
| Reasoning | ARC-C | 59.4 |
| GPQA | 27.2 | |
| Hellaswag | 41.2 | |
| Tool Use | BFCL V2 | 25.7 |
| Nexus | 13.5 | |
| Long Context | InfiniteBench/En.QA | 20.3 |
| InfiniteBench/En.MC | 38.0 | |
| NIH/Multi-needle | 75.0 | |
| Multilingual | MGSM (CoT) | 24.5 |
Content type
Model
Digest
sha256:da80a8418…
Size
1.8 GB
Last updated
9 months ago
docker model pull ai/llama3.2:3B-Q4_0Pulls:
11,912
Last week