Llama-3.1-8B-Instruct-4bit
2 workload results across 2 hardware configurations.
Fastest local config
130.2 decode tok/s
on M3 Ultra (60-core GPU) + 96GB unified via mlx — see full run
Local runs (2 runs)
Runs from contributors' own machines via MLX, llama.cpp, vLLM, exllamav2, or ollama. Signed on the submitter's hardware.
M3 Pro (18-core GPU) + 36GB unified
| Workload | Backend | Quant | decode tok/s | prefill tok/s | TTFT | Run |
|---|---|---|---|---|---|---|
| chat-short | mlx@0.31.3 | — | 29.20tok/s | 203.3tok/s | 669ms | r_h0-use1ypnb |
M3 Ultra (60-core GPU) + 96GB unified
| Workload | Backend | Quant | decode tok/s | prefill tok/s | TTFT | Run |
|---|---|---|---|---|---|---|
| chat-short | mlx@0.31.3 | — | 130.2tok/s | 400.3tok/s | 340ms | r_v2pbc0rq2l4 |