Skip to content
llm-speed

llm-speed

The benchmark suite for local + hosted LLM inference.Run it. See your tok/s. Compare your rig.

$ pipx install llm-speed && llm-speed bench

Runs in 60 seconds. Auto-detects your hardware. Open source. · Pre-launch: pipx install https://api-dev.llm-speed.com/dist/llm_speed-0.0.1-py3-none-any.whl

Latest submissions

13 runs
  • M3 Ultra (60-core GPU) + 96GB unified
    suite-v1
  • M3 Ultra (60-core GPU) + 96GB unified
    suite-v1
  • M3 Pro (18-core GPU) + 36GB unified
    suite-v1
  • M3 Pro (18-core GPU) + 36GB unified
    suite-v1
  • M3 Ultra (60-core GPU) + 96GB unified
    suite-v1
  • M3 Ultra (60-core GPU) + 96GB unified
    suite-v1
  • M3 Ultra (60-core GPU) + 96GB unified
    suite-v1
  • M3 Pro (18-core GPU) + 36GB unified
    suite-v1
  • M3 Ultra (60-core GPU) + 96GB unified
    suite-v1
  • M3 Ultra (60-core GPU) + 96GB unified
    suite-v1
  • M3 Pro (18-core GPU) + 36GB unified
    suite-v1
  • M3 Pro (18-core GPU) + 36GB unified
    suite-v1
  • M3 Pro (18-core GPU) + 36GB unified
    suite-v1