Benchmarks
Inspect the newest public uploads
Start with the newest results and narrow the list down from there with filters.
GPU & NPU Benchmarking
Menü
Community benchmark hub
Explore public GPU and NPU benchmark uploads, jump into filtered leaderboards and publish your own reproducible results with screenshots.
Jump in
Start with the newest uploads, jump straight into speed and efficiency views, filter for Apple hardware or browse the uploader leaderboard.
Benchmarks
Start with the newest results and narrow the list down from there with filters.
Views
Compare GPUs and NPUs by prompt throughput and see which local LLM hardware handles long prompts and large context windows the fastest.
Views
See which GPUs and NPUs deliver the highest token generation speed after the first token when streaming output and response speed matter most.
Views
Compare GPUs and NPUs by token generation per watt to find local LLM hardware that balances speed and lower power consumption.
Apple
Open the benchmark list with Apple already selected and compare the available results right away.
Leaderboard
See who has contributed the most results and jump from there into matching benchmark views.
Quick answers
Get the core answers about comparable hardware results, standardized tests, and the upcoming LocalLLMBench software in one place.
LocalLLMBench compares GPU and NPU performance for local LLMs using public benchmark results, filters, and measurable hardware context.
Fair hardware comparisons need reproducible measurements with the same benchmark, ideally a standardized test such as Llama-Bench on llama.cpp.
Yes. A simpler LocalLLMBench software is in development to make standardized benchmarks easier to run and prepare for upload.