See how providers stack up.
Real benchmark runs — latency, throughput, cost, and stability. Every report is reproducible: click any card to view full results or clone the run with your own keys.
Groq vs OpenAI vs Anthropic — Speed Shootout
Three cloud providers head-to-head on a short generation task. Groq dominates on raw throughput; Anthropic leads on coherence.
Local RAIS vs Groq — On-device vs Cloud
A self-hosted RAIS endpoint against Groq's cloud inference. Local wins on latency when on the same machine; Groq wins on throughput.
Groq vs Ollama — Cloud vs Edge Inference
Groq cloud vs local Ollama (Llama 3.2 3B) on a reasoning task. Cloud wins on speed; Ollama wins on cost and privacy.
All 3 Cloud Providers — Long Generation
Extended output task across Groq, OpenAI, and Anthropic. Cost-efficiency ratios become visible at higher token counts.
Submit your benchmark
Run a benchmark on any RAIS-compatible endpoint, generate a share link, and open a PR to add it here. We feature benchmarks that are reproducible, honest, and interesting.