8,500 multi-step grade-school math word problems. Tests basic arithmetic reasoning — adding, subtracting, multiplying with multi-step chains.
Why it matters: Useful baseline for math ability, but now saturated — top models exceed 95%. More useful for evaluating smaller or open-source models.
Top Model
96.8%
Llama 3.1 405B
Average Score
93.9%
Across 14 models
Models Tested
14
Metric: accuracy
Human Baseline
—
Range: 0%–100%
All models with a reported GSM8K score, ranked by highest accuracy.
GSM8K is a standardized evaluation that measures AI model performance on specific tasks. It provides comparable scores across different models, helping developers choose the right model for their needs.
Llama 3.1 405B currently holds the top score on the GSM8K benchmark. See our full rankings table above for the complete leaderboard with 14 models.
We update benchmark data from multiple sources including HuggingFace Open LLM Leaderboard and LMArena. Scores are refreshed regularly as new evaluations are published and new models are released.
No. While GSM8K is an important indicator, real-world performance depends on many factors including pricing, latency, context window, and specific task requirements. We recommend using our composite score which weighs multiple benchmarks and practical factors.