Statistical analysis of how 300 AI model composite scores are distributed. Explore the mean, median, percentiles, and tier breakdowns to understand the AI model landscape.
Summary statistics across all 300 scored models.
Mean Score
68.2
+/- 14.6 stddev
Median Score
70.8
Score Range
32—94
95th Percentile
86.9
Above Median
150
of 300 models
Number of models in each 10-point score bucket.
Models grouped by performance tier with summary statistics.
| Tier | Range | Count | % of Total |
|---|---|---|---|
| Elite | 90–100 | 11 | 3.7% |
| Strong | 70–89 | 141 | 47.0% |
| Average | 50–69 | 102 | 34.0% |
| Below Average | 30–49 | 37 | 12.3% |
| Weak | 0–29 | 0 | 0.0% |
Score thresholds at key percentile levels.
| Percentile | Score | Position |
|---|---|---|
| P5 | 39.0 | 3294 |
| P10 | 44.7 | 3294 |
| P25 | 59.3 | 3294 |
| P50 | 70.8 | 3294 |
| P75 | 80.3 | 3294 |
| P90 | 85.0 | 3294 |
| P95 | 86.9 | 3294 |
Providers with 3+ models, ranked by average composite score.
| Provider | Models | Avg Score |
|---|---|---|
1Xiaomi | 3 | 84.2 |
2ByteDance | 5 | 80.5 |
3xAI | 10 | 78.8 |
4Anthropic | 13 | 77.3 |
5OpenAI | 60 | 72.5 |
6MiniMax | 8 | 72.3 |
7Google | 23 | 71.9 |
8Moonshot AI | 4 | 71.5 |
9DeepSeek | 11 | 71.5 |
10NVIDIA | 11 | 70.6 |
How models are distributed across the top 20%, middle 60%, and bottom 20% of scores.
How scores are computed and what the distribution reveals.
Each model receives a composite score from 0 to 100, calculated as a weighted combination of six signals: capabilities (25%), pricing tier (25%), context window (15%), recency (15%), output capacity (10%), and versatility (10%). The score is designed to capture overall model quality and value in a single number.
The score distribution reveals the competitive landscape of AI models. A tight cluster near the median suggests many similarly capable models, while a wide spread indicates clear differentiation between tiers. The shape of the distribution, its skew, and the gap between mean and median all provide insight into whether the market is top-heavy, bottom-heavy, or evenly distributed.
Continue exploring AI model data with benchmarks, capabilities, and the full leaderboard.
The score distribution shows how all 290+ tracked AI models are spread across the 0-100 SignalScore scale. Most models cluster in the 40-70 range, with a small elite group scoring above 80 and budget/older models falling below 30.
SignalScore is a composite metric combining six weighted factors: capability breadth (25%), pricing tier (25%), context window (15%), recency (15%), output capacity (10%), and versatility (10%). Each factor is normalized to a 0-100 scale before weighting.
Models scoring above the 75th percentile (typically 65+ SignalScore) are considered strong performers. The top 10% of models score above 78, while the median score across all tracked models sits around 52-55.