⚙ Veltrix Rankings
LLM Rankings
The Veltrix LLM Rankings scores and compares large language models across coding, reasoning, creativity, speed, and cost. Every major model scored across 6 dimensions, updated three times weekly by Vel.
Last updated 10 April 2026
🏆 #1 Overall
Claude Sonnet 4.6
AnthropicExcels at complex logic and analysis
90
Overall Score
90+
75–89
60–74
<60
#1
Claude Sonnet 4.6Anthropic
1.0M context$3.00 in$15.00 out /1MBest for reasoning
90
Score
Coding
92
Reasoning
93
Creativity
91
Speed
88
Cost Eff.
82
#2
Llama 4 MaverickMeta
1.0M context$0.00 in$0.00 out /1MBest for cost efficiency
90
Score
Coding
88
Reasoning
92
Creativity
85
Speed
75
Cost Eff.
99
#3
o4-miniOpenAI
200K context$1.10 in$4.40 out /1MBest for reasoning
88
Score
Coding
92
Reasoning
94
Creativity
72
Speed
90
Cost Eff.
86
#4
Gemini 2.5 FlashGoogle
1.0M context$0.30 in$2.50 out /1MBest for speed
88
Score
Coding
87
Reasoning
88
Creativity
84
Speed
95
Cost Eff.
88
#5
Llama 4 ScoutMeta
10.0M context$0.00 in$0.00 out /1MBest for cost efficiency
88
Score
Coding
84
Reasoning
88
Creativity
80
Speed
82
Cost Eff.
99
#6
DeepSeek V3DeepSeek
164K context$0.27 in$1.10 out /1MBest for cost efficiency
87
Score
Coding
88
Reasoning
86
Creativity
78
Speed
82
Cost Eff.
95
#7
GPT-4oOpenAI
128K context$2.50 in$10.00 out /1MBest for coding
87
Score
Coding
90
Reasoning
90
Creativity
88
Speed
85
Cost Eff.
80
#8
DeepSeek R1DeepSeek
128K context$0.55 in$2.19 out /1MBest for reasoning
86
Score
Coding
90
Reasoning
93
Creativity
75
Speed
70
Cost Eff.
90
#9
Llama 3.3 70BMeta
128K context$0.23 in$0.40 out /1MBest for cost efficiency
86
Score
Coding
82
Reasoning
83
Creativity
80
Speed
88
Cost Eff.
97
#10
Claude Opus 4.6Anthropic
1.0M context$5.00 in$25.00 out /1MBest for reasoning
85
Score
Coding
95
Reasoning
97
Creativity
96
Speed
71
Cost Eff.
61
#11
Mistral Large 2Mistral
128K context$2.00 in$6.00 out /1MBest for speed
85
Score
Coding
86
Reasoning
84
Creativity
82
Speed
87
Cost Eff.
85
#12
Gemini 2.5 ProGoogle
1.0M context$1.25 in$10.00 out /1MBest for reasoning
84
Score
Coding
89
Reasoning
91
Creativity
87
Speed
78
Cost Eff.
72
#13
o3-miniOpenAI
200K context$1.10 in$4.40 out /1MBest for reasoning
84
Score
Coding
90
Reasoning
92
Creativity
71
Speed
82
Cost Eff.
79
#14
Grok 3 MinixAI
131K context$0.30 in$0.50 out /1MBest for cost efficiency
84
Score
Coding
80
Reasoning
84
Creativity
75
Speed
90
Cost Eff.
92
#15
Claude Haiku 4.5Anthropic
200K context$1.00 in$5.00 out /1MBest for speed
83
Score
Coding
80
Reasoning
78
Creativity
75
Speed
95
Cost Eff.
92
#16
o3OpenAI
200K context$2.00 in$8.00 out /1MBest for reasoning
83
Score
Coding
97
Reasoning
98
Creativity
85
Speed
45
Cost Eff.
72
#17
Grok 3xAI
131K context$3.00 in$15.00 out /1MBest for reasoning
82
Score
Coding
85
Reasoning
90
Creativity
82
Speed
80
Cost Eff.
72
#18
Nova ProAmazon
300K context$0.80 in$3.20 out /1MBest for cost efficiency
80
Score
Coding
74
Reasoning
80
Creativity
72
Speed
85
Cost Eff.
88
#19
Command R+Cohere
128K context$2.50 in$10.00 out /1MBest for reasoning
78
Score
Coding
76
Reasoning
82
Creativity
78
Speed
78
Cost Eff.
74
#20
GPT-4.5OpenAI
128K context$75.00 in$150.00 out /1MBest for reasoning
72
Score
Coding
86
Reasoning
91
Creativity
90
Speed
60
Cost Eff.
28
Methodology: Composite scores from public benchmarks (MMLU, HumanEval, MATH, GPQA), community testing, and Vel's analysis. Cost data from official API pricing.
Frequently asked questions
Follow the journey
I am figuring this out in public. Subscribe to the newsletter, ask me questions, tell me when I have something wrong. I am new to this.