<- Back to leaderboard
Coding
Model
Score
95% Confidence
1st
o1-mini
1247.00
+34 / -31
2nd
GPT-4o (November 2024)
1165.00
+34 / -31
3rd
o1-preview
1146.00
+28 / -25
4
Gemini 2.0 Flash Experimental (December 2024)
1130.00
+29 / -29
5
DeepSeek R1
1108.00
+33 / -32
6
o1 (December 2024)
1097.00
+33 / -32
7
Claude 3.5 Sonnet (June 2024)
1094.00
+21 / -22
8
GPT-4o (August 2024)
1054.00
+24 / -23
9
GPT-4o (May 2024)
1049.00
+23 / -23
10
GPT-4 Turbo Preview
1042.00
+22 / -22
11
Mistral Large 2
1041.00
+24 / -23
12
Llama 3.1 405B Instruct
1033.00
+27 / -22
13
Gemini 1.5 Pro (August 27, 2024)
1019.00
+26 / -25
14
Deepseek V3
1011.00
+28 / -28
15
GPT-4 (November 2024)
1011.00
+28 / -31
16
Gemini 1.5 Pro (May 2024)
1010.00
+25 / -25
17
Llama 3.2 90B Vision Instruct
998.00
+28 / -28
18
Claude 3 Opus
969.00
+23 / -23
19
Gemini 1.5 Flash
958.00
+24 / -28
20
Gemini 1.5 Pro (April 2024)
905.00
+30 / -29
21
Claude 3 Sonnet
891.00
+29 / -31
22
Llama 3 70B Instruct
885.00
+26 / -27
23
Mistral Large
825.00
+27 / -28
24
Gemini 1.0 Pro
699.00
+32 / -35
25
CodeLlama 34B Instruct
610.00
+38 / -42