<- Back to leaderboard
Coding
Model
Score
95% Confidence
1st
o1-mini
1245
+32 / -31
2nd
GPT-4o (November 2024)
1168
+36 / -35
3rd
o1 Preview
1165
+30 / -29
4
Gemini Pro Flash 2
1140
+31 / -31
5
Claude 3.5 Sonnet (June 2024)
1097
+22 / -24
6
GPT-4o (August 2024)
1069
+23 / -27
7
GPT-4o (May 2024)
1058
+24 / -24
8
Mistral Large 2
1052
+25 / -25
9
GPT-4 Turbo Preview
1052
+22 / -21
10
Llama 3.1 405B Instruct
1046
+23 / -24
11
Deepseek V3
1025
+31 / -31
12
GPT-4 (November 2024)
1016
+35 / -33
13
Gemini 1.5 Pro (August 27, 2024)
1015
+27 / -28
14
Gemini 1.5 Pro (May 2024)
1011
+26 / -26
15
Llama 3.2 90B Vision Instruct
1006
+28 / -32
16
Claude 3 Opus
979
+23 / -23
17
Gemini 1.5 Flash
967
+25 / -25
18
Gemini 1.5 Pro (April 2024)
927
+26 / -29
19
Claude 3 Sonnet
903
+27 / -30
20
Llama 3 70B Instruct
896
+26 / -26
21
Mistral Large
836
+28 / -30
22
Gemini 1.0 Pro
710
+35 / -32
23
CodeLlama 34B Instruct
621
+35 / -39