Skip to content

Latest commit

 

History

History
120 lines (120 loc) · 18 KB

bench.md

File metadata and controls

120 lines (120 loc) · 18 KB
model code crm docs integrate marketing reason final 🏆 Cost Speed
GPT o1 pro (manual) ☁️ 100 100 97 100 95 87 97 0.20 € 1.00 rps
GPT o1-preview v1/2024-09-12 ☁️ 95 92 94 95 88 87 92 52.32 € 0.08 rps
GPT o1 v1/2024-12-17 ☁️ 100 95 94 91 82 83 91 30.63 € 0.17 rps
GPT o1-mini v1/2024-09-12 ☁️ 93 96 94 83 82 87 89 8.15 € 0.16 rps
GPT-4o v3/2024-11-20 ☁️ 86 97 94 95 88 72 89 0.63 € 1.14 rps
GPT-4o v1/2024-05-13 ☁️ 90 96 100 92 78 74 88 1.21 € 1.44 rps
Google Gemini 1.5 Pro v2 ☁️ 86 97 94 99 78 74 88 1.00 € 1.18 rps
X-AI Grok 2 v2/1212 ⚠️ 66 95 97 97 88 78 87 0.58 € 0.99 rps
GPT-4 Turbo v5/2024-04-09 ☁️ 86 99 98 96 88 43 85 2.45 € 0.84 rps
Google Gemini 2.0 Flash Exp ☁️ 63 96 100 100 82 62 84 0.03 € 0.85 rps
Google Gemini Exp 1121 ☁️ 70 97 97 95 72 72 84 0.89 € 0.49 rps
GPT-4o v2/2024-08-06 ☁️ 90 84 97 86 82 59 83 0.63 € 1.49 rps
Google Gemini 1.5 Pro 0801 ☁️ 84 92 79 100 70 74 83 0.90 € 0.83 rps
Qwen 2.5 72B Instruct ⚠️ 79 92 94 97 71 59 82 0.10 € 0.66 rps
Llama 3.1 405B Hermes 3🦙 68 93 89 98 88 53 81 0.54 € 0.49 rps
Claude 3.5 Sonnet v2 ☁️ 82 97 93 84 71 57 81 0.95 € 0.09 rps
GPT-4 v1/0314 ☁️ 90 88 98 73 88 45 80 7.04 € 1.31 rps
X-AI Grok 2 v1/1012 ⚠️ 63 93 87 90 88 58 80 1.03 € 0.31 rps
GPT-4 v2/0613 ☁️ 90 83 95 73 88 45 79 7.04 € 2.16 rps
DeepSeek v3 671B ⚠️ 62 95 97 85 75 55 78 0.03 € 0.49 rps
GPT-4o Mini ☁️ 63 87 80 73 100 65 78 0.04 € 1.46 rps
Claude 3.5 Sonnet v1 ☁️ 72 83 89 87 80 58 78 0.94 € 0.09 rps
Claude 3 Opus ☁️ 69 88 100 74 76 58 77 4.69 € 0.41 rps
Meta Llama3.1 405B Instruct🦙 81 93 92 75 75 48 77 2.39 € 1.16 rps
GPT-4 Turbo v4/0125-preview ☁️ 66 97 100 83 75 43 77 2.45 € 0.84 rps
Google LearnLM 1.5 Pro Experimental ⚠️ 48 97 85 96 64 72 77 0.31 € 0.83 rps
GPT-4 Turbo v3/1106-preview ☁️ 66 75 98 73 88 60 76 2.46 € 0.68 rps
Google Gemini Exp 1206 ☁️ 52 100 85 77 75 69 76 0.88 € 0.16 rps
Qwen 2.5 32B Coder Instruct ⚠️ 43 94 98 98 76 46 76 0.05 € 0.82 rps
DeepSeek v2.5 236B ⚠️ 57 80 91 80 88 57 75 0.03 € 0.42 rps
Meta Llama 3.1 70B Instruct f16🦙 74 89 90 75 75 48 75 1.79 € 0.90 rps
Google Gemini 1.5 Flash v2 ☁️ 64 96 89 76 81 44 75 0.06 € 2.01 rps
Google Gemini 1.5 Pro 0409 ☁️ 68 97 96 80 75 26 74 0.95 € 0.59 rps
Meta Llama 3 70B Instruct🦙 81 83 84 67 81 45 73 0.06 € 0.85 rps
GPT-3.5 v2/0613 ☁️ 68 81 73 87 81 50 73 0.34 € 1.46 rps
Amazon Nova Lite ⚠️ 67 78 74 94 62 62 73 0.02 € 2.19 rps
Mistral Large 123B v2/2407 ☁️ 68 79 68 75 75 70 72 0.57 € 1.02 rps
Google Gemini Flash 1.5 8B ☁️ 70 93 78 67 76 48 72 0.01 € 1.19 rps
Google Gemini 1.5 Pro 0514 ☁️ 73 96 79 100 25 60 72 1.07 € 0.92 rps
Google Gemini 1.5 Flash 0514 ☁️ 32 97 100 76 72 52 72 0.06 € 1.77 rps
Google Gemini 1.0 Pro ☁️ 66 86 83 79 88 28 71 0.37 € 1.36 rps
Meta Llama 3.2 90B Vision🦙 74 84 87 77 71 32 71 0.23 € 1.10 rps
GPT-3.5 v3/1106 ☁️ 68 70 71 81 78 58 71 0.24 € 2.33 rps
Claude 3.5 Haiku ☁️ 52 80 72 75 75 68 70 0.32 € 1.24 rps
Meta Llama 3.3 70B Instruct🦙 74 78 74 77 71 46 70 0.10 € 0.71 rps
GPT-3.5 v4/0125 ☁️ 63 87 71 77 78 43 70 0.12 € 1.43 rps
Cohere Command R+ ☁️ 63 80 76 72 70 58 70 0.83 € 1.90 rps
Mistral Large 123B v3/2411 ☁️ 68 75 64 76 82 51 70 0.56 € 0.66 rps
Qwen1.5 32B Chat f16 ⚠️ 70 90 82 76 78 20 69 0.97 € 1.66 rps
Gemma 2 27B IT ⚠️ 61 72 87 74 89 32 69 0.07 € 0.90 rps
Mistral 7B OpenChat-3.5 v3 0106 f16 ✅ 68 87 67 74 88 25 68 0.32 € 3.39 rps
Meta Llama 3 8B Instruct f16🦙 79 62 68 70 80 41 67 0.32 € 3.33 rps
Gemma 7B OpenChat-3.5 v3 0106 f16 ✅ 63 67 84 58 81 46 67 0.21 € 5.09 rps
GPT-3.5-instruct 0914 ☁️ 47 92 69 69 88 33 66 0.35 € 2.15 rps
Amazon Nova Pro ⚠️ 64 78 82 79 52 41 66 0.22 € 1.34 rps
GPT-3.5 v1/0301 ☁️ 55 82 69 81 82 26 66 0.35 € 4.12 rps
Llama 3 8B OpenChat-3.6 20240522 f16 ✅ 76 51 76 65 88 38 66 0.28 € 3.79 rps
Mistral 7B OpenChat-3.5 v1 f16 ✅ 58 72 72 71 88 33 66 0.49 € 2.20 rps
Mistral 7B OpenChat-3.5 v2 1210 f16 ✅ 63 73 72 66 88 30 65 0.32 € 3.40 rps
Qwen 2.5 7B Instruct ⚠️ 48 77 80 68 69 47 65 0.07 € 1.25 rps
Starling 7B-alpha f16 ⚠️ 58 66 67 73 88 34 64 0.58 € 1.85 rps
Mistral Nemo 12B v1/2407 ☁️ 54 58 51 99 75 49 64 0.03 € 1.22 rps
Meta Llama 3.2 11B Vision🦙 70 71 65 70 71 36 64 0.04 € 1.49 rps
Llama 3 8B Hermes 2 Theta🦙 61 73 74 74 85 16 64 0.05 € 0.55 rps
Claude 3 Haiku ☁️ 64 69 64 75 75 35 64 0.08 € 0.52 rps
Yi 1.5 34B Chat f16 ⚠️ 47 78 70 74 86 26 64 1.18 € 1.37 rps
Liquid: LFM 40B MoE ⚠️ 72 69 65 63 82 24 63 0.00 € 1.45 rps
Meta Llama 3.1 8B Instruct f16🦙 57 74 62 74 74 32 62 0.45 € 2.41 rps
Qwen2 7B Instruct f32 ⚠️ 50 81 81 61 66 31 62 0.46 € 2.36 rps
Claude 3 Sonnet ☁️ 72 41 74 74 78 28 61 0.95 € 0.85 rps
Mistral Small v3/2409 ☁️ 43 75 71 74 75 26 61 0.06 € 0.81 rps
Mistral Pixtral 12B ✅ 53 69 73 63 64 40 60 0.03 € 0.83 rps
Mixtral 8x22B API (Instruct) ☁️ 53 62 62 97 75 7 59 0.17 € 3.12 rps
Anthropic Claude Instant v1.2 ☁️ 58 75 65 77 65 16 59 2.10 € 1.49 rps
Codestral Mamba 7B v1 ✅ 53 66 51 97 71 17 59 0.30 € 2.82 rps
Inflection 3 Productivity ⚠️ 46 59 39 70 79 61 59 0.92 € 0.17 rps
Anthropic Claude v2.0 ☁️ 63 52 55 67 84 34 59 2.19 € 0.40 rps
Cohere Command R ☁️ 45 66 57 74 84 27 59 0.13 € 2.50 rps
Amazon Nova Micro ⚠️ 58 68 64 71 59 31 59 0.01 € 2.41 rps
Qwen1.5 7B Chat f16 ⚠️ 56 81 60 56 60 36 58 0.29 € 3.76 rps
Mistral Large v1/2402 ☁️ 37 49 70 83 84 25 58 0.58 € 2.11 rps
Microsoft WizardLM 2 8x22B ⚠️ 48 76 79 59 62 22 58 0.13 € 0.70 rps
Qwen1.5 14B Chat f16 ⚠️ 50 58 51 72 84 22 56 0.36 € 3.03 rps
MistralAI Ministral 8B ✅ 56 55 41 82 68 30 55 0.02 € 1.02 rps
Anthropic Claude v2.1 ☁️ 29 58 59 78 75 32 55 2.25 € 0.35 rps
Mistral 7B OpenOrca f16 ☁️ 54 57 76 36 78 27 55 0.41 € 2.65 rps
MistralAI Ministral 3B ✅ 50 48 39 89 60 41 54 0.01 € 1.02 rps
Llama2 13B Vicuna-1.5 f16🦙 50 37 55 62 82 37 54 0.99 € 1.09 rps
Mistral 7B Instruct v0.1 f16 ☁️ 34 71 69 63 62 23 54 0.75 € 1.43 rps
Meta Llama 3.2 3B🦙 52 71 66 71 44 14 53 0.01 € 1.25 rps
Google Recurrent Gemma 9B IT f16 ⚠️ 58 27 71 64 56 23 50 0.89 € 1.21 rps
Codestral 22B v1 ✅ 38 47 44 84 66 13 49 0.06 € 4.03 rps
Qwen: QwQ 32B Preview ⚠️ 43 32 74 52 48 40 48 0.05 € 0.63 rps
Llama2 13B Hermes f16🦙 50 24 37 75 60 42 48 1.00 € 1.07 rps
IBM Granite 34B Code Instruct f16 ☁️ 63 49 34 67 57 7 46 1.07 € 1.51 rps
Meta Llama 3.2 1B🦙 32 40 33 53 68 51 46 0.02 € 1.69 rps
Mistral Small v2/2402 ☁️ 33 42 45 88 56 8 46 0.06 € 3.21 rps
Mistral Small v1/2312 (Mixtral) ☁️ 10 67 63 65 56 8 45 0.06 € 2.21 rps
DBRX 132B Instruct ⚠️ 43 39 43 74 59 10 45 0.26 € 1.31 rps
NVIDIA Llama 3.1 Nemotron 70B Instruct🦙 68 54 25 72 28 21 45 0.09 € 0.53 rps
Mistral Medium v1/2312 ☁️ 41 43 44 59 62 12 44 0.81 € 0.35 rps
Microsoft WizardLM 2 7B ⚠️ 53 34 42 66 53 13 43 0.02 € 0.89 rps
Llama2 13B Puffin f16🦙 37 15 44 67 56 39 43 4.70 € 0.23 rps
Mistral Tiny v1/2312 (7B Instruct v0.2) ☁️ 22 47 59 53 62 8 42 0.05 € 2.39 rps
Gemma 2 9B IT ⚠️ 45 25 47 36 68 13 39 0.02 € 0.88 rps
Meta Llama2 13B chat f16🦙 22 38 17 65 75 6 37 0.75 € 1.44 rps
Mistral 7B Zephyr-β f16 ✅ 37 34 46 62 29 4 35 0.46 € 2.34 rps
Meta Llama2 7B chat f16🦙 22 33 20 62 50 18 34 0.56 € 1.93 rps
Mistral 7B Notus-v1 f16 ⚠️ 10 54 25 60 48 4 33 0.75 € 1.43 rps
Orca 2 13B f16 ⚠️ 18 22 32 29 67 20 31 0.95 € 1.14 rps
Mistral 7B Instruct v0.2 f16 ☁️ 11 30 54 25 58 8 31 0.96 € 1.12 rps
Mistral 7B v0.1 f16 ☁️ 0 9 48 63 52 12 31 0.87 € 1.23 rps
Google Gemma 2B IT f16 ⚠️ 33 28 16 47 15 20 27 0.30 € 3.54 rps
Microsoft Phi 3 Medium 4K Instruct 14B f16 ⚠️ 5 34 30 32 47 8 26 0.82 € 1.32 rps
Orca 2 7B f16 ⚠️ 22 0 26 26 52 4 22 0.78 € 1.38 rps
Google Gemma 7B IT f16 ⚠️ 0 0 0 6 62 0 11 0.99 € 1.08 rps
Meta Llama2 7B f16🦙 0 5 22 3 28 2 10 0.95 € 1.13 rps
Yi 1.5 9B Chat f16 ⚠️ 0 4 29 17 0 8 10 1.41 € 0.76 rps