Speed & Latency

Output speed (tokens per second) and latency (time to first token) benchmarks across models and providers.

Filters:
Provider:
Creators:

Speed vs Intelligence

EUGlobal
Drag to zoom ยท Click to pin
55 models
#
Model
Creator
Output Speed (t/s)โ†“
Latency TTFT (s)
Total Response (s)
Intelligence
Price $/1M
1
Gemini 2.5 Flash-LiteEU
Google250.4 t/s19.12s-
17.6
$0.17
2
Gemini 2.5 FlashREU
Google193.7 t/s12.75s-
27
$0.85
3
GPT-4.1 NanoEU
OpenAI181 t/s0.48s-
13
$0.17
4
Nova 2 LiteREU
Amazon174.7 t/s13.30s-
34.5
$0.85
5
Gemini 3 FlashRGlobal
Google170 t/s5.43s-
46.4
$1.13
6
GPT-5.4 NanoRGlobal
OpenAI162.2 t/s2.66s-
44
$0.46
7
GPT-5.4 MiniRGlobal
OpenAI161 t/s4.37s-
48.9
$1.69
8
GPT-5 NanoREU
OpenAI159.6 t/s62.37s-
26.8
$0.14
9
Mistral Small 4ROpenGlobal
Mistral155.2 t/s0.53s-
27.8
$0.26
10
Nemotron 3 SuperROpenGlobal
NVIDIA152.5 t/s0.77s-
36
$0.41
11
o3-miniREU
OpenAI148.6 t/s19.20s-
25.2
$1.93
12
Grok 4.20RGlobal
xAI140 t/s12.56s-
49.3
$3.00
13
GPT-4oEU
OpenAI136 t/s0.49s-
17.3
$4.38
14
o4-miniREU
OpenAI130.1 t/s13.44s-
33.1
$1.93
15
Gemini 3 ProRGlobal
Google125.8 t/s27.92s-
48.4
$4.50
16
Gemini 3.1 ProRGlobal
Google123.3 t/s19.86s-
57.2
$4.50
17
Grok 4.1 FastRGlobal
xAI116.5 t/s9.48s-
38.6
$0.28
18
Llama 4 MaverickOpenGlobal
Meta113 t/s0.61s-
18.4
$0.47
19
Gemini 2.5 ProREU
Google110.3 t/s25.80s-
34.6
$3.44
20
o1REU
OpenAI108.5 t/s19.13s-
30.8
$26.25
21
Claude Haiku 4.5REU
Anthropic106.6 t/s9.60s-
37.1
$2.00
22
GPT-5.1REU
OpenAI105.6 t/s13.78s-
47.7
$3.44
23
Llama 3.3 70BOpenGlobal
Meta96.1 t/s0.55s-
14.5
$0.61
24
GPT-4.1EU
OpenAI86.3 t/s0.70s-
26.3
$3.50
25
GPT-4.1 MiniEU
OpenAI85 t/s0.59s-
22.9
$0.70
26
GPT-5REU
OpenAI81.3 t/s64.94s-
44.6
$3.44
27
o3REU
OpenAI79.8 t/s6.27s-
38.4
$3.50
28
GPT-5.4RGlobal
OpenAI76.5 t/s169.45s-
56.8
$5.63
29
GPT-5 MiniREU
OpenAI76.3 t/s69.23s-
41.2
$0.69
30
GLM-5RGlobal
Z AI71 t/s0.77s-
49.8
$1.55
31
GPT-5.3 CodexRGlobal
OpenAI65.1 t/s53.13s-
53.6
$4.81
32
GPT-5.2RGlobal
OpenAI60.8 t/s48.31s-
51.3
$4.81
33
MiMo V2 ProROpenGlobal
Xiaomi58.2 t/s2.59s-
49.2
$1.50
34
GPT-4o MiniEU
OpenAI58.1 t/s0.55s-
12.6
$0.26
35
MiniMax M2.5RGlobal
MiniMax57 t/s2.46s-
41.9
$0.53
36
Grok 4RGlobal
xAI54.2 t/s6.77s-
41.5
$6.00
37
Mistral Large 3OpenGlobal
Mistral53.9 t/s0.63s-
22.8
$0.75
38
Qwen3.6 PlusRGlobal
Alibaba52.1 t/s1.54s-
50
$1.13
39
Qwen3.5 397BROpenGlobal
Alibaba52 t/s1.52s-
45
$1.35
40
Claude Sonnet 4.6REU
Anthropic51.2 t/s32.92s-
51.7
$6.00
41
Claude Opus 4.5REU
Anthropic50.9 t/s12.65s-
49.7
$10.00
42
GLM-5.1RGlobal
Z AI45 t/s0.93s-
51.4
$2.15
43
Claude Opus 4.6REU
Anthropic42.3 t/s10.52s-
53
$10.00
44
Claude Sonnet 4REU
Anthropic41.3 t/s8.79s-
38.7
$6.00
45
Claude Sonnet 4.5REU
Anthropic39.3 t/s8.51s-
43
$6.00
46
MiniMax M2.7RGlobal
MiniMax37.7 t/s2.31s-
49.6
$0.53
47
Kimi K2.5RGlobal
Kimi32.4 t/s3.99s-
46.8
$1.20
48
DeepSeek V3.2 (Non-reasoning)OpenGlobal
DeepSeek31.9 t/s1.43s-
32.1
$0.32
49
DeepSeek V3.2ROpenGlobal
DeepSeek31.4 t/s1.38s-
41.7
$0.32
50
Claude 3.7 SonnetREU
Anthropic---
34.7
$6.00
51
Muse SparkRGlobal
Meta---
52.1
$0.00
52
SonarGlobal
Perplexity----$1.00
53
Sonar ProGlobal
Perplexity----$12.00
54
Sonar Reasoning ProRGlobal
Perplexity----$6.50
55
Sonar Deep ResearchRGlobal
Perplexity----$6.50