BE
← Leaderboard

o1-mini

Closed
OpenAI
Proprietary
text
o-seriesReleased 2y ago
Avg score
72.6
/ 100
Context
128k
Output limit
66k
Input price
$3.00 /M
Output price
$12.00 /M

Pricing verified 1y ago · source

Benchmarks

preference

Crowdsourced pairwise human preference rankings of LLM responses. Higher Elo means more frequently preferred by users.

math

American Invitational Mathematics Examination 2024 problems. Three-digit integer answers; very hard for non-reasoning models.

coding

HumanEval% pass@1

164 hand-written Python programming problems scored by passing unit tests. Saturated for frontier models.

vision

Massive Multi-discipline Multimodal Understanding; college-exam level questions with images across 30+ subjects.

Math reasoning over visual contexts (charts, figures, geometry).

long context

Long-context retrieval and reasoning suite. We report the 128k token effective-context score.

performance

Median sustained output speed in tokens per second on the model's first-party API for medium-length prompts. Higher is faster.

Median time from request to first output chunk in milliseconds on the model's first-party API for medium-length prompts. Lower is snappier; reasoning models are penalised here because they think before talking.

Providers

Only available from OpenAI — no third-party hosts found.

Compare with...