OpenAI

o4 Mini — Benchmark Scores, Pricing & Performance Analysis

Chatbot Arena ELO
1350
Output Speed
120 tok/s
Input Cost
$1.1/1M
Output Cost
$4.4/1M
Context Window
200K

o4 Mini by OpenAI demonstrates strong general intelligence, solid coding performance, outstanding mathematical reasoning. View detailed benchmark data including scores across coding, math, reasoning, speed, and cost metrics.

o4 Mini — Benchmark Scores Overview

Scores normalized to percentage scale for visual comparison. ELO scores mapped to 0-100 range (1100-1500).

o4 Mini — Frequently Asked Questions

How intelligent is o4 Mini?

o4 Mini scores 1350 on the Chatbot Arena ELO rating, making it a high-performing AI model. This score is based on blind head-to-head human preference voting.

How much does o4 Mini cost?

o4 Mini costs $1.1 per 1M input tokens and $4.4 per 1M output tokens. This is mid-range pricing for its capability level.

How fast is o4 Mini?

o4 Mini generates output at 120 tokens per second, which is moderate compared to other models. The time to first token is 300 ms.

How good is o4 Mini at coding?

o4 Mini achieves 68.1% on SWE-bench Verified, demonstrating strong real-world software engineering capability. This benchmark tests the model's ability to resolve actual GitHub issues.

How good is o4 Mini at math and reasoning?

o4 Mini scores 93.0% on the MATH benchmark (competition-level mathematics). It also achieves 81.4% on GPQA Diamond, a graduate-level science reasoning benchmark.

What is the context window of o4 Mini?

o4 Mini has a context window of 200K tokens. This determines how much text, conversation history, and code the model can process in a single request.

Who created o4 Mini?

o4 Mini was created by OpenAI. It is classified as a mid model in the AI Value Index.

Is o4 Mini open source?

No, o4 Mini is a proprietary model. It is available through OpenAI's API and compatible providers.