o1 by OpenAI demonstrates strong general intelligence, outstanding mathematical reasoning. View detailed benchmark data including scores across coding, math, reasoning, speed, and cost metrics.
General Benchmarks
Coding Benchmarks
Reasoning Benchmarks
Speed Benchmarks
Cost Benchmarks
Context Benchmarks
o1 — Benchmark Scores Overview
Scores normalized to percentage scale for visual comparison. ELO scores mapped to 0-100 range (1100-1500).
Compare o1 With
o1 — Frequently Asked Questions
How intelligent is o1?
o1 scores 1360 on the Chatbot Arena ELO rating, making it a high-performing AI model. This score is based on blind head-to-head human preference voting.
How much does o1 cost?
o1 costs $15.0 per 1M input tokens and $60.0 per 1M output tokens. This places it in the premium pricing tier.
How fast is o1?
o1 generates output at 35 tokens per second, which is slower, prioritizing quality over speed compared to other models. The time to first token is 1000 ms.
How good is o1 at coding?
o1 achieves 48.9% on SWE-bench Verified, demonstrating moderate real-world software engineering capability. This benchmark tests the model's ability to resolve actual GitHub issues.
How good is o1 at math and reasoning?
o1 scores 94.8% on the MATH benchmark (competition-level mathematics). It also achieves 77.3% on GPQA Diamond, a graduate-level science reasoning benchmark.
What is the context window of o1?
o1 has a context window of 200K tokens. This determines how much text, conversation history, and code the model can process in a single request.
Who created o1?
o1 was created by OpenAI. It is classified as a flagship model in the AI Value Index.
Is o1 open source?
No, o1 is a proprietary model. It is available through OpenAI's API and compatible providers.