OpenAI

GPT-4.1 — Benchmark Scores, Pricing & Performance Analysis

Chatbot Arena ELO
1340
Output Speed
70 tok/s
Input Cost
$2.0/1M
Output Cost
$8.0/1M
Context Window
1.0M

GPT-4.1 by OpenAI demonstrates strong general intelligence, solid coding performance. View detailed benchmark data including scores across coding, math, reasoning, speed, and cost metrics.

GPT-4.1 — Benchmark Scores Overview

Scores normalized to percentage scale for visual comparison. ELO scores mapped to 0-100 range (1100-1500).

GPT-4.1 — Frequently Asked Questions

How intelligent is GPT-4.1?

GPT-4.1 scores 1340 on the Chatbot Arena ELO rating, making it a mid-tier AI model. This score is based on blind head-to-head human preference voting.

How much does GPT-4.1 cost?

GPT-4.1 costs $2.0 per 1M input tokens and $8.0 per 1M output tokens. This is mid-range pricing for its capability level.

How fast is GPT-4.1?

GPT-4.1 generates output at 70 tokens per second, which is slower, prioritizing quality over speed compared to other models. The time to first token is 450 ms.

How good is GPT-4.1 at coding?

GPT-4.1 achieves 50.0% on SWE-bench Verified, demonstrating strong real-world software engineering capability. This benchmark tests the model's ability to resolve actual GitHub issues.

How good is GPT-4.1 at math and reasoning?

GPT-4.1 scores 83.0% on the MATH benchmark (competition-level mathematics). It also achieves 58.0% on GPQA Diamond, a graduate-level science reasoning benchmark.

What is the context window of GPT-4.1?

GPT-4.1 has a context window of 1.0M tokens. This determines how much text, conversation history, and code the model can process in a single request.

Who created GPT-4.1?

GPT-4.1 was created by OpenAI. It is classified as a mid model in the AI Value Index.

Is GPT-4.1 open source?

No, GPT-4.1 is a proprietary model. It is available through OpenAI's API and compatible providers.