When Anthropic unveiled Claude 4 Opus on May 15, 2026, the AI research community took notice immediately. The numbers were not incremental improvements. They were categorical leaps. On GPQA Diamond, the benchmark widely regarded as the toughest test of graduate-level scientific reasoning, Claude 4 Opus scored 94.2%, a full 11 points above the previous best result from any publicly evaluated model. On HumanEval, the canonical Python code generation suite, it reached 91.3%. These are not just impressive figures; they represent a real shift in what AI systems can reliably accomplish.

The release marks the culmination of an 18-month development cycle that Anthropic describes as its most intensive to date. Unlike previous Claude generations, which iterated primarily on scale and fine-tuning approaches, Claude 4 Opus was trained with a thoroughly redesigned architecture featuring improved long-range attention, a reworked chain-of-thought scaffolding system, and the second generation of Anthropic's Constitutional AI framework. The result is a model that does not merely retrieve and recombine information but genuinely reasons through problems in a structured, verifiable way.

Key Benchmark Results at a Glance

  • GPQA Diamond (graduate-level science)94.2%
  • HumanEval (Python code generation)91.3%
  • MATH (competition mathematics)88.4%
  • SWE-bench (real software engineering)48.9%

Benchmark Results at a Glance

GPQA Diamond is particularly meaningful. The dataset consists of questions written by domain experts, specifically PhD-level biologists, chemists, and physicists, that are specifically designed to be difficult even for people with advanced degrees outside the specialty. Human expert accuracy on GPQA Diamond hovers around 65–70%. Claude 4 Opus at 94.2% is not just beating AI models; it is clearly exceeding the average domain expert. This is the clearest indicator yet that frontier models have crossed a qualitative capability threshold in scientific reasoning.

HumanEval at 91.3% continues a trend of rapid improvement in code generation, but the more telling result is SWE-bench: 48.9%. SWE-bench evaluates models on real GitHub issues, tasks that require understanding a large codebase, diagnosing a bug, writing a fix, and ensuring it does not break other tests. It is considerably harder than HumanEval because it demands agentic, multi-step reasoning rather than single-function completion. Reaching nearly 49% on SWE-bench represents a dramatic step toward models that can perform meaningful, unsupervised software engineering work.

"Claude 4 Opus represents our most significant capability advance to date, but we are equally proud that it maintains our strongest safety profile. These goals are not in tension — they reinforce each other." — Anthropic Research Team, Model Card Statement, May 2026

What Changed Under the Hood

Anthropic has been unusually transparent about the architectural choices behind Claude 4 Opus. The model features an expanded context window of 200,000 tokens, consistent with previous Claude models, but with markedly improved coherence across that entire context. Earlier models showed degraded performance when critical information appeared in the middle of a very long context (the so-called "lost in the middle" problem). Internal evaluations suggest Claude 4 Opus largely eliminates this effect, maintaining near-uniform attention quality across positions.

The integration of Extended Thinking is also central to the performance gains. On tasks where Claude 4 Opus is given a generous thinking budget, benchmark scores improve further, in some cases by 5–8 percentage points. For MATH, a benchmark requiring multi-step symbolic reasoning, Extended Thinking pushes the score from 88.4% to an estimated 93% when a 16,000-token thinking budget is enabled. This suggests the architectural improvements and the reasoning scaffold are compounding rather than additive.

Safety First, Still

Anthropic is careful to contextualize the benchmark results within its broader safety mission. Claude 4 Opus was evaluated against the criteria laid out in Anthropic's Responsible Scaling Policy before deployment. Critically, it passed all ASL-3 evaluations, meaning it did not demonstrate capability levels that would require more restrictive deployment controls under Anthropic's internal framework.

On alignment metrics, Claude 4 Opus shows notably improved scores on harmlessness evaluations compared to Claude 3 Opus, despite being far more capable. Anthropic's safety team attributes this to Constitutional AI v2, which tightens the connection between stated values and actual model behavior across a wider range of adversarial prompting strategies. The result is a model that is simultaneously more capable and better aligned. Anthropic argues this combination is achievable by design, not in spite of each other.

For developers and researchers, Claude 4 Opus is available now via the Anthropic API and through Amazon Bedrock. Pricing is set at $15 per million input tokens and $75 per million output tokens, reflecting its position as a premium, research-grade model. For most production use cases, Claude Sonnet 4.5 remains the recommended choice. But for the most demanding reasoning tasks, including scientific research, complex agentic workflows, and advanced code generation, Claude 4 Opus sets a new standard.

Further reading: Learn more about Claude's model family, read our background on Anthropic, or browse the latest Claude AI news.