Anthropic has launched Claude Opus 4.7, it’s latest frontier model and a direct successor to Claude Opus 4.6. The release is positioned as a focused improvement rather than a full generational leap, but the gains it delivers are substantial in the areas that matter most to developers building real-world AI-powered applications: agentic software engineering, multimodal […] The post Anthropic Releas

Anthropic has launched Claude Opus 4.7, it’s latest frontier model and a direct successor to Claude Opus 4.6. The release is positioned as a focused improvement rather than a full generational leap, but the gains it delivers are substantial in the areas that matter most to developers building real-world AI-powered applications: agentic software engineering, multimodal reasoning, and long-running autonomous task execution. https://www.anthropic.com/news/claude-opus-4-7 What Exactly is Claude Opus 4.7? Anthropic maintains a model family with tiers — Haiku (fast and lightweight), Sonnet (balanced), and Opus (highest capability).

Opus 4.7 sits at the top of this stack, below only the newly previewed Claude Mythos, which Anthropic has kept in a restricted release. Opus 4.7 represents a notable improvement on Opus 4.6 in advanced software engineering, with particular gains on the most difficult tasks. Crucially, users report being able to hand off their hardest coding work — the kind that previously needed close supervision — to Opus 4.7 with confidence, as it handles complex, long-running tasks with rigor and consistency, pays precise attention to instructions, and devises ways to verify its own outputs before reporting back.

The model verifying its own outputs is a meaningful behavioral shift. Earlier models often produced results without internal sanity checks; Opus 4.7 appears to close that loop autonomously, which has significant implications for CI/CD pipelines and multi-step agentic workflows. Stronger Coding Benchmarks Early testers have put some sharp numbers on the coding improvements.

On a 93-task coding benchmark, Opus 4.7 lifted resolution by 13% over Opus 4.6, including four tasks that neither Opus 4.6 nor Sonnet 4.6 could solve. On CursorBench — a widely-used developer evaluation harness — Opus 4.7 cleared 70% versus Opus 4.6 at 58%. And for complex multi-step workflows, one tester observed a 14% gain over Opus 4.6 at fewer tokens and a third of the tool errors — and notably, Opus 4.7 was the first model to pass their implicit-need tests, continuing to execute through tool failures that used to stop Opus cold.

Improved Vision: 3× the Resolution of Prior Models One of the most technically concrete upgrades in Opus 4.7 is its multimodal capability. Opus 4.7 can now accept images up to 2,576 pixels on the long edge (~3.75 megapixels), more than three times as many pixels as prior Claude models. Many real-world applications — from computer-use agents reading dense UI screenshots to data extraction from complex engineering diagrams — fail not because the model lacks reasoning ability, but because it can’t resolve fine visual detail.

This opens up a wealth of multimodal uses that depend on fine visual detail: computer-use agents reading dense screenshots, data extractions from complex diagrams, and work that needs pixel-perfect references. The impact in production has already been dramatic. One tester working on computer-use workflows reported that Opus 4.7 scored 98.5% on their visual-acuity benchmark versus 54.5% for Opus 4.6 — effectively eliminating their single biggest Opus pain point.

This is a model-level change rather than an API parameter, so images users send to Claude will simply be processed at higher fidelity — though because higher-resolution images consume more tokens, users who don’t require the extra detail can downsample images before sending them to the model. https://www.anthropic.com/news/claude-opus-4-7 A New Effort Level: xhigh, Plus Task Budgets Developers working with the Claude API will notice two new levers for controlling compute spend. First, Opus 4.7 introduces a new xhigh (‘extra high’) effort level between high and max, giving users finer control over the tradeoff between reasoning and latency on hard problems.

In Claude Code, Anthropic team has raised the default effort level to xhigh for all plans. When testing Opus 4.7 for coding and agentic use cases, Anthropic recommends starting with high or xhigh effort. Second, task budgets are now launching in public beta on the Claude Platform API, giving developers a way to guide Claude’s token spend so it can prioritize work across longer runs.

Together, these two controls give developer teams meaningful production levers — especially relevant when running parallelized agent pipelines where per-call cost and latency must be managed carefully. New in Claude Code: /ultrareview and Auto Mode for Max Users Two new Claude Code features ship alongside Opus 4.7 that are worth flagging for devs who use it as part of their development workflow. The new /ultrareview slash command produces a dedicated review session that reads through changes and flags bugs and design issues that a careful reviewer would catch.

Anthropic is giving Pro and Max Claude Code users three free ultrareviews to try it out. Think of it as a senior engineer review pass on demand — useful before merging complex PRs or shipping to production. Additionally, auto mode