Claude Opus 4.6
FrontierAnthropic•Released on 2026-02-05
Anthropic's flagship model with 1M token context (now default), adaptive thinking, and the highest agentic coding scores. Introduced Agent Teams for parallel autonomous coding. Nearly doubled ARC-AGI-2 score over Opus 4.5 (68.8% vs 37.6%).
91
Overall Score
Core Specs
1000K
Context Window
128K
Max Output
✓
Reasoning
✗
Open Source
Multimodal Support
textimage
Scenario Scores
User Feedback Highlights
Based on community feedback. Hover to see original reviews.
+ Highest SWE-bench score (80.8%)+ 128K max output (doubled from 4.5)− Response prefilling removed (breaking change)− Extended thinking deprecated+ Adaptive thinking with effort levels+ Agent Teams for parallel coding+ Best instruction following in complex contexts− Rate limits can be hit quickly on entry-level plans− 2x price of GPT-5.4
Sentiment:👍 80%😐 13%👎 7%
Pros & Cons
Pros
- +Highest SWE-bench score (80.8%)
- +128K max output (doubled from 4.5)
- +Adaptive thinking with effort levels
- +Agent Teams for parallel coding
- +Best instruction following in complex contexts
Cons
- −2x price of GPT-5.4
- −Response prefilling removed (breaking change)
- −Extended thinking deprecated
- −Rate limits can be hit quickly on entry-level plans
Reliability
SLA99.9%
Incidents (30d)44
Last Incident2026-04-08
4/8 故障已解决(监控中)。4月前9天累计 9 次故障,连续3天出问题(4/6-4/8)。
View Status Page →Pricing
Input (per 1M tokens)$5.00
Output (per 1M tokens)$25.00
Subscription$20/month
Updated on 2026-04-21
Get Started with Claude
💡 Free tier uses Sonnet. Upgrade to Pro for Opus.
Tools Supporting This Model
Compare with Others
Benchmarks
sweBenchVerified80.8%
terminalBench265.4%
browseComp84%
gdpvalElo1606%
arcAgi268.8%
gpqaDiamond91.3%
bigLawBench90.2%