Anthropic has released Claude Opus 4.6, a direct response to OpenAI’s dominance, specifically targeting complex “agentic AI” workflows. Instead of focusing purely on speed, the model relies on a context window of one million tokens and “adaptive thinking” to solve deep architectural problems like a senior engineer, rather than just delivering fast boilerplate code. We have summarized the technical data, criticism of high latency, and a direct comparison with GPT-5.3 Codex.
Claude Opus 4.6: The most important information
- Anthropic positions Claude Opus 4.6 as a strategic “agentic engine” that stands out from classic chatbots through long-term planning and a massive 1 million token memory.
- Thanks to the new “Context Compaction,” the model now processes entire code repositories or complex project histories in memory without suffering the usual loss of quality under high load.
- In direct comparison to the faster GPT-5.3 Codex, Opus operates more slowly but methodically like a “senior architect” who dynamically validates solutions before producing code.
- For your company, this means maximum precision in critical refactorings and legacy projects, as the model recognizes logical dependencies that pure “speed models” often overlook.
- However, the increased autonomy carries a financial risk: since Opus decides for itself how much computing time (“thinking tokens”) it invests, strict budget controls are necessary to prevent cost explosions.
- Therefore, use Opus 4.6 specifically for complex architecture issues or the modernization of monolithic applications, while continuing to use less expensive models for simple routine tasks.
- Activate “Adaptive Thinking” in the API for maximum problem-solving competence, but be sure to define hard limits for the amount of output to avoid getting caught in an expensive optimization loop.
- Start your first pilot project via the Claude Code CLI by initiating an “agent team” that processes code, database adjustments, and tests in parallel and synchronized.
Summary
- 76% accuracy at full capacity: Thanks to “Context Compaction,” Opus 4.6 dominates the MRCR v2 retrieval benchmark at 1 million context tokens, while Sonnet 4.5 slumped to 18.5%.
- Architect vs. Grinder: In a pure coding comparison (Terminal Bench 2.0), Opus lags significantly behind GPT-5.3 Codex (77.3%) at 65.4%, but scores well on complex dependencies.
- Autonomy as a cost trap: The output limit was doubled to 128,000 tokens, but with the price remaining at $25.00/1M output tokens, autonomous “adaptive thinking” loops risk exploding API costs.
Claude Opus 4.6 marks the transition from a pure language model to an agentic engine. While earlier models were primarily trained on the next token, the architecture here has been fundamentally optimized for long-term planning and autonomous workflows. Two core technologies make this possible: context compaction and adaptive thinking.
Context Compaction: Combating “Context Rot”
A context window of 1,000,000 tokens (beta) sounds impressive on paper, but in the past it often led to the “lost-in-the-middle” phenomenon or “context rot” – accuracy decreased as the memory became fuller.
Anthropic addresses this with Context Compaction. This is a server-side process that automatically summarizes and compresses older parts of conversations without losing their semantic essence. The result is measurable: In the MRCR v2 (retrieval) benchmark, Opus 4.6 achieves 76% accuracy at full load, while its direct predecessor, Sonnet 4.5, dropped to 18.5% under the same load. This allows the model to keep entire repositories in RAM and actively work with them, rather than just passively searching them.
Adaptive Thinking: Dynamic Computational Load
Instead of giving the model a fixed budget of “thinking tokens,” Opus 4.6 introduces adaptive thinking. The model analyzes the complexity of the prompt and independently decides on the necessary “effort level.”
Developers no longer have to guess how much thinking time is needed in the API:
{
"model": "claude-opus-4-6",
"thinking": {
"type": "adaptive"
},
"messages": [...]
}
However, this architecture comes at a price: in “Critical Reception,” the model is sometimes referred to as a “slowpoke” because it “thinks” significantly longer than GPT-5.3 Codex, for example. In return, it acts less as a boilerplate generator and more as a strategic partner that validates architectural decisions before executing them.
Hard facts & economics of agents
To enable autonomous “agent teams” to operate economically, technical limits and prices have been adjusted. The economic efficiency of Opus 4.6 is based on three pillars:
- Doubled output: The limit has been increased to 128,000 tokens (previously 64k). This enables the generation of entire modules in one go, which is essential for agentic loops.
- Stable price structure:
- Input: $5.00 / 1M tokens
- Output: $25.00 / 1M tokens
- Cost efficiency: Context compaction effectively makes the input “cheaper” because fewer redundant tokens need to be processed. However, critics warn that the output price ($25) remains the same as its predecessor, which can quickly become a cost trap with uncontrolled “adaptive thinking” loops.
The almost simultaneous release of Claude Opus 4.6 and GPT-5.3 Codex (only 20 minutes apart) marks a split in the AI market. While OpenAI focuses on speed and raw output, Anthropic positions Opus 4.6 as a methodical strategist – or, in the developer analogy: senior architect vs. speed coder.
Philosophy: Methodology beats speed
Users on HackerNews and Reddit often describe Opus 4.6 as a “collaborator” that actively asks questions instead of blindly generating code. This “latency,” criticized by some as “slowpoke,” is a result of the new Adaptive Thinking. The model dynamically decides how much computing time (“thinking tokens”) to invest in planning before writing the first line of code.
In contrast, GPT-5.3 Codex acts as a “speed demon.” It generates boilerplate code almost instantaneously and is ideal for isolated, clearly defined tickets. Opus, on the other hand, tends toward a defensive coding strategy: it questions architectural decisions and refuses to implement potential anti-patterns until the user explicitly confirms them.
The benchmark reality
The bare figures confirm this qualitative perception. GPT-5.3 Codex clearly beat Opus in the pure Terminal Bench 2.0 (77.3% vs. 65.4%). So if you want to automate “grind” tasks, go for Codex.
However, Opus 4.6 dominates where context and nuances count:
- Humanity’s Last Exam: Here, Opus achieves 40% and leads in multidisciplinary reasoning.
- MRCR v2 (retrieval): With an accuracy of 76% on 1 million tokens (compared to 18.5% for Sonnet 4.5), Opus is the only model that reliably detects dependencies in huge legacy codebases without hallucinating.
Direct comparison: When to use which model?
| Feature | Claude Opus 4.6 (The Architect) | GPT-5.3 Codex (The Grinder) |
|---|---|---|
| Primary focus | Deep reasoning & long-term planning: Simulates a senior engineer who performs code reviews and anticipates race conditions. | Speed & Execution: Simulates a fast mid-level developer who works through tickets (“Get sh\*t done”). |
| Context Handling | 1M Token Compaction: Keeps entire repositories in RAM. Thanks to “context compaction,” the risk of “context rot” during long conversations is reduced. | 128k – 200k: Relies more on RAG (retrieval) than on a massive active window. |
| Coding Style | Cautious & Defensive: Asks: “Should I really do X?” According to the system card, tends toward “over-optimization.” | Aggressive & Fast: Generates code that works immediately, often “good enough,” but less sustainable. |
| Special feature | Agentic Teams: Can split into specialized sub-agents (e.g., API, DB, Test) via CLI, which monitor each other. | Low Latency: Unbeatable at generating standard functions and unit tests. |
Data conclusion: If you need to refactor a monolithic legacy application and are concerned about race conditions or complex dependencies, choose Opus. If you have a blank slate and want to build prototypes quickly, choose GPT-5.3.
Practical guide: Setting up an autonomous dev team with Claude Code
This workflow uses the advanced capabilities of Claude Opus 4.6 to not only edit a legacy codebase linearly, but also to refactor it using parallel agents. Access to the Claude Code CLI is required.
1. Configuration: Enable experimental features
To unlock multitasking agents in the CLI, you must set the experimental flags in your environment. Since Opus 4.6 tends to “over-optimize,” the Adaptive Thinking setting is also crucial. This prevents the model from unnecessarily burning tokens or arguing too flatly.
Activating the agent teams (settings.json):
Navigate to your configuration file and force multi-agent mode:
{
"env": {
"CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS": "1"
}
}
Controlling reasoning depth (API level):
For backend communication (or if you orchestrate the team via API), replace static token budgets with the new dynamic type. This allows Opus 4.6 to determine the “effort level” itself:
{
"model": "claude-opus-4-6",
"max_tokens": 128000, // Use increased output limit!
"thinking": {
"type": "adaptive"
}
}
2. Initialization: Start the agent trio
Instead of a single chat, we spawn specialized instances. In a terminal environment with tmux, Claude can open separate panes for different responsibilities.
The prompt:
Enter the following command in the CLI to enforce role distribution:
“Create an agent team. Spawn three teammates: one for the API layer regarding auth, one for database migration to fix race conditions, and one for test coverage integration.”
The system then initiates three parallel sessions:
- API agent: Focus on endpoints and security.
- DB agent: Focus on schema integrity and locking mechanisms.
- QA agent: Writes tests during development.
3. Execution: Synchronous dependencies
The key advantage over GPT-5.3 Codex here is not speed, but inter-agent communication. Opus 4.6 automatically recognizes dependencies between the generated modules.
- Workflow example: The API agent does not generate mock code, but instead enters a wait state. It sends a signal to the DB agent and waits until the migration of the
userstable is confirmed and the new schema is available. - Context management: Thanks to the 1M token context window and the new context compaction (beta), all three agents effectively share knowledge about the entire repo without older decisions made by the DB agent disappearing in “context rot.”
- Result: In the end, you don’t get isolated code snippets that first have to be integrated manually, but a synchronized module update in which the tests are already adapted to the new API structure.
The latency debate: thinkers vs. doers
While marketing talks about “deep reasoning,” the developer community on Reddit (r/LocalLLaMA) and HackerNews calls it what it is: “Slowpoke. “ Claude Opus 4.6 feels sluggish in direct comparison to the competition.
The reason lies in the architecture. Opus 4.6 acts as a “collaborator” that questions tasks, conducts internal monologues, and weighs architectural decisions. This stands in stark contrast to the almost simultaneously released GPT-5.3 Codex, which is perceived as a “speed demon” and spits out boilerplate code almost instantly.
The raw numbers from Terminal Bench 2.0 reinforce this feeling. Opus achieved a strong 65.4%, but was surpassed by GPT-5.3 Codex with 77.3% just 20 minutes after its release. Anyone who wants to perform fast “grind” tasks (e.g., writing unit tests, customizing CSS) will find the wait time with Opus to be a hindrance.
When AI “over-optimizes” code
Anthropic documents a paradoxical problem in its own System Card: over-optimization.
Opus 4.6 tends to want to further “improve” functioning code in the late stages of generation, even when the requirements have already been met.
- The scenario: The agent has found a solution.
- The problem: Instead of stopping, the model tries to make the code more elegant or compact.
- The result: This often introduces new bugs or overlooks edge cases that were handled correctly in the first, “less elegant” draft.
Developers need to be vigilant here: with Opus 4.6, “better” is sometimes the enemy of “good enough.”
The cost trap: Autonomy comes at a price
At first glance, the pricing structure ($5.00 input / $25.00 output per 1M token) appears identical to its predecessor, Opus 4.5. However, the danger lurks in the new Agentic Architecture.
Features such as Adaptive Thinking now allow the model to decide independently on the depth of reasoning (“Effort Level”). There is no longer a fixed token budget dictated by the user per request. Combined with Agentic Loops (e.g., in Claude Code CLI), where the model autonomously breaks down tasks into sub-steps, this creates a multiplier effect:
- Adaptive Thinking: The model decides to generate thousands of “thinking tokens” for a complex problem.
- Repetitive loops: The agent performs internal reviews and corrects itself multiple times.
- Billing: What used to be an API call is now effectively dozens of internal cycles.
Those who do not set strict limits for max_tokens or budget caps in settings.json risk exploding API costs. Developers report that unsupervised agent teams can quickly rack up 5-digit sums if they get stuck in an optimization loop.
Conclusion
Claude Opus 4.6 is not a simple performance upgrade, but a strategic pivot. Anthropic is deliberately stepping out of the race for sheer generation speed and instead delivering the most stable “thinking engine” to date for complex software architectures. The model is less of a chatbot that spits out code and more of a digital senior developer who critically questions your requirements before implementing them. The result is impressively accurate, but also noticeably sluggish and potentially expensive.
The decision aid:
- Use Opus 4.6 if you need to refactor legacy code where “context rot” has been fatal in the past. If you’re hunting race conditions, need to keep massive repositories in RAM, or need a “second opinion” on system architecture, this model is unrivaled.
- Stay away from it if you just want to do some quick “ticket grinding.” For boilerplate code, standard unit tests, or quickly pulling up prototypes, GPT-5.3 Codex is superior. Opus is too slow for this (“slowpoke” effect) and simply too expensive due to its internal thought processes.
Action:
Don’t go “all-in” on Opus. The professional workflow for the coming months is hybrid: Use Opus 4.6 as an architect and control instance in the planning phase and GPT-5.3 (or Sonnet) as the executing “work drone” for implementation.
Caution: Be sure to set hard budget limits in your API configuration! The new Adaptive Thinking is powerful, but without supervision it can end up in expensive optimization loops that “improve” working code for the worse. Trust is good, cost control is better.





