Why Claude Opus 4.6 Feels Slower and Less Reliable in OpenClaw Right Now
If your OpenClaw agent on Claude Opus 4.6 has felt slower, more prone to stopping mid-task, or more token-hungry than it was a few weeks ago โ you're not imagining it. Power users are reporting the same thing across VentureBeat, r/ClaudeAI, and Hacker News. Here's what's actually going on and what to do about it.
What Users Are Reporting
The complaints follow a consistent pattern: Opus 4.6 responses that used to feel crisp and confident now feel hedged, slower, or incomplete. Claude Code sessions that used to run autonomously are stopping more frequently to ask for confirmation. Token usage per task has crept up noticeably. Some users describe it as "nerfed," though that framing is probably wrong.
The Real Explanation: Compute Rationing
This is what happens when a lab is preparing to launch a new frontier model. Anthropic is actively preparing Claude Opus 4.7 โ a distinct model from Mythos, expected as soon as this week. When frontier compute is being routed toward training and evaluation of the next model, inference capacity for the current production model gets squeezed.
Anthropic hasn't confirmed this publicly, but the timing matches exactly. The degradation complaints started several weeks ago โ which aligns with Opus 4.7 entering its final training and evaluation phases. It's a predictable symptom, not a deliberate "nerfing."
There's a secondary factor: Uber's CTO revealed this week that Claude Code usage alone has maxed out Uber's entire 2026 AI budget months ahead of schedule. Uber is one account. Multiply that by thousands of enterprise customers and the compute pressure Anthropic is under becomes clear. More demand on the same inference infrastructure = degraded latency and reliability for everyone.
What This Means for OpenClaw Users
If you're running Opus 4.6 as your primary model in OpenClaw, a few practical responses:
1. Route non-critical tasks to a cheaper model
OpenClaw lets you set per-agent model overrides. If your heartbeat cycles, research
tasks, or document drafting don't require Opus-level quality, route them to
claude/sonnet-4-6 or a local Ollama model instead. Reserve Opus for
the work that actually needs it.
// Per-agent model override in openclaw.json
{
"agents": {
"research-agent": {
"model": "claude/sonnet-4-6"
},
"main": {
"model": "claude/opus-4-6"
}
}
}
2. Set explicit timeouts for long tasks
If your agent is stopping mid-task more than before, adding explicit continuation prompts to your SOUL.md helps. Something like: "When a task is taking long, continue without asking for confirmation unless you genuinely need input." The degradation often manifests as unnecessary check-ins, not actual failure.
3. Check your token usage patterns
If token usage has crept up, review your workspace file sizes โ the context bloat bug (#67419) means bootstrap files are re-injected every turn. A 4,000-token workspace overhead compounds fast when the model is already running slower. Trim aggressively. (See our context bloat guide.)
4. Wait it out โ Opus 4.7 is coming
The degradation is temporary. Once Opus 4.7 launches, compute pressure on 4.6 should ease as traffic migrates to the new model. If Opus 4.7 is meaningfully better (early signals suggest it is), you'll want to upgrade anyway.
Should You Switch Models Now?
If the degradation is actively breaking your workflows, yes โ drop to Sonnet 4.6 as your primary model for now. Sonnet is faster, cheaper, and in the current environment may actually be performing comparably to a compute-squeezed Opus 4.6.
If it's minor โ some extra latency, occasional extra confirmation ask โ just wait. Opus 4.7 is days away at most.
TL;DR
- Opus 4.6 degradation is real โ slower, more token-hungry, more stop-and-confirm behavior
- Cause: compute rationing as Anthropic prepares Opus 4.7 (coming this week)
- Compounded by: enterprise demand surge (Uber alone maxed its 2026 AI budget)
- Fix now: route non-critical tasks to Sonnet 4.6 or local Ollama; trim context bloat
- Fix soon: Opus 4.7 expected to land within days โ upgrade when it drops
Need help tuning your OpenClaw model routing?
ClawReady can configure smart model routing for your setup โ Opus for high-value tasks, Sonnet or local models for everything else. Usually cuts API spend 40โ60% with no quality loss on the things that matter.
Book a Free Call โ