What's Happening
Over the past 48-96 hours, multiple Claude Code users have reported dramatic, unexplained increases in usage consumption. Developers describe identical prompts on identical repository states that previously consumed ~5% of their 5-hour window now consuming the entire allocation. One user reported that simply typing "hey" in a previously opened session increased usage by 22%.
This follows Claude Code's recent surge in adoption, with the tool surpassing 100,000 stars on GitHub just last week. The timing suggests this could be related to increased load or backend changes as Anthropic scales infrastructure.
Immediate Diagnostic Steps
First, verify if you're affected. Run this command to check your current session usage:
claude code status --detailed
Look for unusual patterns in the "Usage" section. If you see:
- Single messages consuming >20% of your window
- Idle sessions accumulating usage
- Identical tasks costing significantly more than yesterday
You're likely experiencing the issue. The Reddit user's screenshot shows 46% usage from just 4-5 messages across sessions left open overnight.
Workarounds That Work Right Now
1. Close and Recreate Sessions
Don't leave Claude Code sessions open overnight or idle for extended periods. The Reddit report suggests idle sessions may continue accumulating usage. Instead:
# Instead of keeping sessions open
claude code sessions list
claude code sessions close <session_id>
# Create fresh sessions for new work
claude code --new-session
2. Use the /compact Flag Aggressively
When working with large repositories, force token optimization:
claude code --compact "refactor this module"
The /compact flag reduces context window usage by up to 40% in our previous testing. This is crucial when backend changes might be affecting token counting.
3. Monitor with CC-Lens
Install the open-source dashboard we covered last week:
# Install CC-Lens for real-time usage monitoring
pip install cc-lens
cc-lens --monitor --interval 30
CC-Lens gives you granular visibility into which operations consume the most usage, helping identify problematic patterns.
4. Switch to Local Execution Temporarily
For critical work, consider running Claude Code locally with Ollama (as covered in our March 25 guide):
# Set up local execution to bypass usage limits entirely
claude code --local --model ollama:codellama
While local models may be less capable than Opus 4.6, they're unlimited and consistent.
What Anthropic Needs to Address
Based on user reports, the issue appears to be:
- Session persistence problems: Idle sessions accumulating usage
- Token counting inconsistencies: Identical prompts costing dramatically more
- Model switching issues: Users report "Switching to Sonnet doesn't even help"
Until Anthropic provides an official fix or explanation, these workarounds are your best defense against unexpected limit exhaustion.
Long-Term Protection
Update your CLAUDE.md to include usage optimization directives:
# Usage Optimization Directives
## Session Management
- Close sessions after 30 minutes of inactivity
- Use `/compact` mode for all file operations
- Batch related changes into single prompts
## Token Management
- Prefer Sonnet for exploratory work
- Reserve Opus 4.6 for complex reasoning only
- Use `--max-tokens 2000` for simple responses
This aligns with our March 26 coverage of configuring memory systems for productivity gains—now it's about configuring for cost control.
Next Steps
- Report issues: Use
claude code feedbackwith detailed reproduction steps - Monitor announcements: Watch for updates on the Claude Code GitHub repository
- Adjust workflows: Assume higher costs until resolved; plan sessions accordingly
The community needs transparency on whether this is a bug, intentional pricing change, or infrastructure scaling issue. In the meantime, protect your development time with these defensive measures.






