Anthropic made a significant pricing move in mid-March 2026, removing the long-context premium on its 1M-token context window for Claude Opus 4.6 and Claude Sonnet 4.6. The change applies to all Max, Team, and Enterprise plan users - meaning organizations can now process up to one million tokens of input at standard per-token rates.
The decision removes one of the last major cost barriers to enterprise adoption of long-context AI models, and it arrives at a moment when organizations are increasingly deploying AI agents for document-heavy workflows in finance, legal, and compliance.
What Changed With Claude Opus 4.6
1M-Token Context at Standard Pricing
Previously, using Claude's full 1M-token context window carried a premium charge that could significantly increase per-query costs for enterprise workloads. Anthropic's March 2026 update eliminated that surcharge entirely, making the extended context available at the same rate as shorter queries.
For context, one million tokens is roughly equivalent to 750,000 words - enough to process an entire codebase, a full set of legal contracts, or months of customer interaction logs in a single prompt.
Performance Benchmarks
In internal testing on the MRCR v2 benchmark - a rigorous "needle-in-a-haystack" retrieval test - Opus 4.6 achieved 76% retrieval accuracy at the full one million token depth. This represents a meaningful improvement over earlier models that degraded significantly at extreme context lengths.
Doubled Output Capacity
Claude Opus 4.6 supports up to 128K output tokens, doubling the previous 64K limit. This enables longer thinking budgets for complex reasoning tasks and more comprehensive responses for enterprise workflows that require detailed analysis.
Key Enterprise Features
| Feature | Specification |
|---|---|
| Context Window | 1,000,000 tokens (standard pricing) |
| Max Output Tokens | 128,000 tokens |
| MRCR v2 Accuracy (1M depth) | 76% |
| Adaptive Thinking | Dynamic reasoning allocation |
| Context Compaction | Beta - auto-summarizes older context |
| Effort Levels | Max, High, Medium, Low |
Adaptive Thinking
Anthropic introduced adaptive thinking, which allows Claude to dynamically decide when and how much reasoning is required for a given task. Simple queries receive fast, direct answers, while complex analytical tasks trigger deeper reasoning chains. For enterprise users, this means lower average costs without sacrificing quality on difficult problems.
Context Compaction
A new beta feature called context compaction supports long-running conversations and agentic workflows by automatically summarizing older context as token limits are reached. This is particularly relevant for enterprise agent deployments where conversations span hours or days.
Max Effort Control
A new max effort level joins the existing high, medium, and low settings, offering finer control over token allocation across thinking, tool use, and output generation. Enterprise teams can now calibrate exactly how much computational budget each task type receives.
Enterprise Deployment Options
Claude Opus 4.6 is available across multiple deployment channels, expanding enterprise accessibility:
- Anthropic API - Direct access with standard and batched pricing
- Microsoft Azure Foundry - Claude Opus 4.6 is now available in Microsoft Foundry on Azure, giving Azure customers access within their existing cloud infrastructure
- Amazon Bedrock - Available through AWS for organizations in the Amazon ecosystem
- Google Cloud Vertex AI - Cross-cloud availability for multi-cloud enterprise strategies
This multi-cloud strategy means organizations are not locked into a single vendor to access Anthropic's most capable model.
Industry Applications
Finance and Legal
The combination of 1M-token context and high retrieval accuracy makes Opus 4.6 particularly suited for document-intensive industries. A legal team can load an entire merger agreement - often 500+ pages - into a single prompt and ask specific questions with confidence that the model will locate relevant clauses.
Software Development
Opus 4.6 is considered Anthropic's best model for coding, with the extended context enabling analysis of entire codebases rather than individual files. Development teams report using it for code review, refactoring planning, and architecture analysis across repositories with hundreds of thousands of lines of code.
Compliance and Audit
Organizations managing regulatory compliance across multiple jurisdictions can process entire policy frameworks in a single session, cross-referencing requirements against internal procedures and flagging gaps.
Competitive Landscape
The pricing move comes amid intensifying competition in the enterprise AI market. OpenAI, Google, and other providers have been expanding their own context windows and adjusting pricing, but Anthropic's decision to eliminate the long-context premium entirely - rather than simply reducing it - represents an aggressive positioning strategy.
| Provider | Max Context | Long-Context Premium |
|---|---|---|
| Anthropic Claude Opus 4.6 | 1M tokens | None (removed March 2026) |
| OpenAI GPT-5 | 256K tokens | Tiered pricing |
| Google Gemini 2.5 Pro | 1M tokens | Premium on extended context |
What This Means for Virtual Assistant Services
The standardization of 1M-token context at no premium fundamentally changes what virtual assistant services can deliver. Virtual assistants who leverage Claude Opus 4.6 can now process entire project histories, complete document sets, and months of communications in a single analysis pass - without incurring prohibitive AI costs.
For businesses working with professional virtual assistants, this means faster turnaround on complex research tasks, more comprehensive document analysis, and the ability to maintain context across long-running projects without losing critical details.
The practical impact is significant: a virtual assistant handling contract review, competitive analysis, or financial modeling can now feed entire datasets into AI-assisted workflows at standard rates, dramatically reducing both time and cost for enterprise-grade deliverables. Organizations that pair skilled professional virtual assistants with advanced AI tools like Claude Opus 4.6 are positioned to achieve output quality that previously required much larger - and more expensive - in-house teams.