The Full Claude Plan Structure in 2026

Anthropic now structures Claude access across two parallel tracks: consumer plans for individuals and business plans for organisations. The two tracks are structurally separate — a business on the Team or Enterprise plan is not simply scaling up individual user accounts; they are accessing a different commercial framework with different features, different data handling terms, and different support structures.

Consumer plans (individual users):

  • Free: No charge. Access to Sonnet 4.5 with basic capabilities and usage limits. Appropriate for individual evaluation and low-frequency personal use. Free accounts are subject to usage limits and are deprioritised during peak demand. No enterprise data governance features.
  • Pro ($20/month): Full model access including Opus 4.5, Sonnet 4.5, and Haiku 4.5. Claude Code access. Extended features including Google Workspace integration. Approximately 5x the usage allowance of the Free tier. No SSO, SCIM, or enterprise admin controls.
  • Max ($100–$200/month): Multiplied usage capacity — 5x Pro at the $100 tier, 20x Pro at the $200 tier. Priority access during high demand. Designed for power users with heavy Claude usage who need reliable availability and extended context. Not an enterprise plan; no organisational admin features.

Business plans (organisations):

  • Team Standard ($25/seat/month): Minimum five seats. 200,000-token context window. 1.25x Pro usage allocation. Basic admin console. Collaborative workspace features. Does not include SSO, SCIM, audit logs, DPA, or HIPAA BAA. The correct tier for SMB deployments or early-stage enterprise pilots where compliance requirements are modest.
  • Team Premium ($125/seat/month): Minimum five seats. 6.25x Pro usage allocation. Claude Code included for all seats. Priority support. Appropriate for development-heavy teams where Claude Code is a primary value driver and the Team Standard usage cap is constraining.
  • Enterprise (custom pricing): Minimum 50 seats. Custom token rates. Up to 500,000-token context window. Full compliance stack: SSO, SCIM, audit logging, DPA, HIPAA BAA, Compliance API. Named account management. Custom data retention policies. Role-based access control. Enterprise SLA (99.99% uptime). Pricing negotiated; starts at approximately $50,000/year.

API Token Pricing: The Current Rate Card

The Anthropic API is priced per million tokens of input and output across three model tiers. All figures are per-million-token rates as of early 2026:

  • Claude Haiku 4.5: $1.00 input / $5.00 output per million tokens. Fastest and most economical. Designed for high-volume, latency-sensitive workloads.
  • Claude Sonnet 4.6: $3.00 input / $15.00 output per million tokens. Balanced performance and cost. Anthropic's recommended model for general enterprise API workloads.
  • Claude Opus 4.6: $5.00 input / $25.00 output per million tokens. Premium reasoning capability. Significant reduction from Opus 4.1 era pricing of $15/$75 — a 67% price reduction that reflects model efficiency improvements.

Prompt cache read pricing reduces cost substantially for repeated context: approximately 10% of the standard input rate for cache hits. Cache write pricing (when building the cache) is approximately 25% above standard input rate. For applications with large, stable system prompts or reference documents, prompt caching is typically the single highest-ROI API optimisation available.

Batch API pricing: a flat 50% discount on both input and output tokens for asynchronous workloads. No latency guarantee (completion within 24 hours). The combination of prompt caching and batch API can reduce costs by up to 95% for qualifying high-volume asynchronous workloads.

"Consumption billing creates budget unpredictability that requires contractual safeguards — spending caps, monitoring, and governance — all of which carry their own implementation costs. Include these in your total cost model."

How to Choose Between the Plans: The Decision Framework

The right plan for an enterprise organisation depends primarily on four variables: use case type, user volume, compliance requirements, and consumption patterns. The following framework maps these variables to plan recommendations:

For productivity-focused deployments (knowledge workers, drafting, research, analysis): Subscription plans — Team Standard or Enterprise — are typically more economical than API-direct for conversational productivity use cases. The fixed per-seat cost provides budget predictability and the usage allocation is sufficient for typical knowledge worker interaction patterns. Use Team Standard if compliance requirements are modest. Use Enterprise if your organisation requires SSO, audit logging, DPA, HIPAA BAA, or custom data retention.

For application development and programmatic AI workloads: API access is typically more economical than subscription plans for programmatic use cases at scale. Model choice (Haiku, Sonnet, or Opus) should be driven by the quality requirement of your specific workload, not by default. Haiku for classification and triage; Sonnet for analysis, drafting, and agentic workflows; Opus for complex reasoning tasks where quality demonstrably improves outcomes.

For mixed deployments (both knowledge workers and custom applications): A hybrid model is most common at enterprise scale — Enterprise subscription for knowledge workers, API access for applications. The procurement decision is whether both relationships are managed under a single Enterprise agreement (which simplifies governance and can unlock combined volume pricing) or as separate contracts (which preserves flexibility but creates parallel commercial relationships).

For regulated industry deployments (financial services, healthcare, legal, public sector): Enterprise plan is required to access the compliance features (HIPAA BAA, DPA, audit logs) that make Claude permissible for sensitive workload processing. Do not attempt to deploy sensitive data workloads on Team Standard or Team Premium — they do not include the data handling commitments required for regulated use.

Azure OpenAI vs Direct Anthropic: Pricing Model Comparison

A significant portion of enterprise AI procurement decisions in 2026 involve a choice between accessing Claude through Anthropic directly, accessing OpenAI models through Azure, or both. Understanding how Azure OpenAI and direct Anthropic API pricing compare is essential for making an economically rational platform decision.

Azure OpenAI service prices Claude models (where available) at rates that may differ from direct Anthropic API rates. For organisations with large Azure committed-use discounts (EDP), routing Claude API consumption through AWS Bedrock or Azure — where available — can reduce the effective per-token cost below direct API rates. The AWS EDP discount model, which begins delivering meaningful discounts at approximately $2M or more in annual committed spend, is the most common mechanism for achieving below-rate-card Claude pricing through a cloud channel.

However, Azure OpenAI's primary model availability is OpenAI GPT models, not Anthropic Claude. For most organisations, the "Azure OpenAI vs direct Anthropic" decision is actually a decision about whether to use GPT models (via Azure OpenAI) or Claude models (via direct Anthropic API or AWS Bedrock). These are different AI providers with different pricing structures and different model capabilities — not interchangeable access routes to the same underlying models.

OpenAI enterprise agreements have lock-in provisions that always require scrutiny when comparing to Claude alternatives. Always flag these in your evaluation. The dedicated deployment model, where enterprise customers pay for reserved capacity, creates a different cost structure from Claude's consumption-based API. Consumption billing creates budget unpredictability in both cases — the contractual safeguards and monitoring requirements are similar regardless of provider. The choice between the two providers should be driven by capability fit for your workload, compliance requirements, and the total three-year cost of ownership, not by initial rate card comparison.

Real Enterprise Cost Examples

To ground the pricing framework in practical terms, here are representative cost scenarios based on enterprise deployments we have advised on:

Mid-size professional services firm (200 knowledge workers, moderate AI use): Team Standard at $25/seat/month = $5,000/month = $60,000/year. If compliance requirements trigger a move to Enterprise (SSO, audit logs, DPA): minimum $50,000/year, typically $80,000–$120,000/year at this scale depending on negotiated rate and API consumption volume.

Financial services firm (500 knowledge workers, regulated data workloads): Enterprise plan required. Estimated $200,000–$400,000/year in seat and consumption costs, depending on active user rate, average interaction depth, and whether production applications access the API directly. First-year total cost of ownership including implementation: $350,000–$650,000.

Technology company building Claude-powered product (API-first, 10M interactions/month): At Sonnet 4.6 rates ($3/$15 per million tokens) with an estimated 500 tokens per interaction average, 10M interactions generates 5 billion tokens/month. At standard rate: $15,000–$75,000/month depending on input/output ratio. With prompt caching for large system prompts (60% reduction): $6,000–$30,000/month. With batch API for asynchronous interactions (50% reduction): further reductions to $3,000–$15,000/month for qualifying workloads. Volume discounts under enterprise agreement: additional 10–20% at this consumption level.

What the Pricing Does Not Tell You

Rate card pricing is the starting point for enterprise cost analysis, not the conclusion. Several dimensions of Claude's true cost of ownership are invisible in the pricing table:

Implementation cost: Building production-grade AI applications on Claude requires engineering investment — prompt engineering, evaluation frameworks, integration development, safety testing, and monitoring. This is typically 20 to 40% of first-year licence and API cost for new deployments. Budget accordingly.

Consumption variability: Consumption billing creates budget unpredictability that requires active management — spending caps, monitoring dashboards, and governance processes. The operational cost of this management infrastructure is real and persistent.

Migration cost: If you build significant application infrastructure on Claude's API — custom system prompts, RAG architectures, agentic workflow definitions — the cost of migrating to an alternative provider at contract renewal is substantial. This creates technical lock-in even where contractual lock-in is limited. Factor migration cost into your three-year TCO model.

Model deprecation: Anthropic regularly deprecates older model versions as new generations are released. Migration from a deprecated model to its replacement requires re-evaluation, potentially re-fine-tuning, and quality validation. Budget for model migration cycles — typically once every 12 to 18 months at current development pace.

Need an independent Claude cost model?

We build three-year TCO models for enterprise AI decisions — Anthropic, OpenAI, Google, AWS. No vendor affiliations.

Talk to an Advisor →

Related Reading

For the complete enterprise licensing framework, see our Anthropic Claude Enterprise Licensing Guide 2026. For API pricing detail including rate limits and discount mechanisms, see Anthropic API Pricing: Token Costs, Rate Limits and Enterprise Discounts. For contract negotiation guidance, see 7 Contract Clauses to Negotiate Before You Sign.