Why Token Pricing Is a Different Beast

Enterprise software budgets have always been built on predictable inputs: seat counts, server counts, transaction volumes. These are metrics procurement teams know how to model. Token-based AI billing introduces a fundamentally different variable — one that scales not with the number of users or the number of calls, but with the depth and complexity of every individual interaction. A user who asks a simple question consumes tens of tokens. A user who asks the same system to analyse a 50-page contract consumes tens of thousands. The same feature, the same user, the same application — but radically different cost.

This nonlinearity is the source of token pricing risk. Most enterprise budget models assume linear scaling from pilot to production. Token consumption is rarely linear. Pilot users — typically early adopters chosen for enthusiasm — tend to use AI tools in structured, optimised ways. Production users use them in the ways that naturally occur to them: long conversations, large document uploads, repeated iteration, complex multi-step reasoning chains. The difference between pilot consumption and production consumption is frequently a factor of three to five. We have seen it reach ten.

Consumption billing creates budget unpredictability at precisely the moment when AI adoption is accelerating fastest and when the organisational appetite to constrain usage is lowest. That is the structural risk of token pricing.

The Hidden Multipliers in Token Consumption

Beyond the basic input-output token count, several factors multiply consumption in ways that are not visible in vendor pricing pages or pilot metrics:

Context window depth: Large language models operate on context — the full history of a conversation, plus any documents or data provided as input. The longer the context window, the more tokens every subsequent exchange consumes. An LLM with a 200,000-token context window enabled for a complex research workflow will consume dramatically more tokens per session than the same model used for short-form queries. Enterprises enabling extended context windows for legal, financial, or scientific use cases should model context-depth separately in their cost projections.

Retry behaviour: Applications that automatically retry failed or low-quality responses — a common pattern in agentic AI systems — multiply token consumption by a factor equal to the average retry rate. A system that retries 20% of calls effectively adds 20% to your token bill without adding 20% to visible output. This is a cost that lives in your AI middleware and is effectively invisible to budget owners unless they instrument for it.

Prompt engineering overhead: Well-engineered prompts include extensive system instructions, few-shot examples, and structured output requirements. These add tokens to every call, often matching or exceeding the user-visible input in volume. We have reviewed production AI applications where the system prompt consumed 40% of total token budget per call — a cost that was not modelled in the initial procurement.

Hybrid pricing compounding: Nearly half of AI vendors now combine subscription fees with usage-based charges. When the usage charge is token-based, it creates a compounding cost that is difficult to attribute to either the subscription or the consumption budget line. Finance teams managing AI contracts need to understand which portion of cost is fixed and which is variable, and track them separately.

"Pilot users consume tokens in structured, optimised ways. Production users do not. The consumption gap between pilot and production is frequently a factor of three to five."

The Budget Accountability Gap

Token pricing creates a budget accountability gap that traditional enterprise IT governance structures do not close. In a seat-based licensing world, IT procurement owns the budget and the contract. In a consumption-based world, the actual cost is driven by business users, developers, and application behaviours that are outside procurement's direct control.

The CFO who approves a $500,000 annual AI budget based on vendor pricing sheets and pilot data may be approving an actual cost that could reach $1.5 million under production conditions — and neither the CFO nor the IT procurement team will know until the invoice arrives. By that point, usage is embedded in business processes, and the cost reduction options are either constrain user access (which drives business pushback) or renegotiate (which requires leverage you no longer have).

The governance solution is to build token cost monitoring into AI application architectures from the start — not as an afterthought. Every production AI application should have cost dashboards that show token consumption by team, by application, by use case, and by user cohort. These dashboards should be reviewed by business owners and budget holders monthly, not just by IT. When consumption is owned by the business, the business is motivated to optimise it.

How to Model Token Risk Before You Commit

Accurate token cost modelling before an enterprise AI commitment is possible, but it requires a more granular approach than most procurement teams apply. The following methodology has been effective across our client base:

Step 1 — Define interaction archetypes: Rather than modelling average token consumption, identify the two or three most common interaction types for each AI use case. A legal contract review use case might have: (a) short queries on contract terms — 2,000 tokens average; (b) full contract analysis — 40,000 tokens average; (c) redline and commentary generation — 80,000 tokens average. Model each archetype separately and estimate the proportion of total sessions each represents.

Step 2 — Apply a production multiplier: For each archetype, apply a 1.5x to 2.5x multiplier to account for context overhead, retry behaviour, and real-world prompt complexity versus pilot conditions. Use 1.5x for well-engineered, structured applications and 2.5x for exploratory or conversational use cases.

Step 3 — Model adoption curves, not flat rates: Token consumption scales with adoption, but adoption is not linear. Model three curves — conservative (20% of target users by month 6, 60% by month 18), base (50% by month 6, 90% by month 12), and aggressive (80% by month 6, 100% by month 9). Compare total three-year token cost across all three curves to understand your budget range rather than your budget point.

Step 4 — Price test against alternatives: OpenAI enterprise agreements have lock-in provisions — always flag these in your evaluation. Azure OpenAI and direct OpenAI API pricing differ materially, and the model that is cheapest for your archetype mix may not be the market-leading model by benchmarks. Compare pricing models explicitly: Azure OpenAI offers predictable committed-use pricing structures that differ from the pay-as-you-go API pricing on OpenAI direct. The consumption billing creates budget unpredictability in both cases, but the mechanisms differ and the procurement leverage differs.

Contractual Provisions That Reduce Token Pricing Risk

Once you have modelled token risk, you need contractual provisions that protect against the scenarios at the upper end of your consumption range. These are the provisions we negotiate most consistently for enterprise AI clients:

In one engagement, a Fortune 500 financial services company negotiated a $2.4M annual Azure OpenAI consumption budget but experienced 47% token overspend in Q1 due to unexpected context window usage in their compliance document review application. Redress advised on contract amendments that capped quarterly consumption at 1.8x baseline, included most-favoured-nation pricing, and secured annual volume tier re-evaluation rights. The modified agreement reduced projected three-year costs by $1.1M while protecting the vendor's volume commitments. The engagement fee represented 3.2% of the identified cost savings.

Monthly spending caps with auto-pause: A hard monthly cap on API spending, with automatic throttling or suspension of consumption once the cap is reached, is the single most effective protection against runaway token costs. Most vendors offer this natively at the API level. Ensure it is contractually confirmed — not just a feature that can be changed — and that your cap levels are set at a meaningful multiple of your expected consumption, not at a level that would interrupt business operations.

Price protection clauses: Token prices have generally fallen over time as model efficiency improves, but that is not a contractual guarantee. Enterprise AI agreements should include a most-favoured-nation pricing provision: if the vendor reduces its published token prices, your contracted rate decreases to match within 30 days. This prevents the situation — which we have seen — where enterprise customers are paying legacy contracted rates that are materially above the published API rate.

Volume tier re-evaluation rights: If your consumption grows significantly, your pricing should improve. Include an annual right to re-evaluate your volume tier — defined by rolling 90-day consumption — and to access the appropriate volume discount without renegotiating the full agreement. This prevents consumption growth from being an unilateral benefit to the vendor.

Termination for material cost increase: If consumption-based charges in any quarter exceed a defined multiple (typically 1.5x to 2x) of the contracted estimate without a corresponding business justification, include a right to trigger a pricing review and, if agreement is not reached within 60 days, to terminate without penalty. This provision is your ultimate leverage against consumption shock scenarios.

AI cost model review

We review AI consumption contracts and cost models as part of our GenAI advisory service. Identify the token risk in your current agreements before it becomes a budget crisis.

Learn More →

What the Market Data Tells Us

The market data on AI token cost overruns is unambiguous. Industry research shows that 65% of IT leaders report unexpected charges from consumption-based AI pricing models, with actual costs frequently exceeding initial estimates by 30 to 50%. Average monthly AI spending reached $85,521 in 2025 — a 36% increase from 2024 — and the acceleration is continuing. Organisations managing two or three different AI vendor contracts are managing two or three different pricing structures simultaneously, with different token definitions, different context window costs, and different rate card structures.

The Deloitte AI Tokens report specifically identifies that AI costs are measured not in licences or cores but in tokens — a fundamental departure from the cost models enterprises have relied on for decades. The governance and financial management infrastructure for token-based costs does not yet exist in most organisations. Building it is the immediate priority for any enterprise that has moved AI from pilot to production deployment.

The practical implication: do not manage AI token costs retrospectively. Build the monitoring, the governance, and the contractual protections before consumption scales. The enterprise procurement teams that get ahead of this now will have cost structures and negotiating positions that their counterparts will spend the next two years trying to recover.

Related Reading

For more on the commercial dimensions of enterprise AI procurement, see our GenAI Advisory Services and our analysis of Anthropic API Pricing and Token Costs. For OpenAI-specific guidance, our CIO Playbook on OpenAI Negotiations covers the contract clauses that matter most.