Understanding the Licensing Architecture

Before comparing Azure OpenAI and direct OpenAI licensing, it is essential to understand that these are not competing products — they are two commercial access routes to the same underlying model infrastructure. OpenAI develops and trains the foundational models. Microsoft provides Azure OpenAI as a hosted, enterprise-grade wrapper around those same models, backed by Microsoft's cloud infrastructure, compliance certifications, and commercial framework.

This architecture means the licensing decision is fundamentally a commercial and governance decision, not a technical one. The models you access are the same. The APIs are nearly identical. The difference lies in who you have a commercial relationship with, which compliance certifications cover your deployment, what your billing structure looks like, and how much flexibility you retain to change direction if the AI landscape shifts.

Azure OpenAI Licensing Models

Pay-As-You-Go (Standard Deployment)

Standard deployment through Azure OpenAI operates on a consumption pricing model where you are billed per thousand input and output tokens. There is no upfront commitment, no reserved capacity, and no minimum monthly spend. For organisations piloting AI use cases, running infrequent batch workloads, or exploring which AI applications will deliver production value, pay-as-you-go provides maximum flexibility with minimum commitment risk.

The tradeoff is performance variability. Standard deployments share capacity with all other Azure OpenAI customers, meaning response latency can spike during peak demand periods. Microsoft provides no latency SLA for standard deployments. For user-facing applications where consistent sub-second response time is required, standard deployment is unreliable for production use at scale.

Consumption billing also creates budget unpredictability. A production AI application that handles significantly more traffic than projected will generate proportionally higher token costs with no mechanism to prevent overruns beyond manually configured Azure budget alerts. Enterprises must build token rate limiting and budget governance into their application architecture before standard deployments go to production.

Provisioned Throughput Units (PTU)

Provisioned Throughput Units are Azure OpenAI's reserved capacity model. PTU purchases allocate dedicated compute resources to your deployment, ensuring consistent throughput and enabling Microsoft's latency SLA (99th percentile token generation commitment). PTU pricing is calculated per unit per hour, with one-year and three-year reservation options delivering 25 to 30 percent and 35 to 40 percent discounts respectively compared to pay-as-you-go equivalent consumption.

PTU is the correct licensing model for production AI applications that serve users, process customer data, or support mission-critical workflows. The cost efficiency of PTU over pay-as-you-go depends on utilisation: PTU delivers cost savings when capacity is utilised at approximately 65 percent or higher of provisioned throughput. Below this utilisation threshold, pay-as-you-go typically costs less in total even without the latency guarantee.

For large enterprises with predictable AI workload volumes, PTU combined with EA discount structures delivers the lowest total cost for Azure OpenAI. The capital commitment — $2,448 per PTU per month at list before EA discounts — requires CFO-level approval processes and procurement planning that development teams running pilots are not accustomed to managing.

Batch API Pricing

Azure OpenAI offers a Batch API tier that processes requests asynchronously within a 24-hour window, at a 50 percent discount to standard pay-as-you-go token pricing. For AI use cases that do not require real-time responses — document analysis, bulk content generation, data enrichment pipelines, model evaluation — Batch API pricing substantially reduces token costs at the expense of latency. Enterprise procurement teams that overlook Batch API for eligible workloads consistently leave cost reduction opportunities on the table.

Direct OpenAI Licensing Models

API Pay-As-You-Go

Direct OpenAI API access follows the same consumption model as Azure OpenAI standard deployment: pay per thousand tokens, no commitment required, billing through OpenAI directly. List token pricing is identical to Azure OpenAI at comparable model tiers. Direct OpenAI API includes hard billing limits — a control mechanism that Azure OpenAI standard deployment does not match — allowing organisations to cap monthly expenditure at a defined threshold after which API access is suspended until the limit is manually raised.

For organisations that are not Azure customers, direct API access provides the simplest possible path to model capabilities without the overhead of Azure account management, resource group administration, or Azure network configuration. For developers building proofs of concept and internal tools, the direct API's simplicity is a genuine advantage.

ChatGPT Enterprise

ChatGPT Enterprise is OpenAI's SaaS product for enterprise users — it is not an API licensing model. It provides employees with an enterprise-grade interface to ChatGPT with unlimited access to frontier models, custom GPT creation, enterprise admin controls, SSO integration, and data processing agreements that commit OpenAI not to use customer conversations for model training. ChatGPT Enterprise is priced per seat per month on annually negotiated contracts, typically in the $20 to $60 per seat per month range depending on volume.

There is no Azure equivalent to ChatGPT Enterprise. If your requirement is a deployed AI assistant for employee productivity that does not require custom API integration, ChatGPT Enterprise is the relevant product category and it is only available through direct OpenAI agreement. This is a critical distinction that procurement teams conflate with API licensing decisions.

Direct API Enterprise Agreements

Organisations with significant API consumption — above $500,000 annually — qualify for direct OpenAI enterprise API agreements with custom pricing, dedicated support, enhanced SLAs, and custom data governance terms. These agreements are negotiated directly with OpenAI's enterprise sales team and provide volume-based token discounts, model version pinning rights (available through negotiation), and custom BAA and DPA terms for regulated industries.

Direct OpenAI enterprise agreements contain lock-in provisions that enterprise procurement teams must scrutinise carefully. Annual consumption commitments, auto-renewal clauses, model deprecation timelines, and restrictions on fine-tuned model portability are standard features of OpenAI's enterprise template agreement that are disadvantageous to the buyer in their default form. These provisions require active negotiation — they are not offered as concessions, they must be requested explicitly.

Need independent advisory on which licensing model fits your organisation?

We have assessed Azure OpenAI and direct OpenAI contracts across regulated industries.
Talk to Us →

The Lock-In Provisions Both Routes Share

The single most consistently underestimated dimension of GenAI enterprise licensing is vendor lock-in. Both Azure OpenAI and direct OpenAI access routes create lock-in at multiple levels that become progressively more expensive to unwind as AI applications are built on top of specific model APIs, prompt engineering chains, fine-tuned models, and integrated compliance infrastructure.

Model-Level Lock-In

Building production AI applications against specific model versions — GPT-4o, GPT-4 Turbo, embedding models — creates dependency on those model's behaviour. When OpenAI deprecates a model version (typically with 30 to 90 days notice), applications must be retested, prompts may need to be redesigned, and model-specific fine-tuning investments may not transfer to successor models. This rework cost is not covered by any SLA on either access route.

Version pinning rights — the contractual right to continue using a deprecated model version for a defined period after the official deprecation date — can be negotiated into both direct OpenAI enterprise agreements and Azure OpenAI EA addenda. These rights are not standard; they must be requested explicitly during contract negotiation. Organisations that do not negotiate version pinning rights are exposed to forced migration timelines dictated by OpenAI's product roadmap rather than their own operational schedule.

Fine-Tuning Lock-In

Organisations that invest in fine-tuning OpenAI models on proprietary data create a significant lock-in asset. Fine-tuned models on Azure OpenAI are stored as Azure resources and are technically portable within the Azure platform, but cannot be exported as model weights and deployed independently. Fine-tuned models on direct OpenAI API are similarly locked to OpenAI's infrastructure — the training investment cannot be migrated to an alternative AI platform without retraining from scratch.

For organisations making material fine-tuning investments, this lock-in should be assessed carefully before committing to either platform. The alternative — using open-source base models (Llama, Mistral, Falcon) fine-tuned on your own infrastructure — provides portability that neither Azure OpenAI nor direct OpenAI can match, at the cost of higher operational complexity and infrastructure responsibility.

Commercial Commitment Lock-In

Both Azure OpenAI PTU commitments and direct OpenAI enterprise agreement consumption commitments create financial lock-in through prepaid or committed usage. PTU reservations are typically non-refundable; if your AI workload decreases due to project cancellation, business changes, or model substitution, you continue paying for reserved capacity you are not utilising. Direct OpenAI enterprise commitments have similar characteristics. Downward adjustment clauses — the right to reduce committed volume at renewal based on actual usage — should be negotiated explicitly in any multi-year AI licensing agreement.

Compliance and Regulatory Licensing Considerations

Compliance requirements frequently determine the licensing route before commercial analysis begins. Organisations subject to HIPAA are required to execute a Business Associate Agreement (BAA) before processing protected health information through AI services. Both Azure OpenAI and direct OpenAI enterprise agreements can include BAAs, but Azure OpenAI's BAA coverage applies broadly across the Azure platform, while direct OpenAI's BAA requires specific agreement at the enterprise level and is not available to API customers below enterprise threshold.

GDPR compliance requirements introduce further complexity. Azure OpenAI's data processing addendum (DPA) is integrated into Microsoft's standard GDPR-compliant commercial framework, with data processing aligned to Azure's EU Data Boundary for European customers. Direct OpenAI's GDPR compliance requires a standalone DPA with OpenAI, which is available for enterprise customers but requires explicit execution and is not automatic.

For FedRAMP-authorised environments, Azure OpenAI is the only viable route. Direct OpenAI API access does not carry FedRAMP authorisation. US federal agencies and contractors with FedRAMP requirements must access OpenAI models through Azure Government or Azure OpenAI in commercial configurations that have received appropriate authorisation.

"Consumption billing creates budget unpredictability at both OpenAI and Azure. But only direct OpenAI provides a hard billing cap that can prevent your developers from accidentally spending next quarter's AI budget in a week."

Negotiation Leverage: Azure vs Direct OpenAI

Enterprise negotiating leverage differs materially between the two licensing routes, and this asymmetry affects commercial outcomes significantly.

Azure OpenAI negotiations benefit from being embedded within the broader Microsoft commercial relationship. Organisations with existing EA commitments — covering Microsoft 365, Windows, Azure compute, or other Microsoft products — have leverage to negotiate Azure OpenAI pricing, PTU discounts, and support coverage within the EA amendment framework. Microsoft's EA account teams are incentivised to grow total Azure consumption, which creates negotiating space for AI workloads that would be unavailable if Azure OpenAI were negotiated as a standalone procurement.

Direct OpenAI enterprise negotiations are standalone — OpenAI's enterprise sales team negotiates AI platform agreements independently of any other vendor relationship. This simplifies the negotiation process but reduces leverage. An organisation's strongest negotiating position with OpenAI is a credible alternative: Azure OpenAI, Google Vertex AI with Gemini, or Anthropic Claude on AWS Bedrock. Without demonstrating willingness to use an alternative platform, there is limited commercial pressure on OpenAI to deviate from standard enterprise terms.

The strategic recommendation for organisations above $1,000,000 in annual AI platform spend is to run a parallel evaluation of both routes before committing to either, and to use the resulting competitive dynamic to negotiate better terms from whichever vendor is preferred. Both Microsoft and OpenAI will provide improved commercial terms when facing genuine competition — neither will do so proactively without pressure.

Making the Decision: A Licensing Framework

The licensing choice framework reduces to four primary questions:

  • What is your Azure footprint? If you are a large Azure customer with an active EA, Azure OpenAI pricing through EA amendment is almost certainly more competitive than direct OpenAI list pricing. Quantify the EA discount impact before comparing headline token prices.
  • What compliance certifications are required? If FedRAMP, private networking, or EU Data Boundary requirements apply, Azure OpenAI may be the only compliant option regardless of commercial economics.
  • Is ChatGPT Enterprise required? If your use case is deployed as a user-facing AI assistant rather than API integration, ChatGPT Enterprise is a direct OpenAI product category with no Azure equivalent. Confirm whether API licensing or SaaS product deployment is the correct frame.
  • How large is the consumption commitment? For commitments above $500,000 annually, both routes provide enterprise agreement frameworks with custom pricing. For commitments below this threshold, pay-as-you-go pricing applies on both routes and direct OpenAI's billing controls may provide better budget governance.

Download our AI Platform Contract Negotiation Guide

Covers lock-in provisions, SLA gaps, and negotiation tactics for both Azure OpenAI and direct OpenAI.
Download Free →