Understanding Google Cloud Committed Use Discounts

Google Cloud Committed Use Discounts are contractual commitments to reserve and pay for a specified amount of compute capacity, memory, storage, or data transfer over a fixed contract term (one or three years). In return for the upfront commitment, Google offers tiered discount rates ranging from 15 to 57 percent below on-demand pricing, depending on the resource type, machine series, and commitment duration.

CUDs are fundamentally different from Reserved Instances (AWS) or Capacity Reservations (Azure) because they are discount mechanisms, not resource reservations. A CUD commitment to 100 vCPUs does not require you to provision exactly 100 vCPUs at all times. Rather, you receive a discount on any usage up to that commitment level, and you pay on-demand rates for any usage beyond the commitment. This flexibility is the primary strategic advantage of CUDs in cloud-native and variable workload environments.

Google Cloud offers two primary types of CUDs: resource-based and spend-based. Resource-based CUDs commit to a specific quantity of vCPU, memory, GPU, or TPU resources. Spend-based CUDs commit to a monthly spend amount across a defined set of services (compute, BigQuery, Cloud Storage) without specifying individual resource quantities. Spend-based CUDs are ideal for organisations with diverse and unpredictable consumption patterns, while resource-based CUDs deliver maximum savings for stable, predictable workloads.

The discount tiers vary significantly by machine series. N1 machines, the oldest generation, offer discounts of up to 57 percent on three-year commitments. N2 machines deliver up to 55 percent discounts. C3 machines, the most recent generation, offer up to 55 percent discounts but do not qualify for Sustained Use Discounts (SUDs) when purchased as CUDs—a critical consideration in multi-year planning.

"40 percent of enterprises analysed across 500+ Google Cloud engagements deploy no CUDs at all. The median organisation leaves $180,000 to $420,000 annually in potential savings unused." — Morten Andersen, Co-Founder Redress Compliance

The July 2025 CUD Revolution: Multi-Price SKU Model

In July 2025, Google fundamentally restructured its CUD pricing model to accommodate dynamic pricing across machine types and SKUs. The legacy flat-rate CUD model assumed a single price per resource type, which created opaque and inflexible pricing. The new multi-price or multi-SKU CUD model breaks pricing down by individual resource type, region, and machine series, providing far greater transparency and enabling more granular commitment planning.

This change has three major implications for enterprise cost management. First, it enables more accurate cost modelling because you can now commit to specific machine series rather than aggregating costs across heterogeneous types. Second, it allows dynamic resource substitution—you can apply commitments to whichever machine series provides the best price-to-performance ratio in real-time, without lock-in to a specific SKU family. Third, it integrates more naturally with the new CUD Analysis tool, which provides hourly granularity data for optimisation.

Organisations with existing CUD contracts signed before July 2025 should perform a renegotiation review during the next annual contract renewal. The multi-SKU model often permits price reductions of 3 to 8 percent compared to legacy flat-rate structures, and Google sales teams routinely apply these improvements only upon customer request.

The Google Cloud CUD Analysis Tool: Unlocking Hourly Granularity

In response to the complexity created by multi-SKU pricing, Google released the new CUD Analysis tool, which provides hourly breakdown of commitment utilisation up to 30 days of historical data. This tool is the single most powerful instrument for diagnosing under-commitment and over-commitment scenarios and sizing optimal CUD amounts.

The traditional approach to CUD planning relied on monthly billing summaries, which obscured critical utilisation spikes and troughs. An organisation might show 80 percent average monthly utilisation when, in fact, it operates at 40 percent on nights and weekends and spikes to 120 percent during peak hours. The CUD Analysis tool exposes these patterns with hourly resolution, enabling commitment sizing that matches actual usage curves rather than monthly averages.

Using the CUD Analysis tool, you can identify the baseline utilisation level—the minimum capacity you consistently require—and commit that amount. You then size your on-demand capacity to cover the peak utilisation above the baseline. This approach minimises wasted commitment capacity and maximises savings.

The tool integrates directly with the GCP negotiation leverage framework because it provides quantifiable data required for enterprise pricing discussions. A Google Cloud account with consistent 400 vCPU baseline utilisation and 800 vCPU peaks has a very different discount justification than an account showing random daily spikes. That quantifiable data becomes leverage in negotiation.

Commitment Coverage Strategy: 60-80 Percent Optimal Ratio

The optimal commitment coverage ratio—the percentage of expected usage you should commit to—falls in the 70-80 percent range for most enterprise workloads. This ratio balances several competing objectives: maximising savings through commitment discounts, maintaining flexibility for variable usage, and avoiding wasted commitment capacity.

If you commit to 100 percent of your expected usage, you lose flexibility entirely. Any workload decrease leaves you with unused commitment. Any workload increase forces you to pay on-demand premiums for spillover capacity. If you commit to only 40 percent, you forgo the 55 to 57 percent discounts on the majority of your baseline utilisation and pay full on-demand pricing, which eliminates most savings opportunities.

The 70-80 percent sweet spot works as follows. You identify your baseline utilisation—the minimum capacity you consistently consume—and commit to 70-80 percent of that baseline. You leave 20-30 percent of baseline capacity unbilled on-demand, providing flexibility without significant cost impact. You size your aggregate on-demand budget to cover expected peaks above the committed amount. This strategy is particularly effective when combined with Google Cloud sustained use discounts on the uncommitted on-demand portion.

C3 machines introduce a tactical complexity here. Because C3 does not qualify for SUDs when used with CUDs, organisations often commit to 60 percent of C3 usage and leave 40 percent on-demand, accepting higher on-demand costs in exchange for C3's superior price-to-performance ratio.

Resource-Based Versus Spend-Based CUDs: Choosing the Right Model

Resource-based CUDs commit to specific quantities of vCPU, memory, GPU, or storage. These commitments offer maximum transparency because you know exactly what you are paying for and can track utilisation against commitment. They are ideal for organisations with stable, predictable workloads (batch processing, legacy application servers, RDBMS databases) and for cost attribution models that require line-item visibility.

Spend-based CUDs commit to a monthly spend amount (e.g., $50,000 per month) across a defined set of services. They are ideal for organisations with highly variable consumption patterns, diverse service portfolios, and dynamic workloads (Kubernetes clusters, data pipelines, ML training jobs) because they allow flexibility in which services consume the commitment without penalty.

A hybrid approach—combining resource-based commitments for stable workloads with spend-based commitments for variable ones—often yields the best results. You might commit to 300 vCPUs of N2 capacity for your stable ERP and CRM systems, then apply a $100,000 per month spend-based commitment to cover Kubernetes, BigQuery, and other variable services. This strategy locks in savings on predictable costs while maintaining flexibility where it matters most.

The C3 Machine Series Decision: Strategic Implications

Google's third-generation machine series (C3) represents a significant leap in price-to-performance, offering approximately 25 to 30 percent better compute density than N2 at similar pricing. However, C3 machines do not qualify for Sustained Use Discounts—only CUDs. This creates a counterintuitive strategic decision: organisations must choose between C3's superior performance or maximising SUD benefits on N2.

For workloads that require maximum performance or where CPU density directly impacts business outcomes (real-time analytics, low-latency trading systems, high-throughput data processing), C3 with CUDs is optimal. For general-purpose compute where performance is adequate and cost minimisation is the priority, N2 with both CUDs and SUDs often proves more economical.

This decision becomes a line-item negotiation point in Google Cloud CUD negotiation because it affects both discount rates and utilisation guarantees. A Google Cloud commercial team may propose C3 to improve margins, but the customer's true economic interest may lie in N2 to preserve SUD eligibility.

BigQuery CUDs: Slot Commitments and New Pricing Models

BigQuery, Google's serverless data warehouse, launched Committed Use Discounts at Google Cloud Next 2025 in the form of slot commitments. Rather than committing to query volume or data scanned, you commit to purchasing "slots"—units of query processing capacity available continuously. BigQuery Editions introduced three tier levels: Standard Edition at $0.04 per slot-hour, Enterprise Edition at $0.06 per slot-hour, and Enterprise Plus Edition at $0.08 per slot-hour.

The economic break-even point for BigQuery on-demand vs capacity reservations occurs at approximately 467 TiB per month. Any organisation querying more than this volume should evaluate slot commitments, as capacity pricing becomes more economical than on-demand query billing. For organisations with sustained BigQuery workloads, slot commitments deliver 25 to 40 percent savings compared to on-demand pricing, depending on query complexity and the edition selected.

BigQuery CUDs represent a significant evolution because they are the first service-level CUD offering that is independent of compute machine types. This opens opportunities for data-centric organisations to stack CUDs across compute infrastructure, BigQuery capacity, and storage in a unified commitment strategy.

Enterprise Discounts and Private Pricing Agreements

At the $500,000 annual spend threshold and above, enterprise organisations become eligible for layered discounts combining CUDs with Google Cloud Private Pricing Agreements (PPAs). Where CUDs provide 15 to 57 percent discounts based on commitment duration and machine type, PPAs provide an additional 20 to 40 percent discount at the enterprise tier, negotiated on a case-by-case basis.

The mechanics are straightforward: you commit to a CUD for your baseline consumption and obtain a negotiated PPA discount rate applied to both committed and on-demand usage. A $1.2 million annual GCP bill might commit to $700,000 in CUDs (60 percent of spend) at 40 percent discount, resulting in $420,000 committed cost. The remaining $500,000 of on-demand usage receives a negotiated 25 percent PPA discount, reducing it to $375,000. Total annual cost: $795,000 versus list price of $1.2 million, representing a 34 percent blended discount.

Most organisations fail to leverage the dual CUD plus PPA opportunity because Google's sales process typically presents CUDs and PPAs as competing alternatives rather than complementary strategies. In Google Cloud PPA negotiation, explicitly request the combination pricing and provide the CFO-level business case: the committed amount de-risks Google's revenue (valuable to their sales forecast), and the PPA discount on top-line consumption incentivises long-term relationship growth.

Need to optimise your Google Cloud CUD strategy?

Our team has completed 500+ GCP engagements. We'll model your current commitment portfolio and identify savings opportunities.
Speak with our Google Cloud team →

Google Fiscal Year Timing and Negotiation Windows

Google's fiscal year ends on September 30, which creates predictable demand and margin cycles. Q4 (July-September) represents the optimal negotiation window for CUD renewals, contract expansion, and PPA renegotiation. During this period, Google's sales teams are under maximum pressure to meet annual targets, and finance teams are incentivised to lock in large deals rather than push them into the following fiscal year.

An organisation planning a CUD renewal or a $500,000+ annual Google Cloud investment should deliberately schedule negotiations for July through September. A 10-20 percent improvement in PPA rates—a negotiable outcome during Q4—can translate to $100,000 to $300,000 in savings over three years.

The Google Cloud Cost Governance Framework

CUDs optimise pricing, but cost governance prevents waste. The FinOps Framework 2025/2026, expanded to cover Cloud+, defines governance as the systematic control of cloud spending through organisational policies, tooling, and accountability.

Effective cost governance for Google Cloud requires four components: cost allocation, budget controls, usage monitoring, and chargeback. Cost allocation begins with comprehensive labeling and project hierarchies that map cloud resources to business units, cost centres, and projects. Google Cloud's billing export feature, which exports billing data to BigQuery, enables sophisticated cost attribution and trend analysis across dimensional hierarchies.

Budget alerts and spend velocity monitoring prevent surprises. Google Cloud's Budget API integrates with alerting systems to notify teams when forecasted spending exceeds thresholds. This is particularly critical for services with consumption-based pricing (BigQuery, Dataflow, Cloud Run) where usage spikes can dramatically increase costs in single billing cycles.

Usage monitoring across the Google Cloud cost allocation and tagging framework enables right-sizing decisions. Organisations should establish monthly cost reviews by business unit, identifying under-utilised resources and optimisation opportunities. The CUD Analysis tool feeds directly into these reviews, providing quantitative data on commitment utilisation and spillover costs.

Chargeback—allocating cloud costs to internal business units based on consumption—creates accountability and incentivises cost-conscious architecture decisions. Without chargeback, engineering teams optimise for speed and feature velocity without considering cloud economics. With chargeback, they balance speed and cost.

Cloud Billing Export and FOCUS 1.2 Standardisation

The FinOps Framework 2025/2026 emphasises the FOCUS 1.2 specification, a unified billing data format designed to consolidate costs across cloud, SaaS, and PaaS consumption into a single schema. Google Cloud's billing export now supports FOCUS format, enabling organisations to build unified cost analytics across multiple cloud vendors (Google Cloud, AWS, Azure) and SaaS applications within a single data warehouse.

This standardisation is strategically important because it enables cost governance at enterprise scale. Rather than maintaining separate cost models for GCP, AWS, and Microsoft, you maintain a single FOCUS-compliant data model. BI tools and forecasting engines can consume FOCUS data natively, reducing integration overhead and accelerating time-to-insight.

Gemini AI Licensing and Google Cloud Workspace Dynamics

Google Workspace underwent a 17-22 percent price increase in January 2025, embedding Gemini AI into all standard plans. Gemini has five separate licensing channels: basic Gemini in Workspace (standard), Gemini Advanced (premium), Gemini API (consumption-based), Gemini Business (enterprise), and Gemini Workspace Intelligence (enterprise workflows). Understanding these channels is critical because they affect total cost of ownership for organisations leveraging Google Cloud infrastructure alongside Workspace productivity.

For organisations purchasing Google Workspace at scale, the embedded Gemini features may provide sufficient AI capability, eliminating the need to purchase separate Gemini subscriptions. Conversely, organisations deploying Gemini APIs for custom applications must model API consumption alongside CUD and PPA strategies. The Google Gemini enterprise licensing guide provides detailed pricing models and breakeven analyses across these channels.

Google Workspace licensing negotiation typically occurs on a separate track from Google Cloud negotiation, but organisations should consider bundling both into a unified RFP process. Larger combined commitments (e.g., Workspace + Cloud + Gemini) often unlock multi-product discounts in the 15-25 percent range that individual products cannot achieve alone.

Advanced Strategies: Commitment Layering and Portfolio Optimisation

Sophisticated organisations layer multiple commitment vehicles to optimise total savings. A typical enterprise GCP portfolio might include: (1) three-year resource-based CUDs for stable baseline compute, (2) one-year spend-based CUDs for variable services, (3) a PPA discount rate applied to all usage, and (4) spot/preemptible VM discounts for non-critical batch workloads. This layered approach captures savings across multiple dimensions without sacrificing flexibility.

Portfolio optimisation also means actively rebalancing commitments as workload patterns evolve. The CUD Analysis tool should be run quarterly, feeding into a commitment review process that adjusts commitment amounts as baseline utilisation grows or declines. Many organisations lock in multi-year commitments and never revisit them, leaving significant savings on the table if utilisation grows 30-50 percent above the original commitment baseline.

The most significant hidden cost for enterprises operating multi-region Google Cloud infrastructure is reducing Google Cloud egress costs, which represents another optimisation vector often overlooked in CUD discussions. Data transfer costs—egress to the internet, cross-region replication, and inter-zone data movement—can consume 10-25 percent of total GCP costs in data-intensive environments. For enterprises running data lakes, analytics pipelines, and real-time replication across regions, egress costs often exceed compute costs. Architectural decisions about data placement, regional distribution, caching layers, and content delivery integration directly impact egress costs and should factor into commitment planning.

A 5,000-person enterprise running a multi-region Google Cloud deployment with distributed data pipelines might incur $400,000 annually in compute costs and $200,000 in egress costs. Optimising egress architecture—consolidating data flows, implementing caching strategies, and designing efficient replication—can reduce egress costs by 30-40 percent, offsetting additional commitment capacity for compute CUDs. The tradeoff between data movement costs and compute placement becomes a strategic decision point in portfolio optimisation.

The Role of Preemptible and Spot VMs in Commitment Strategy

Google's Spot VMs represent a complementary discount mechanism that should be integrated into CUD planning. Spot VMs provide up to 80-90 percent discounts versus on-demand pricing in exchange for interruptibility—Google can terminate them at any time. For batch workloads, data processing pipelines, and non-critical background jobs, Spot VMs offer substantial savings without requiring commitment contracts.

The optimal portfolio structure reserves CUD capacity for mission-critical, always-on workloads (databases, application servers, monitoring infrastructure) and allocates Spot VMs to fault-tolerant batch processing. This hybrid approach—combining CUDs for stable workloads, on-demand for peak spillover, and Spot for batch—achieves optimal cost efficiency. An enterprise might commit to 500 vCPUs via CUD, maintain 200 vCPUs on-demand for flexibility, and run 300+ vCPUs of batch workload on Spot. The combined approach delivers blended costs 40-50 percent below full on-demand pricing.

When negotiating CUDs plus PPA discounts, explicitly include Spot VM pricing in the discussion. Some Google Cloud commercial teams offer discounts on Spot VM pricing when layered with large CUD and PPA commitments. While Spot pricing is typically less negotiable than CUD and PPA rates, enterprise-scale customers may achieve 5-10 percent reductions through bundled negotiations.

Contractual Considerations: Lock-In, Flexibility, and Renegotiation Rights

CUD contracts are legally binding financial commitments, and the terms beyond pricing require careful review. Key contract considerations include lock-in period (one or three years), renewal options, price escalation terms, and renegotiation triggers. Most organisations focus exclusively on discount rate negotiation and overlook structural terms that can provide significant flexibility.

Three-year CUD commitments lock in pricing for the entire contract term, which is attractive when discount rates are high but risky if workload patterns shift dramatically. One-year commitments provide greater flexibility at the cost of lower discount rates. Some enterprises employ a staggered approach: committing 60 percent of capacity to three-year terms for maximum discounts on proven stable workloads, while maintaining 40 percent on one-year renewals for workloads with higher volatility.

Renegotiation rights and mid-contract adjustment provisions are often negotiable at enterprise scale. A $1.5 million three-year commitment might include a clause permitting a 15 percent increase in commitment amount at the original discount rate if business needs grow, and a corresponding reduction to 70 percent of committed capacity if workload declines. These flexibility provisions reduce financial risk without forfeiting discount rates.

Common CUD Planning Mistakes and Mitigation

Four mistakes consistently emerge in CUD planning:

  • Mistake 1: Over-Commitment. Committing to 100 percent of expected usage eliminates flexibility entirely. Mitigation: establish a 70-80 percent target coverage ratio and maintain on-demand spillover capacity for growth and variability.
  • Mistake 2: No Annual Review. Failing to review commitment amounts annually leaves organisations with misaligned capacity as workloads evolve. Mitigation: establish a quarterly CUD review cadence using the CUD Analysis tool to identify optimization opportunities and adjust commitment amounts based on actual utilisation trends.
  • Mistake 3: CUD vs PPA False Choice. Treating CUDs and PPAs as alternatives rather than complementary strategies leaves 15-25 percent savings uncaptured. Mitigation: explicitly negotiate CUD plus PPA layering at annual review and demand separate line-item pricing for both mechanisms.
  • Mistake 4: Narrow Scope. Ignoring service-level CUDs (BigQuery slots, Cloud Storage) while focusing exclusively on compute CUDs misses secondary savings opportunities. Mitigation: conduct a service portfolio audit to identify secondary services with consumption-based pricing and evaluate CUD opportunities across the entire stack.

The most successful organisations execute a comprehensive portfolio review twice annually—once during budget planning cycles and again during Google's Q4 (July-September) negotiation window—using the CUD Analysis tool to validate commitment sizing and identify renegotiation opportunities.

Executive Priorities for CUD Optimisation

Three immediate actions should be prioritised. First, commission an independent assessment of your current CUD portfolio using the CUD Analysis tool. Identify under-commitment and over-commitment scenarios, and model optimal commitment amounts based on historical and forecasted utilisation. Second, negotiate dual CUD plus PPA discounting in your next annual review. If you are not receiving both mechanisms, you are likely leaving 15-25 percent in potential savings unused. Third, establish a quarterly cost governance and commitment review process. CUDs are not fire-and-forget; they require continuous optimization to remain aligned with evolving workloads.

Download our CUD negotiation guide to access template frameworks for commitment sizing, enterprise discount justification, and multi-year deal structuring. Our team has advised 500+ organisations through Google Cloud negotiations and can provide cost optimisation specialists to validate your strategy and negotiate with Google directly.

Stay Informed on Google Cloud Cost Optimisation

CUD pricing, FinOps frameworks, and Google Cloud cost governance evolve continuously. Join our community to receive quarterly updates on cost optimisation strategy, new commitment models, and enterprise discount techniques.