Executive Summary

Five key findings on cloud provider AI bundling and negotiation strategy:

  • All three hyperscalers now use AI commitments as the primary mechanism for inflating cloud consumption minimums at renewal. AWS embeds AI spend (Bedrock, SageMaker) into EDP commitments. Microsoft includes Azure OpenAI Service consumption within the Microsoft Azure Consumption Commitment (MACC). Google counts Vertex AI consumption toward CUD spend thresholds.
  • Enterprises that negotiate AI pricing independently of their cloud agreement achieve 25–40% better unit economics on AI services. This improvement is driven by the ability to compare pricing across providers, leverage multi-provider evaluation as a negotiation tactic, and avoid bundled pricing that obscures true AI costs.
  • Provider AI pricing is more negotiable than any other cloud service category. Because adoption targets are aggressive and market share is unsettled, providers are willing to improve pricing to gain committed AI consumption. Token-level pricing, committed capacity rates, and model availability are all negotiable.
  • The cross-subsidy between cloud infrastructure discounts and AI commitment premiums is the defining commercial tactic of 2025–2026 cloud renewals. Providers offer infrastructure discount improvements funded by AI pricing that carries higher margins than the enterprise realises.
  • Multi-provider AI evaluation is the most effective negotiation lever, yet fewer than 15% of enterprises deploy it. Model portability (Anthropic Claude and Meta Llama are available on all three platforms) enables genuine competitive pricing data. This translates directly into better AI pricing and protective contract terms.

How Cloud Providers Bundle AI into Enterprise Agreements

The Bundling Architecture

Each provider has adapted its existing enterprise commitment vehicle to accommodate AI services. Microsoft includes Azure OpenAI Service consumption within the Microsoft Azure Consumption Commitment (MACC), meaning that AI spend counts toward — and inflates — the same commitment floor as compute, storage, and platform services. AWS embeds AI spend (Bedrock, SageMaker) into EDP commitments. Google counts Vertex AI consumption toward CUD spend thresholds.

The commercial consequence is that AI spend is not treated as a new category requiring independent evaluation — it is absorbed into existing commitment frameworks where the enterprise has already established pricing expectations, discount tiers, and consumption patterns.

Why This Matters Now

The urgency is driven by timing. Enterprise AI adoption is moving from experimentation to production deployment, which means AI consumption is scaling from insignificant to material within a single contract cycle. The decisions being made in 2025 and 2026 cloud renewals will establish the commercial framework for AI spending for years.

Key Insight

The cloud providers are not doing anything technically deceptive — AI services are legitimately part of their cloud platforms. The issue is that treating AI as "just another cloud service" within existing commitment frameworks prevents enterprises from evaluating AI pricing on its own merits, comparing it across providers, or negotiating it with the leverage that comes from genuine alternatives.

The AI Commitment Architecture: Bedrock, Azure OpenAI & Vertex AI

Each hyperscaler has built a distinct AI services platform with its own pricing model, commitment structure, and bundling mechanism.

AWS Approach: EDP Inclusion

AWS embeds AI spend into EDP commitments. Bedrock and SageMaker consumption counts toward the total spend threshold that determines compute and storage discount tiers. Bundling mechanism: EDP spend threshold inclusion. AI consumption inflates the commitment floor, which in turn justifies the infrastructure discount tier — creating a circular dependency where removing AI spend would threaten the infrastructure discount.

Microsoft Approach: MACC Integration

Microsoft includes Azure OpenAI consumption within MACC commitments. This is the most aggressive bundling of the three providers — Azure OpenAI spend is treated identically to compute or storage consumption for MACC attainment purposes. Microsoft additionally ties Azure OpenAI access and quota allocation to the overall commercial relationship. Bundling mechanism: MACC inclusion plus quota allocation leverage.

Google Approach: CUD Alignment

Google counts Vertex AI consumption toward spend-based committed use discounts. While Google's bundling is currently less aggressive than Microsoft's MACC integration, the trajectory is toward tighter coupling as Vertex AI consumption scales. Google is currently the most willing to discuss standalone AI pricing, partly because their AI market share position incentivises competitive pricing to gain adoption.

Model Access as a Negotiation Variable

A critical difference between providers: all three now offer access to many of the same foundation models. Anthropic Claude is available on both AWS Bedrock and Google Vertex AI; Meta Llama is available on all three. This model portability creates a genuine competitive dynamic that didn't exist when AI infrastructure was provider-proprietary.

See how enterprises separate AI and cloud commitments

Our team has helped Fortune 500 companies achieve 25–40% better AI unit economics by separating AI from their cloud agreement.

8 Bundling Tactics That Obscure True AI Costs

1. Commitment Floor Inflation

The provider proposes a higher overall cloud commitment (MACC/EDP/CUD) at renewal, presenting the increase as necessary to maintain current discount tiers. The incremental commitment is primarily driven by projected AI consumption — but this isn't explicitly separated.

2. Cross-Subsidy Packaging

The provider offers enhanced discounts on mature services (compute, storage, networking) in exchange for large AI commitments. The infrastructure discount improvement is real — but it's funded by AI pricing that carries higher margins than the enterprise realises.

3. Blended Unit Rate Obfuscation

The provider presents a "blended rate" across all cloud services that includes AI consumption. This blended rate appears competitive when viewed in aggregate, but disguises significant variation in pricing quality across service categories.

4. Provisioned Capacity Lock-In

For AI workloads requiring predictable throughput, providers offer Provisioned Throughput (AWS) or PTUs (Azure) at committed rates, typically 1–3 year terms with limited flexibility.

5. Model-Specific Quota Gating

Providers control access to high-demand models through quota allocation systems. Quota availability is influenced by the overall commercial relationship.

6. Adjacent Service Bundling

AI inference pricing is bundled with adjacent services (vector databases, embedding generation, fine-tuning compute) as a "complete AI platform" offer.

7. Consumption Credit Fungibility

Providers offer "AI credits" or "innovation funds" that are structured as pre-committed consumption counting toward the overall agreement commitment.

8. Renewal Timing Compression

When AI is bundled into the cloud agreement, the provider argues that AI pricing must be negotiated within the cloud renewal timeline, preventing independent evaluation.

Download: Cloud AI Commitment Negotiation Playbook

How to structure your cloud agreement to separate AI pricing, protect infrastructure discounts, and benchmark AI spend independently.

Pricing Mechanics: How AI Commitments Inflate Cloud Minimums

The Commitment Inflation Cycle

Consider a typical enterprise cloud renewal scenario. The current agreement has $8M annual committed spend with an 18% average infrastructure discount. At renewal, the provider projects that AI consumption will add $3–4M annually. Rather than pricing AI separately, the provider proposes an $12M overall commitment, presenting the increase as necessary to maintain the infrastructure discount tier. Bundling creates the appearance of a better overall deal by improving infrastructure discounts, while embedding AI at pricing that would not survive independent scrutiny.

Token Economics and the Opacity Problem

AI pricing is inherently more complex than infrastructure pricing because it is model-dependent, use-case-dependent, and rapidly evolving. A per-token price for GPT-4o is not comparable to a per-token price for Claude 3.5 Sonnet without normalising for output quality, latency, and task-specific performance. This opacity is not accidental.

The FinOps Gap

Most enterprise FinOps practices are designed around infrastructure optimisation — rightsizing compute, managing reserved instances, eliminating idle resources. AI FinOps requires a fundamentally different approach: tracking cost-per-inference, comparing model economics for equivalent tasks, managing prompt efficiency, and evaluating model routing strategies.

Negotiation Strategy for Standalone AI Pricing

1. Establish an Independent AI Consumption Baseline

Before entering any negotiation, build a detailed model of your AI consumption: which models, what volumes (tokens/requests), what latency requirements, what growth trajectory. This baseline should be provider-agnostic — expressed in terms of workload requirements, not provider-specific units.

2. Request Itemised AI Pricing Separate from Infrastructure

Formally request that all AI services be priced as a separate line item, independent of the cloud infrastructure commitment. Be explicit: "We require independent pricing schedules for AI services and infrastructure services, with separate discount frameworks for each."

3. Run Parallel Multi-Provider AI Evaluations

Evaluate the same workloads across at least two providers. Because Anthropic Claude is available on both AWS Bedrock and Google Vertex AI, and Meta Llama is available on all three platforms, you can run genuine like-for-like pricing comparisons for identical models on different platforms.

4. Negotiate AI and Cloud on Separate Timelines

Decouple the AI commitment timeline from the cloud renewal timeline. Your cloud infrastructure agreement can be renewed on schedule while AI pricing is negotiated independently. The provider will argue this isn't possible; it is.

5. Demand Consumption Flexibility and Model Portability Clauses

AI workloads evolve rapidly. Negotiate clauses that allow you to shift consumption between models without penalty, adjust committed volumes with reasonable notice periods, and access new models at the same or better pricing terms.

6. Benchmark Against Direct API Pricing

For models available via direct API access, compare provider-hosted pricing against direct API pricing for the same models. The delta between direct API pricing and cloud-hosted pricing represents the platform premium — which should be justified by measurable additional value.

Maintaining Cloud Volume Discounts While Unbundling AI

Understanding the Provider's Position

The provider's argument is: "Unbundling AI will reduce your overall commitment, which means your infrastructure discount drops from Tier 3 to Tier 2." Your counter: "Our infrastructure spend independently supports Tier 3 pricing. AI spend should be priced on its own merits — and we're evaluating it across providers."

Five Strategies for Discount Preservation

1. Infrastructure-Only Commitment Modelling

Build a forward-looking infrastructure spend model that excludes AI. Demonstrate that infrastructure consumption alone supports the discount tier.

2. Parallel Agreement Structures

Propose two parallel commercial agreements: an infrastructure commitment covering compute, storage, and platform services, and a separate AI services agreement.

3. Tiered Commitment with AI Carve-Out

If parallel agreements aren't achievable, negotiate a single agreement with explicit AI carve-out provisions: AI spend is included in the total for commitment attainment, but AI pricing is benchmarked independently.

4. Commitment Floor Protection Clauses

Negotiate clauses that protect your infrastructure discount tier even if total spend fluctuates due to AI consumption variability.

5. Strategic Concession Trading

Offer the provider something they value — longer term, earlier payment, migration commitment — in exchange for accommodating standalone AI pricing.

Recommendations: 7 Priority Actions

  1. Demand itemised AI pricing at the start of every cloud renewal — make it a non-negotiable requirement from the first engagement.
  2. Build an AI consumption baseline before negotiating — map current and projected AI consumption in provider-agnostic terms.
  3. Run multi-provider AI evaluations for at least two key workloads — leverage model portability to produce genuine competitive pricing data.
  4. Separate AI and infrastructure renewal timelines — do not allow AI commitments to be forced into the cloud infrastructure renewal timeline.
  5. Develop AI-specific FinOps capabilities — invest in the capability to track AI cost-per-inference and compare model economics.
  6. Negotiate model flexibility and downgrade provisions — avoid long-term commitments to specific model versions.
  7. Engage independent advisory for AI-inclusive cloud renewals — AI pricing requires specialised expertise that most internal procurement teams do not yet have.

How Redress Can Help

Redress Compliance's Cloud & AI Practice provides independent advisory services for enterprises navigating AI-inclusive cloud agreements. We operate with zero provider affiliations, no reseller agreements, and no referral fees. Services include:

  • AI Cost Benchmarking — independent pricing benchmarks across all three hyperscalers for every major foundation model
  • Cloud Agreement Restructuring — designs commercial structures that separate AI and infrastructure commitments while preserving cloud volume discounts
  • AI FinOps Advisory — build AI-specific cost governance capabilities
  • Renewal Negotiation Support — shadow advisory or active negotiation support for cloud renewals
  • Multi-Provider AI Evaluation — structured evaluations across Bedrock, Azure OpenAI, and Vertex AI
  • Contract & Term Analysis — detailed review of AI-specific terms: model access provisions, quota guarantees, consumption flexibility clauses

Want help separating AI pricing from your cloud deal?

Our Cloud & AI Practice works exclusively with enterprise buyers to structure AI and cloud agreements independently, preserve infrastructure discounts, and negotiate better unit economics.