Table of Contents
- The Gap Between Quoted Price and Actual Cost
- Component 1: Licence and Consumption Fees
- Component 2: Infrastructure and Compute Costs
- Component 3: Integration and Implementation
- Component 4: Talent, Training, and Capability Building
- Component 5: Governance and Compliance Costs
- Component 6: Ongoing Operational Costs
- Building Your Enterprise AI TCO Model
The Gap Between Quoted Price and Actual Cost
In our advisory work with enterprise AI procurement, we consistently see the same pattern: a business case built on vendor-quoted licence fees, approved by a board that believes the investment is well-understood, followed by actual costs that are 3–5× the original budget by year two. The additional costs are not surprises to anyone who has built enterprise technology before — but AI has a specific cost structure that many organisations are encountering for the first time.
The licence fee is visible because vendors quote it. The other cost components — infrastructure, integration, talent, governance — are invisible in the vendor's proposal because the vendor has no obligation to quantify them and significant interest in keeping the initial commitment number low. A $500K AI licence commitment looks very different when the board understands the total 3-year TCO is $2.8M.
Component 1: Licence and Consumption Fees
Licence fees are the most visible cost but also the most variable. Enterprise AI pricing follows several models, each with different cost dynamics:
Per-User Subscription Pricing
Microsoft Copilot ($30/user/month), Salesforce Einstein ($50/user/month), and similar per-user AI subscriptions appear straightforward but contain two significant cost drivers: the user expansion problem (initially deployed for power users, gradually expanded to broader populations as business value is demonstrated) and the feature tier escalation problem (initial licences are base tier; advanced features require higher tiers at 40–80% premiums).
Model your per-user costs against a realistic adoption curve — not the maximum theoretical deployment — and include a 30% contingency for tier escalation. A deployment that begins with 500 users at $30/user/month but expands to 3,000 users over 24 months with 40% upgrading to advanced tiers has a very different cost trajectory than the initial business case suggests.
Consumption-Based Pricing
Token-based pricing (OpenAI, Anthropic, Azure OpenAI) and API call pricing are the most difficult to model accurately. Token consumption varies with: prompt engineering complexity, user behaviour, use case type, and the degree to which users adopt advanced features. Initial pilots routinely underestimate production consumption by 2–5×, because pilots involve carefully designed use cases while production involves unguided user behaviour.
Build consumption models from use case sizing — estimate average tokens per interaction, interactions per user per day, and user count — then apply a production scaling factor of at least 3× your pilot consumption rate. Cap consumption contractually through pre-purchased credit blocks with overage pricing limits.
| Pricing Model | Vendors | Key Cost Driver | Common Budget Miss |
|---|---|---|---|
| Per-user subscription | Microsoft, Salesforce, ServiceNow | User adoption growth + tier escalation | 2-3× initial user count by year 2 |
| Token consumption | OpenAI, Anthropic, Azure OpenAI | Production vs. pilot consumption ratio | 3-5× underestimation vs. pilot |
| Platform + compute | AWS Bedrock, Google Vertex AI | GPU instance costs for hosted models | 50-80% underestimated GPU costs |
| Embedded AI uplift | SAP, Oracle, Workday | Edition upgrades + BTP consumption | Integration costs underestimated 5× |
Component 2: Infrastructure and Compute Costs
Infrastructure is the most underestimated AI cost category, and the one that most consistently blows enterprise AI budgets. There are three infrastructure cost layers for enterprise AI:
Model Serving Infrastructure
If you are self-hosting models (open-source LLMs, fine-tuned models, or models from vendors like Mistral or Meta), you need GPU infrastructure for inference. A single A100 GPU costs approximately $2.50–4.00/hour on major cloud providers. A production-grade model serving setup for an enterprise use case typically requires 4–16 GPUs depending on model size, traffic volume, and latency requirements. At $3/hour × 8 GPUs × 8,760 hours/year, you are spending $210K/year on compute alone — before redundancy, development environments, or monitoring infrastructure.
Data Infrastructure
Enterprise AI requires substantial data infrastructure beyond what most organisations have in place: vector databases for retrieval-augmented generation (Pinecone, Weaviate, Chroma), data lakes or feature stores for training data management, real-time data pipelines for AI systems consuming live business data, and enhanced storage for model weights, embeddings, and inference logs. These infrastructure components typically cost $50–300K to build and $80–200K/year to operate, depending on scale.
Network and Egress Costs
AI-heavy architectures generate significant cloud egress costs as data moves between your systems and AI model endpoints. In our analysis, organisations consuming $500K/year in AI API services spend an average of $80–150K in associated egress and data transfer costs — a line item that rarely appears in the initial business case. See our article on cloud egress cost negotiation for strategies to reduce this component.
Component 3: Integration and Implementation
Integration is the largest single cost component in most enterprise AI deployments and the one most consistently underestimated. The vendor's professional services estimate — typically provided as part of the sales process — assumes a greenfield, well-documented environment. Enterprise reality is different.
System Integration
Connecting an AI system to your enterprise data sources — CRM, ERP, HR systems, document repositories — requires data mapping, API development, identity management integration, and data quality remediation. For a typical mid-size enterprise AI deployment connecting 3–5 source systems, integration effort runs 6–18 months at $150–250K per system in professional services cost, or equivalent internal engineering time.
Legacy System Accommodation
AI systems typically assume modern, API-accessible data sources. Your data may live in legacy systems with poor API support, inconsistent data quality, undocumented schemas, or batch-only update cycles. Legacy accommodation adds 40–80% to integration cost and timeline — and is rarely surfaced in vendor proposals that assume your environment matches their reference architecture.
Custom Development and Prompt Engineering
Production AI deployments require significant prompt engineering and custom development work: prompt design and optimisation for each use case, context injection and RAG pipeline development, output formatting and post-processing, edge case handling and safety filtering, and UI/UX development for user interfaces. A production deployment typically involves 3–6 months of development work beyond the vendor's implementation services — at $100–200K in engineering cost.
Component 4: Talent, Training, and Capability Building
Talent is a cost component that most AI business cases underestimate because it manifests gradually and is distributed across departments. The talent cost of enterprise AI has three elements:
Specialist AI Talent
Production AI deployments require at minimum: an ML engineer or AI platform engineer to manage the AI infrastructure and deployment pipeline, a data engineer to manage the data pipelines feeding the AI system, and a prompt engineer or AI product manager to own use case development and optimisation. At current market rates ($180–280K total compensation each), a minimal 3-person team costs $540–840K/year. Most organisations either hire these roles (high cost, competitive market) or pay consulting firms for them (often higher cost, with knowledge transfer risk).
User Training and Change Management
User training for AI tools is more intensive than for traditional software because behaviour change is required — not just feature learning. Effective AI adoption programmes involve: initial training (2–4 hours per user), ongoing coaching for power users, use case development workshops, and change management to overcome resistance from users who fear replacement. For a 1,000-user deployment, budget $150–300K for the first year's training and change management programme.
Management and Governance Overhead
AI deployments require ongoing management attention from business leaders, legal, compliance, and IT security teams — roles that don't appear in vendor proposals. This overhead is real but difficult to quantify; a reasonable estimate is 0.25 FTE per major AI use case across these functions, costing $50–100K per use case annually in management time.
Component 5: Governance and Compliance Costs
Governance costs are the fastest-growing component of enterprise AI TCO, driven by the EU AI Act, sector-specific AI regulation, and the growing body of AI-related litigation. Organisations that fail to account for governance costs in their AI TCO create regulatory exposure that can dwarf the deployment cost itself.
EU AI Act Compliance
For high-risk AI systems deployed in European operations, EU AI Act compliance involves: initial risk assessment and classification ($30–80K), conformity assessment documentation ($50–150K one-time), ongoing technical documentation maintenance ($20–50K/year), human oversight implementation ($30–80K implementation), and post-market monitoring ($30–60K/year). Total EU AI Act compliance cost for a high-risk deployment: $130–370K first year, $50–110K annually thereafter.
Bias Testing and Model Evaluation
Regular bias testing and model performance evaluation are required both by regulation (for high-risk systems) and by good governance practice for any AI making consequential decisions. Establish a testing cadence — quarterly for high-risk systems, semi-annually for others — and budget $50–150K per year for evaluation infrastructure and specialist assessment services.
Component 6: Ongoing Operational Costs
Once deployed, AI systems incur ongoing operational costs that compound over time. Organisations that account for 3-year operational costs in their initial TCO model make fundamentally different architecture and vendor decisions than those who focus on implementation cost alone.
Key ongoing cost categories: model retraining and fine-tuning as the deployment environment evolves ($50–200K/year); monitoring and observability infrastructure for production AI systems ($30–80K/year); security testing including adversarial testing and prompt injection defence ($40–100K/year); vendor relationship management for complex AI contracts ($20–50K/year); and continuous improvement engineering as AI capabilities expand ($100–300K/year).
Building Your Enterprise AI TCO Model
A complete enterprise AI TCO model should project costs across all six components over a 3-year horizon, including: Year 1 implementation costs (typically the most expensive year due to integration and talent acquisition), steady-state Year 2–3 operational costs, and a Year 3+ scaling projection based on realistic adoption curves.
Typical total 3-year TCO for a mid-size enterprise AI deployment (1,000–5,000 users, 3–5 integrated systems, one major use case):
- Licence and consumption: $1.5–3M
- Infrastructure: $600K–1.5M
- Integration and implementation: $800K–2M
- Talent and training: $900K–2M
- Governance and compliance: $300K–700K
- Ongoing operational: $400K–900K
- Total 3-year TCO: $4.5M–10M
The licence fee — the number that appears in the vendor's proposal — represents $1.5–3M of this total. Before presenting an AI investment case to leadership, ensure your model accounts for all six components. For a structured review of your AI procurement approach, see our AI Procurement Advisory service and our AI Procurement Checklist.