Two years ago, I forecast ChatGPT’s seismic impact on procurement – and for once, I wasn’t off by miles. In May 2023, my CPOstrategy article prophesied that generative AI would reshape how procurement teams draft contracts, analyse spend, and navigate sourcing. Early demos – ChatGPT spinning out RFP drafts in seconds – generated headlines, but real-world pilots often remained siloed experiments. The rapid evolution of AI in just two years is truly impressive.
Fast-forward to May 2025: AI has graduated from novelty to necessity. “Agentic” bots now auto-approve low-value purchase orders overnight; private GPTs flag noncompliant clauses across hundreds of supplier contracts in minutes; and leading SaaS suites like SAP Ariba, Oracle Cloud, and Coupa have fused GenAI, reasoning modules, and workflow orchestration into unified procurement workbenches. Yet amidst the buzz, procurement leaders still wrestle with critical questions:
Where does real ROI hide?
Which use cases should you pilot first?
How do you scale safely without stifling innovation?
In part one of this two-part series, we answer those questions through four parts:
Tech drivers: From chips to bots – Why modern AI is faster, cheaper, and more capable than ever, making it ubiquitous.
Deployments and use cases: How procurement teams leverage AI today, from ad hoc copilots to autonomous agents. This means AI adoption is not just a theoretical concept, but it’s happening in real-time at scale, making it a practical and feasible solution for procurement teams.
The four pillars of trust: Governance essentials for alignment, data quality, compliance, and sustainability. This will help you understand how to deploy AI responsibly, ensuring its implementation is secure and in line with ethical standards.
Market map and strategic playbook: My view of a vendor landscape and roadmap to turn pilots into enterprise standards.
By the end of this deep dive, you’ll have a better idea of where to focus – optimising inference compute, piloting multimodal AI proofs-of-concept, and standing up an AI governance council – setting the stage for part two’s practical source-to-pay playbook.
Tech drivers: From chips to bots
Modern AI’s power surge stems from the relentless innovation of hardware, software, and architecture. Let’s unpack the three key drivers.
1.1. The GPU Gold Rush
General-purpose old-school CPUs (Central Processing Units) simply couldn’t handle the voluminous matrix math AI demands. Enter Graphical Processing Units (GPUs) and Huang’s Law – Nvidia CEO Jensen Huang’s observation that GPU performance for AI workloads doubles to triples annually, dwarfing Moore’s Law’s transistor pace. The result? Training or running a GPT-3.5 query in late 2022 cost roughly $20 per million tokens; by mid-2025, it plunged to about $0.07. That’s a 280× collapse – transforming AI from a boardroom curiosity into a line-item purchase for even mid-market procurement teams.
Inference vs compute time: Training (compute time) is the heavy-duty process where models learn from data, long, costly tasks handled by AI Hyperscalers, the 10 or so huge AI companies such as OpenAI. Inference is the moment of truth: each query you run taps into trained models, consuming lighter, on-demand compute. While training enhances capabilities, inference generates real-time ROI and must be budgeted per usage to manage ongoing costs.
Your next move:
Break down your AI projects by supplier and contract compute costs. Understand percentage training versus inference, and work with IT to forecast future costs based on planned models. Where could you renegotiate rates or shift workloads to cheaper models?
1.2. Smarter, leaner, faster
Hardware and software co-design have driven dramatic efficiency gains. Design techniques have improved GPU performance and recued power draw. Yet raw scale remains a competitive edge: hyperscalers still harness thousands of GPU pods for training new reasoning models, chasing performance gains.
Reasoning models and chain-of-thought: Unlike generative LLMs that predict the next token in a flash, reasoning-focused models, like Open AI’s “O-series” (O1, O2, O3, etc.) break problems into sequential steps via chain-of-thought prompting. This means that these models don’t just generate a response but follow a logical sequence of steps, similar to how a human would think through a problem. Each step requires a fresh inference call – often 2–3× more compute per query – and drives up power usage by 40–60% compared to standard text generation. For CPOs, this means advanced “slow-thinking” applications like risk analysis or strategic scenario planning demand higher inference budgets and tighter cost controls.
So what for the CPO? Before rolling out reasoning-heavy AI, work with IT to map high-value use cases to their expected inference costs, and only use reasoning where you need it. Negotiate tiered pricing with providers to keep power and cost in check, ensuring “thinking” AI doesn’t break your bottom line.
1.3. Multimodal Mojo
Text-only bots are so 2021. Today’s AI architectures seamlessly fuse text, images, audio, and video:
Document and image fusion: AI reviews technical drawings and spec sheets, extracting key tolerances and auto-populating contract clauses.
Video monitoring: Live camera feeds at supplier factories feed anomaly detectors – scratches, misalignments – triggering real-time alerts to procurement.
Voice and chat blends: Conversational interfaces integrate voice commands with text logs for richer audit trails.
From text-to-image to text-to-video: Developers discovered that they could combine speech to text and image to text to readily create video models like DALL-E and Stable Diffusion with relativity limited effort. That’s the AI scaling laws at work. Once models proved they could generate high-resolution images, increasing depth and data enabled coherent frame-by-frame video synthesis “for free” – supercharging use cases like automated site inspections via short clips.
Case study: A global consumer goods enterprise deployed a multimodal AI solution across six factories. The system ingested HD video streams, identified quality defects within 60 seconds, and automatically initiated reorders, reducing recall-related costs by 35% and accelerating supplier response times by 50%.
Your next move:
Identify one high-volume process – e.g., invoice checking or RFP drafting – and evaluate a multimodal tool’s ability to ingest non-text data (images, PDFs, audio notes).
1.4. Three phases of AI
Building on Nvidia’s framework, AI’s evolution spans three overlapping eras, each with its own unique characteristics and applications:
Perception (pre-2025): Recognising patterns-speech, images, sensor data. Early voice assistants, optical character recognition, and basic anomaly detection fall here.
Cognition (2023–now): Generative models (GPT-4, Claude 3) and reasoning pipelines that write, summarise, and solve complex problems via chain-of-thought approaches. This is also known as “Agentic AI”
Action (future): The future of AI will feature the development of agentic AI and robotics, which will integrate perception and cognition into tangible actions. This means we can expect warehouse robots, self-driving delivery fleets, and automated transactional agents placing orders – all powered by AI and robotics – making procurement processes more efficient and autonomous.
We’re only halfway through the chessboard. The upcoming advancements – AI agents autonomously negotiating multi-round contracts influenced by suppliers aiming to optimise sales – cannot be overlooked.
Deployments and use cases
Procurement teams are weaving AI into every process – drafting RFIs one minute and handling mission-critical risk management the next. A strategic way to visualise this is across Use Case Contexts (what AI delivers) and Deployment Modes (where it lives).

2.1. Use case contexts
A. Desktop assist
Public-data copilots: ChatGPT, Bard, or enterprise Copilots for ad hoc market research, price tracking, or drafting RFP skeletons – without touching confidential IP.
Example: Anna generates a five-minute rare-earth metals briefing, cutting prep from two hours to ten minutes.
B. Private-data insights
Proprietary GPTs: Fine-tuned on internal contracts, purchase orders, and supplier scorecards – enabling clause validation, spend forecasting, and compliance checks.
Example: A private GPT flags nonstandard liability clauses in 200 supplier agreements, slashing legal review time by 80%. Contoso Case Study
C. Autonomous agents
Event or time-driven bots: Auto-source quotes, draft RFI responses, or approve low-risk POs based on set thresholds.
Example (Global Electronics): An invoice-triage agent processes sub-$30K invoices overnight, reducing manual touchpoints by 70% and cutting errors by 60%. Fast-Electronics Case Study
D. SaaS fusion workbenches
Integrated procurement suites: The latest cloud platforms (SAP Ariba’s Joule, Oracle Procurement Cloud, Coupa Nexus) are blending GenAI drafting, reasoning modules, and workflow orchestration into a single interface – no tool-hopping required.
Example: A consumer-goods CPO uses a unified platform to draft contracts, risk analyses, and trigger approval workflows – all within a single pane of glass, reducing context-switching by 50%.
2.2. Deployment modes
Desktop copilots: Ideal for individual tasks – quick drafting, small-scale insights, and experimentation.
ERP embedding: Native AI features in SAP, Oracle, or Microsoft Dynamics deliver in-context recommendations on sourcing screens and purchase orders.
S2P applications: ProcureTech100 innovators bake AI into each module – sourcing, contracting, invoicing, supplier risk – turning AI into the engine rather than an add-on.
Your next move:
Context-mode mapping: Chart your top five procurement processes in a 2×2 grid to pinpoint high-impact AI integration points.
Pilot framework: For each selected use case, define success metrics (cycle-time reduction, error rate decline, compliance coverage) and assemble cross-functional teams to own pilots.
The four pillars of trust
Deploying AI without governance is like unquestioningly sailing stormy seas. Secure AI adoption with these four pillars:
A. Alignment and safety
AI alignment creates the moral compass for AI to ensure models act according to human values, even in edge cases. Vendor philosophies differ:
Anthropic (Claude): Safety-first, red-team tested.
OpenAI (GPT-4): Balanced guardrails.
xAI (Grok): Free expression, higher risk.
Meta (LLaMA): Open-source, relies on community oversight.
Case study (FinServ): A Fortune 500 bank tested three LLMs for supplier risk summaries. Only the Claude model passed compliance checks; others hallucinated regulations.
Your next move:
Build an AI Risk Matrix ranking cross business use cases by impact and misalignment likelihood. Manage any risk with IT & the business.
B. Data quality
AI’s Achilles’ heel is dirty data – siloed, unstructured, or outdated. Combat this with AI-driven MDM:
Informatica & SAP MDG: Self-healing workflows correct supplier duplicates and GL mismatches.
Tealbook and Creactives: Procurement MDM specialists also enrich records with external data, such as certifications and financial health.
Case Study (CPG): A global CPG slashed master-data errors by 90%, halved onboarding time, and improved analytics accuracy by 40% using AI-driven MDM. Tealbook
Your next move:
Pilot AI-driven MDM in one spend category; measure error reduction and time saved.
C. Regulatory and geopolitics
AI is now a strategic asset, and US–China–Taiwan tensions strain the chip supply chain. The US leads the way – backed by scale, investment, and initiatives like the $500 billion “Stargate” infrastructure plan, while China’s government support keeps it close behind.
AI operating across borders faces a patchwork of rules:
EU AI Act: Risk-based obligations and conformity assessments.
US NIST framework and export controls: Voluntary guidelines and chip export limits.
China controls: Data localisation, censorship, security reviews.
Case study (Pharma): A pharma leader mapped data-residency across 12 countries, shifting inference to EU private clouds – avoiding $5M fines. Pharma Compliance
Your next move:
Create a compliance register linking each AI service to its regulatory requirements and data zones.
D. Sustainability
AI’s power demands rivals entire industries. Data centres consume 1.5% of global electricity – projected to 3% by 2030. At it’s peak, Bitcoin consumed 173 TWh – equal to Poland’s electricity use, Switzerland’s water use (and matching the Netherlands’ IT waste), and AI will be way beyond Bitcoin.
Carbon mix: Prefer vendors with renewable-heavy grids.
Latency vs green: Edge inference cuts transfer but may use dirtier local power.
The water consumption from AI is also astronomical, at a time when water will become one of the limiting economic factors in the next 10 years.
Case study (Retail): A retailer migrated inference from the US West Coast (60% fossil) to Northern Europe (80% renewables), slashing CO₂ by 40% with no latency hit. GreenTail
Your next move:
Embed energy-intensity KPIs into AI contracts – aim for 20% yearly kWh and CO₂ reduction.
Market map and strategic playbook
Navigating the crowded AI vendor landscape can feel overwhelming.
A simple speed vs safety matrix helps frame choices:
Manage GenAI as a strategic category
Work with IT to ensure that instead of chasing every shiny new model, adopt a tiered approach first, pilot safety-first vendors for mission-critical tasks like compliance checks or contract reviews. Once confidence grows, layer in speed-optimised models for low-risk, high-volume processes think invoice triage or market research.
Then, build a vendor scorecard to compare providers on cost-per-inference, alignment philosophy, support SLAs, and innovation roadmap. Regularly score and rank vendors to guide renewals and expansions.
Finally, establish a governance council – a cross-functional group from procurement, IT, legal, and sustainability. Meet quarterly to review vendor performance, emerging risks, and new capabilities, ensuring your AI portfolio stays aligned with business goals.
Your next move:
Schedule your first quarterly vendor review and finalise your vendor scorecard template.
Strategic moves:
Pilot pairing: Start with safety-first for core tasks, then explore speed-focused models in low-risk areas.
Vendor scorecard: Evaluate cost-per-inference, guardrails, SLAS, support, and innovation pipeline.
Governance council: The cross-functional team (Procurement, IT, legal, and sustainability) meets monthly to oversee the AI portfolio.
Conclusion and looking ahead to part two
Procurement AI has matured from flashy demos to line-item ROI two years after the ChatGPT whirlwind.
You’ve explored why GPU costs plummeted, witnessed real-world pilots cutting cycles by up to 80%, and learned how to govern AI with trust through alignment, data hygiene, compliance, and green metrics.
Next month in part two – AI-enabled source-to-pay in practice – we’ll unpack how AI is weaving into every S2P tool, explore the rise of agentic bots in procurement, and map out a five-to 10-year roadmap. You’ll see vendor-agnostic case studies showing how smart assistants and autonomous agents will slash cycle times, cut costs, and drive innovation across our deeply structured, expertise-rich function.
Stay tuned: procurement’s AI-powered future is not coming – it’s already here.