Back to Blog

AI Agent ROI: The KPI Framework That Convinces CFOs in 90 Days

Henri Jung, Co-founder at Superkind
Henri Jung

Co-founder at Superkind

A precision dial gauge representing measurement of AI agent ROI

The agent went live in February. By April the operations team called it a success - it handled 4,200 tickets, cut response time, freed staff for harder cases. The CFO sat through the demo, said it looked impressive, and then asked: “What was the cost per ticket before, what is it now, and what hidden costs am I missing?” The room went quiet. Nobody had measured the baseline.

That story is the rule, not the exception. McKinsey’s State of AI 2025 found that 88 percent of organisations now use AI in at least one function, but only 6 percent qualify as “AI high performers” with EBIT impact above 5 percent1. IBM research found that 79 percent of organisations see productivity gains from AI, but only 29 percent can measure ROI confidently14. The productivity is real - the measurement is missing.

This guide is for the CFO, controller, or Geschaeftsfuehrer who has approved an AI agent pilot and now needs to know whether it actually pays off. No vendor pitch. No vanity metrics. Just the four-tier KPI framework, the hidden costs to surface, the 90-day measurement plan, and the three-panel template you take into the next finance review.

TL;DR

Most AI ROI numbers fail the CFO test for three reasons: no baseline, hidden costs, vanity metrics. Fix all three and ROI becomes defensible.

The four-tier KPI framework: Operational (handle time, containment), Quality (CSAT, error rate), Financial (cost per task, hours freed), Strategic (capability, optionality). All four show up in a CFO report.

True total cost of ownership runs 1.4 to 1.7x the build quote. Maintenance is 15-25 percent of build per year. Engineer time for operations adds EUR 3,000-6,000 per month.

90 days is enough to prove or kill a use case. By month three you should see the ROI slope, even if payback is later. If the slope is flat, re-scope or stop.

The CFO presentation is one slide, three panels: baseline vs current vs target, with the financial bridge showing where the value comes from. Anything more is noise.

The 6 Percent Problem

The headline numbers on AI ROI in 2026 contradict each other. Adoption is at record levels, vendors quote breakthrough returns, but at the enterprise level, real bottom-line impact stays narrow. The gap between the two is the 6 percent problem.

  • Adoption is high - 88 percent of organisations now use AI in at least one business function, up from 78 percent in 20241
  • Real EBIT impact is rare - 39 percent of respondents attribute any level of EBIT impact to AI; most of those say less than 5 percent1
  • The high-performer threshold - Only about 6 percent of organisations qualify as “AI high performers”, attributing more than 5 percent EBIT impact to AI1
  • Productivity exceeds measurement - 79 percent of organisations report productivity gains from AI, only 29 percent can measure ROI confidently14
  • Pilots stall before production - Roughly two-thirds of organisations remain in experiment or pilot mode1,21. 88 percent of agent pilots never reach production21
  • Mittelstand is not behind, the world is - Bitkom reports 41 percent of German firms actively use AI, with 62 percent experimenting and 23 percent scaling agents13. The gap between adoption and impact is structural, not regional

Key Data Point

Global AI spend is projected above EUR 2 trillion in 202622. The 6 percent of organisations that translate that spend into real EBIT impact will compound their advantage over the next three years. The other 94 percent will face a CFO who stops approving AI budget.

The CFO’s question is not whether AI works - it is whether your AI works in your company. Generic adoption stats do not answer that. Specific KPIs measured against baselines do.

Metric2025-2026 RealitySource
AI use in at least 1 function88% of organisationsMcKinsey 20251
Any EBIT impact reported39% (mostly <5%)McKinsey 20251
AI high performers (>5% EBIT)~6% of organisationsMcKinsey 20251
Productivity gains reported79% of organisationsIBM via Larridin 202614
ROI measured confidently29% of organisationsIBM via Larridin 202614
Agent pilots reaching production~12%Anaconda/Forrester 202621

Why Most Mittelstand AI ROI Numbers Are Wrong

When a CFO challenges an AI ROI claim, the failure usually traces to one of three patterns. Spotting them in your own numbers before the finance review is the cheapest fix in this article.

1. The baseline was never measured

  • What goes wrong - The team launches the agent without measuring the prior process state. After 90 days they cannot say whether the new cost per ticket is better or worse - because the prior cost per ticket was never quantified
  • Why it happens - Pre-launch energy is spent on the build, not the measurement. “We will figure out the metrics during the pilot” is the most common phrase that ends in a failed CFO review
  • Fix - Spend the first two weeks of any AI project measuring the current state. Without a baseline, ROI is impossible to defend
  • Practical baseline list - Volume, cycle time, error rate, cost per unit, FTE-equivalent hours, customer satisfaction, escalation rate

2. Hidden costs were left out

  • What goes wrong - The build quote covers the agent. The ROI calculation uses the build quote. The actual cost includes maintenance, monitoring, retraining, vendor migrations, and engineer time, none of which were in the quote
  • The 1.4-1.7x rule - Real total cost of ownership lands 40 to 70 percent above the headline build cost9
  • Maintenance reality - Annual maintenance runs 15 to 25 percent of the initial build cost, covering prompt updates, model upgrades, and integration upkeep9
  • Engineer-time burn - Production agents need 20-30 percent of a senior engineer’s time, roughly EUR 3,000-6,000 per month at German rates9
  • Token economics traps - Cheaper models often need longer prompts, more retries, and extra human review. Output tokens cost 3-10x more than input tokens. Reasoning tokens add silent overhead11

3. Vanity metrics replaced business metrics

  • What goes wrong - The dashboard shows “15,000 prompts processed” or “agent uptime 99.9 percent”. None of those translate to euros
  • The trap - Operational metrics are easy to collect; business metrics need work. The team picks what is convenient instead of what matters
  • Fix - For every operational metric, define the corresponding business metric. “Prompts processed” becomes “tasks completed” becomes “cost per task” becomes “EUR savings vs baseline”

Vanity Metrics vs Business Metrics

Vanity (avoid)

  • Prompts processed - volume without outcome
  • Agent uptime - infrastructure metric, not value
  • Tokens consumed - cost driver, not benefit
  • Conversations started - engagement, not resolution
  • Average response speed - latency without context

Business (use)

  • Cost per resolved task - the headline finance number
  • Cycle time reduction - days/hours from start to done
  • FTE-hours freed - hours redirected to higher-value work
  • Containment / completion rate - share fully handled by agent
  • CSAT or quality score - did outcome quality hold?

The CFO Test

If your AI dashboard does not include a euro figure within three clicks, it is built for the IT team, not the CFO. Every dashboard that survives a finance review has a single headline metric in euros, with the bridge to baseline visible directly underneath.

The 4-Tier KPI Framework

A defensible AI agent ROI report has four tiers. Each tier answers a different stakeholder question. Skip a tier and the picture collapses under finance scrutiny.

Tier 1: Operational metrics (how does the agent perform?)

  • Containment / completion rate - Share of interactions the agent handles end-to-end. Target: 60-80% for focused use cases
  • Average handle / cycle time - Time from start to resolution. Compare against human baseline at the same scope
  • Throughput - Volume processed per unit of time. Useful when comparing capacity, not cost
  • Escalation rate - Share of interactions handed off to humans. Lower is not always better - too low can mean the agent is overreaching
  • Latency / response time - Critical for voice and customer-facing agents. For back-office agents, secondary

Tier 2: Quality metrics (is the outcome good?)

  • Resolution rate - Share of interactions where the customer outcome was actually achieved. Different from containment
  • Error or hallucination rate - Frequency of factually wrong or off-policy outputs. Track via human review on a sample
  • CSAT / quality score - Customer-facing agents need CSAT. Internal agents need quality review by domain experts
  • Compliance / audit pass rate - Share of agent actions that pass compliance review. Critical for regulated workflows
  • Rework rate - Share of agent outputs that needed correction by a human. The hidden cost number

Tier 3: Financial metrics (what does it cost and save?)

  • Cost per resolved task - The headline finance KPI. Includes LLM cost, infrastructure, and allocated maintenance
  • FTE-equivalent hours freed - Hours per week redirected from agent-handled work to higher-value tasks. Convert to EUR at fully loaded labour cost
  • Total cost of ownership - Build + maintenance + operations + retraining over a defined period (typically 12 or 24 months)
  • Payback period - Months until cumulative savings exceed cumulative costs. Target: 4-9 months for focused use cases
  • Cost avoidance - EUR value of errors, escalations, or compliance issues prevented. Audit-trailed against historical event cost

Tier 4: Strategic metrics (does this build options?)

  • Capability gain - New capabilities unlocked (24/7 coverage, multi-language, after-hours service). Hard to monetise, real to customers
  • Workforce reallocation - Share of FTE time moved from routine to strategic work. The competitive moat number
  • Customer retention impact - Churn rate change attributable to faster service or 24/7 coverage
  • Competitive optionality - Speed at which you can deploy the next agent because the first one created the foundation
  • Compliance posture - Audit trail completeness, EU AI Act readiness, DSGVO documentation - all reduce future risk cost
TierHeadline KPIStakeholderUpdate Frequency
1. OperationalContainment rateOperations leadDaily
2. QualityResolution rate / CSATService / quality leadWeekly
3. FinancialCost per resolved taskCFO / controllerMonthly
4. StrategicFTE-hours reallocatedGeschaeftsfuehrer / boardQuarterly

“AI does not follow one cost curve, and it does not produce one uniform type of value. CFOs need to account for that if they want a complete picture of what AI is really delivering.”

- Twisha Sharma, Senior Principal Research at Gartner25

Build the ROI report your CFO actually trusts

Book a 30-minute call. We will sketch the four-tier framework against your live agent or planned pilot.

Book a Demo →
Ascending precision weights representing tiered KPIs from operational to strategic

The Hidden Costs CFOs Will Ask About

The first question in any honest CFO review is “what is missing from this number?”. Six cost categories are routinely left out of AI agent ROI calculations. Get ahead of all six before the finance meeting.

1. Maintenance and prompt iteration

  • What it covers - Prompt updates, regression testing, edge case handling, retraining when business processes change
  • Cost rule of thumb - 15-25 percent of initial build cost, per year9
  • Mittelstand reality - Higher than enterprise because Mittelstand workflows tend to evolve continuously rather than in big-bang releases

2. Model and infrastructure cost drift

  • What it covers - LLM token costs, vector DB hosting, telephony for voice agents, observability tooling
  • The token economics trap - Output tokens cost 3-10x input tokens. Reasoning models add silent overhead. Context window inflation as the agent matures11
  • Forecast assumption - 12-month flat baseline, 24-month plus-30-percent stress test

3. Engineer time for operations

  • What it covers - Monitoring, incident response, version upgrades, vendor coordination
  • Allocation rule - 20-30 percent of one senior engineer per production agent9
  • EUR translation - Roughly EUR 3,000-6,000 per month at fully loaded German engineering rates

4. Human review and quality assurance

  • What it covers - Sample-based human review of agent outputs, quality scoring, feedback loop maintenance
  • Why it shows up - Production agents need ongoing QA. Skipping it is the fastest way to silent quality drift
  • Allocation rule - 5-10 percent of an SME-level reviewer per active agent in regulated workflows

5. Vendor migration and lock-in cost

  • What it covers - Cost of switching LLM providers, prompt re-engineering when models change, integration rework
  • Hidden trigger - Models get deprecated. Vendor pricing changes. Your prompts work less well on the next model
  • Mitigation - Architect for portability (MCP-based tooling, abstraction layers). Re-test on alternative models quarterly

6. Compliance and audit overhead

  • What it covers - DSFA preparation, AI inventory maintenance, audit trail review, EU AI Act conformity work
  • Mittelstand reality - Often handled by an external DSB or law firm at billable rates
  • Cost expectation - EUR 5,000-15,000 per agent for initial DSFA, EUR 1,000-3,000 per quarter for ongoing review
Hidden CostAnnual Range (EUR)Where to Document
Maintenance & prompt iteration15-25% of build costOperating budget
Model & infrastructureEUR 3k-30k+Direct OPEX
Engineer operations timeEUR 36k-72kAllocated salary cost
Human QAEUR 5k-25kAllocated salary cost
Vendor migration reserve10-15% of build costRisk reserve
Compliance & auditEUR 9k-25kDirect OPEX

The 1.4-1.7x Rule

Multiply your build cost by 1.4 (light maintenance scenario) to 1.7 (heavy ops/compliance scenario) to get true total cost of ownership for the first year. If your ROI still works at 1.7x, the project is real. If it only works at 1.0x, it is a vendor pitch in disguise.

The 90-Day Measurement Plan

ROI measurement does not start at launch - it starts before week one. The plan below maps to a typical 90-day pilot. By month three you should have a CFO-grade ROI report or a clear signal to kill the use case.

Phase 1: Baseline and instrument (Weeks 1-3)

  1. Week 1: Pre-launch baseline - Measure the current state for every Tier 1-3 KPI you plan to track. Volume, cycle time, cost per task, error rate, FTE-hours, CSAT. Without this, no later ROI claim is defensible
  2. Week 2: Cost forecast with hidden costs - Build the 12-month TCO forecast at 1.0x, 1.4x, and 1.7x scenarios. Document every cost category. Get sign-off from controller before launch
  3. Week 3: Define success and kill criteria - Specific, numerical thresholds for “continue”, “re-scope”, and “stop” decisions at week 12. Without kill criteria, sunk cost takes over and the project lingers

Phase 2: Live measurement (Weeks 4-9)

  1. Weeks 4-5: Soft launch with shadow comparison - Agent runs in parallel with the existing process. KPIs measured for both. Gap to baseline becomes the working ROI signal
  2. Weeks 6-7: Limited live - Route 10-30 percent of in-scope work to the agent. Daily KPI review. Anomalies flagged for human review
  3. Week 8: First financial pulse - Run the cost-per-task math against current volume. Compare to baseline. Update the TCO model with actuals
  4. Week 9: Mid-pilot review - Decision point. If KPIs are trending toward the success threshold, scale to 50-80 percent. If flat, re-scope the use case. If declining, kill

Phase 3: ROI report and CFO presentation (Weeks 10-12)

  1. Week 10: Full rollout (if continuing) - Scale to full in-scope volume. Continue daily Tier 1, weekly Tier 2, monthly Tier 3 cadence
  2. Week 11: ROI calculation and stress tests - Run the financial model at 1.0x, 1.4x, 1.7x cost scenarios. Compute payback at each. If payback exceeds 12 months at 1.7x, escalate to leadership
  3. Week 12: CFO report and decision review - Three-panel one-slide summary (covered in the next section). Decision: continue, expand, sunset

90-Day ROI Readiness Checklist

  • Pre-launch baseline measured for all Tier 1-3 KPIs
  • TCO forecast modelled at 1.0x, 1.4x, 1.7x scenarios
  • Kill criteria defined in writing before launch
  • Named “agent owner” with budget authority and target outcome
  • Daily/weekly/monthly KPI cadence operating from week 4
  • Human review sample (5-10 percent of outputs) running weekly
  • Mid-pilot decision documented at week 9
  • CFO three-panel report drafted by week 11

What success looks like at 90 days

  • Tier 1 (Operational) - Containment 60-80 percent for focused use cases. Cycle time down 30-50 percent vs baseline
  • Tier 2 (Quality) - CSAT or quality score at parity with human baseline or better. Error rate at or below baseline
  • Tier 3 (Financial) - Cost per task down 40-70 percent vs baseline at 1.4x TCO. Payback projection 4-9 months
  • Tier 4 (Strategic) - 30-50 percent of FTE time on the targeted workflow reallocated to higher-value tasks

How to Present to the CFO: The Three-Panel One-Slide Template

CFOs do not read 40-slide AI ROI decks. They read one slide, three panels, with the financial bridge from baseline to current state visible at a glance. Build that slide first; everything else is appendix.

Panel 1: The headline number

  • One metric in EUR - Annualised cost saving or capacity created at current run rate. No percentages without absolute numbers next to them
  • Confidence band - Best-case, mid-case, worst-case based on TCO scenarios
  • Payback period - Months to break-even at mid-case TCO
  • Decision frame - Continue / expand / sunset, with one-sentence rationale

Panel 2: The bridge to baseline

  • Baseline state - Pre-launch numbers for the relevant KPIs in one row
  • Current state - Same KPIs at week 12, in the next row
  • Delta - Absolute and percentage change. EUR conversion where applicable
  • Cost bridge - Build cost + 12-month operating cost = total investment. Annualised savings = return. Net = ROI

Panel 3: The risks and what comes next

  • Top 3 risks - Vendor lock-in, model cost drift, compliance change, quality regression - whichever apply
  • Mitigation - One sentence each. The CFO wants to see the risks named, not hidden
  • Next 90 days - Expansion plan, second use case, scaling cost. Concrete numbers, not aspirations
  • Capital ask - If any. Clearly separated from the current pilot ROI
PanelWhat It ShowsCommon Mistake
1. HeadlineEUR savings, payback months, decisionPercentages without absolute numbers
2. Bridge Baseline → current → delta in EURSkipping baseline because it was not measured
3. Risk & nextTop 3 risks + 90-day planHiding risks behind “positive momentum”

“The companies that get the most value from AI will not be the ones chasing a single breakthrough or forcing every initiative through the same ROI lens. They will be the ones that treat AI like a portfolio - balancing routine productivity gains, targeted process improvements and selective transformational bets, while scaling winners and cutting weak ideas early.”

- Gartner, AI ROI portfolio guidance for CFOs26

How Superkind Fits

Superkind builds custom AI agents for the Mittelstand and delivers the ROI measurement framework with the build, not as a separate consulting engagement. Process-first means the baseline is measured before code is written.

  • Pre-launch baseline included - We spend the first two weeks measuring the current state of the targeted workflow. Volume, cycle time, cost per task, FTE-hours, quality. No baseline, no go-live
  • Four-tier KPI dashboard delivered - Operational, quality, financial, and strategic KPIs measured automatically from launch, with the bridge to baseline visible
  • 1.4-1.7x TCO modelled upfront - We deliver the financial model with all hidden cost categories priced. Maintenance, engineer time, compliance, vendor migration reserve. CFO-ready before week one
  • Kill criteria written in the contract - Specific thresholds at week 12 that trigger continue, re-scope, or sunset. We do not benefit from agents that should not exist
  • EU data residency - Models, telephony, transcripts in EU data centres. Reduces compliance overhead and the audit cost line item
  • Outcome-based pricing - Pricing tied to measurable containment and resolution rates, not seat licences. Aligns vendor incentive with CFO interest
  • Monthly CFO-grade report - Three-panel one-slide template delivered each month, not just at the pilot end. The report is the deliverable, not an add-on
  • Quarterly scope review - Every quarter we re-baseline, re-test on alternative models, and confirm the use case still earns its keep
ApproachGeneric AI VendorSuperkind
Baseline measurementCustomer’s problemTwo-week pre-launch baseline included
TCO modelBuild quote only1.0x / 1.4x / 1.7x scenarios with hidden costs priced
Kill criteriaImplicit, defended at all costsWritten into the contract before launch
PricingPer-seat or per-minute SaaSOutcome-based, tied to KPIs
CFO reportGeneric dashboardMonthly three-panel slide
Scope reviewAnnual contract renewalQuarterly re-baselining and re-test

Superkind

Pros

  • Baseline + TCO included - delivered before launch, not invoiced after
  • Outcome-based pricing - aligned with CFO economics
  • Written kill criteria - removes sunk-cost defence of weak use cases
  • Monthly CFO report - the three-panel slide is the deliverable
  • EU data residency - reduces compliance overhead and audit costs

Cons

  • Not a self-serve SaaS - requires engagement with our team
  • Slower start than off-the-shelf - two weeks of baseline before any agent
  • Honest TCO can scare buyers - we surface hidden costs that vendors hide
  • Capacity-limited - we work with a focused number of clients at a time

Decision Framework: Continue, Re-scope, or Kill?

At week 12 of any AI agent pilot, three numbers decide the fate. Apply this framework strictly. The biggest source of wasted Mittelstand AI budget is sunk-cost defence of pilots that should have been killed at month three.

Signal at Week 12DiagnosisDecision
Containment 60%+, CSAT at or above baseline, payback under 9 months at 1.4x TCOWorking as designedScale to full scope and plan use case #2
Containment 40-60%, quality at baseline, payback 9-15 monthsUse case is workable but scope is wrongRe-scope to a narrower workflow, re-baseline, re-test for 60 days
Containment under 40%, or CSAT below baseline, or payback past 18 months at 1.7x TCOWrong use case or wrong toolKill. Document learnings. Pick the next use case
KPIs unstable, mixed signals across tiersMeasurement system not strong enough to decidePause expansion. Fix observability and re-decide in 30 days
All KPIs trending positive but absolute values still below thresholdUse case is right, learning curve incompleteContinue at current scope for another 60 days, then re-decide

Continue vs Kill

Continue Signals

  • Containment trend - rising month over month
  • Quality stable or rising - CSAT and resolution rate hold
  • Payback in sight - under 9 months at honest TCO
  • Workflow simpler - less rework, fewer escalations

Kill Signals

  • Containment plateau - flat for 60+ days below 40%
  • CSAT regression - customers prefer the old way
  • Cost climbing - TCO grows faster than savings
  • Team workaround - employees route around the agent

Frequently Asked Questions

Most production AI agents focused on a single workflow reach payback within 4 to 9 months. Boards typically expect initial payback within 90 to 180 days for workflow-level deployments. The right comparison is not "is the agent profitable in month one" but "is the curve heading toward payback by month six". If you cannot see the slope by month three, the use case is wrong.

Three reasons. The baseline was never measured before deployment, so there is nothing to compare against. Hidden costs (maintenance, retraining, model upgrades, escalation review) get left out of the calculation. And vanity metrics (calls handled, prompts answered) replace business metrics (cost per resolved case, hours freed for skilled work). Fix all three and ROI becomes measurable.

Six numbers: containment or completion rate, average handle or cycle time, cost per task, error or escalation rate, hours freed per FTE, and CSAT or quality score. Each one needs a baseline measured before launch, a current value, and a 30-day trend. Anything else is supporting context, not headline KPIs.

Add 30 to 40 percent to the vendor or build quote for true total cost of ownership. Annual maintenance runs 15 to 25 percent of the initial build cost. Allocate 20 to 30 percent of a senior engineer time for ongoing operations - roughly EUR 3,000 to 6,000 per month at German rates. If the project still pencils out after these adjustments, the numbers are real.

It is the right reference point. McKinsey reports 88 percent of organisations use AI in at least one function, but only 6 percent attribute more than 5 percent EBIT impact to AI. The Mittelstand is not behind enterprise on this - it is a global problem. The companies that close the gap measure rigorously and scale what works, not what feels good.

Track FTE-equivalent hours freed per week per employee, output volume change at constant headcount, and reallocation of time to higher-value work. Convert hours to euros at fully loaded labour cost (gross wage plus social security plus overhead, typically 1.5 to 1.8 times gross). This translates productivity into a number CFOs accept.

Containment is the share of interactions the agent handles end-to-end without human handoff. Resolution rate is the share of interactions where the customer outcome was actually achieved (problem solved, order placed, ticket closed). A high containment with low resolution means the agent is good at not escalating but bad at solving - a measurement trap.

Both. Human baseline answers "are we better than before?" Absolute targets answer "are we good enough for the customer?" If the agent beats human handle time but customer satisfaction drops, the human comparison is misleading. Use baseline as a milestone, not as the ceiling.

Cost avoidance is real ROI but harder to defend. Pre-deployment, document the historical cost of the avoided event (e.g. average cost of a customer complaint, recall, or compliance fine). Track the rate before and after deployment. Multiply the rate reduction by the unit cost. Audit-trail this calculation - CFOs scrutinise cost-avoidance numbers more than revenue numbers.

Three signals: containment below 50 percent after 90 days, no measurable change in cycle time or cost per task, and CSAT below the human baseline at the same point. Any one of these means the use case scope is wrong. Re-scope or kill it. Sunk-cost defence of weak agents is the largest source of wasted AI budget.

RPA produces faster, narrower payback (often 3 to 6 months) on rigid scripted tasks. AI agents produce slower, wider payback (4 to 9 months) on tasks with exceptions and judgement. They are not substitutes - well-built systems use both. The CFO question is not "AI vs RPA" but "is each tool deployed where its economics work?".

Monthly during the first 6 months, quarterly after that. Re-baseline whenever the underlying process changes (new product, new system, new compliance requirement). Without re-baselining, the agent looks better than it is because the world has moved on.

Model cost has historically dropped year over year (60 to 80 percent annually for similar capability), but this is not guaranteed. Build the financial model with a 12-month flat assumption and a 24-month plus-30-percent stress test. Renegotiate vendor contracts annually. Retain the option to switch models - vendor lock-in becomes a CFO concern when costs move.

Henri Jung, Co-founder at Superkind
Henri Jung

Co-founder of Superkind, where he helps SMEs and enterprises deploy custom AI agents that actually fit how their teams work. Henri is passionate about closing the gap between what AI can do and the value it creates in real companies. He believes the Mittelstand has everything it needs to lead in AI - it just needs the right approach.

Ready to make your next AI agent CFO-defensible?

Book a 30-minute call with Henri. We will walk through your current pilot or planned use case and build the ROI framework together - no commitment, no sales pitch.

Book a Demo →