Skip to content
Diosh Lequiron
AI & Digital Transformation12 min read

How to Measure AI ROI Without Gaming the Numbers

Most AI ROI claims are gamed. A 3-category framework for honest measurement — time recovery, error reduction, capacity expansion — that produces numbers you can actually defend.

AI ROI claims have a credibility problem. Not because AI does not produce real value — in the right conditions, it does. The problem is that most measurement frameworks are designed to confirm the investment decision that was already made, not to tell the organization whether the investment is working.

This shows up in predictable patterns. The measurement scope narrows to the use case where results were strongest. The baseline period is selected to make the improvement look larger. Hours saved are estimated by asking workers how much time they think the tool saves — a number that will always trend optimistic because people want to be seen as adopting the new system effectively. The cost calculation includes the AI tool's license fee but not the process redesign labor, the integration work, or the ongoing maintenance.

The result is an ROI number that no one who did the measurement believes, and that no external observer can verify. It serves the purpose of continuing the AI program. It does not serve the purpose of understanding whether the AI program should be continued, scaled, redirected, or stopped.

I have run AI implementations that produced real, measurable results. The 40-70% delivery reduction and 60% cost reduction I reference in this context are not estimates — they come from measuring comparable work packages before and after AI workflow integration, using the same measurement methodology at both points. That methodology is what this article is about.

The Three Measurement Categories That Matter

Honest AI ROI measurement connects AI implementation to business outcomes. The measurement categories should correspond to outcomes that the business cares about independently of the AI program — not outcomes that are only visible through the lens of AI adoption.

I use three categories: Time Recovery, Error Reduction, and Capacity Expansion. Each category requires baseline data collected before the AI implementation, and outcome data collected after, using the same measurement approach. The comparison is the ROI claim.

These three categories are not exhaustive. Depending on the use case, other outcome categories may be relevant — customer satisfaction, employee retention, revenue per transaction. But Time Recovery, Error Reduction, and Capacity Expansion cover the majority of operational AI use cases and are measurable with data that most organizations already collect.

Measurement Category 1: Time Recovery

Time Recovery is the reduction in wall-clock time required to produce a defined unit of output, measured before and after AI implementation on comparable work packages.

The key word is wall-clock. Not estimated hours. Not self-reported time savings. Actual elapsed time from the moment a work package enters the workflow to the moment it exits as completed output, measured in the same way before and after.

This measurement requires two things: a consistent definition of "work package" and a consistent definition of "completed." If the definition of done changes between the baseline period and the measurement period — if the AI implementation also changed the output standard — the comparison is invalid. You are measuring a different thing, not an improvement to the same thing.

In practice, measuring wall-clock time means tracking work packages from entry to exit in a system that records timestamps. This can be as simple as a spreadsheet with entry date, exit date, and work package ID — as long as the data was collected consistently in both periods.

The Comparable Work Package Requirement

The most common mistake in time recovery measurement is comparing different types of work. If the AI handles the straightforward cases and humans handle the complex ones, the AI-assisted queue will always be faster — but that is not a reduction in delivery time, it is a routing decision.

To get a clean measurement, you need comparable work packages in both periods. For a content production use case, this means comparing articles of similar length, complexity, and research depth. For a customer support use case, this means comparing tickets of similar category and resolution complexity. For a document review use case, this means comparing documents of similar type and length.

If you cannot identify comparable work packages across the two periods, the measurement is not possible from historical data. In that case, run a prospective measurement: assign similar work packages to both the AI-assisted workflow and the baseline workflow simultaneously, measure both, and compare. This prospective design is more rigorous and more convincing to skeptical stakeholders than historical comparisons.

In the implementations where I measured 40-70% delivery reduction, the measurement was prospective. We ran comparable work packages through the AI-assisted workflow and the prior workflow in parallel for a defined period. The comparison was direct and contemporaneous. The reduction percentages reflect median improvement across the comparable-work-package sample, not the best-performing category.

Measurement Category 2: Error Reduction

Error Reduction measures defect rates, rework rates, and escalation rates before and after AI implementation. This category is frequently omitted from AI ROI claims, for an obvious reason: AI implementations sometimes increase error rates, at least initially, and that does not make for a compelling case study.

Including error reduction in the measurement framework is not optional. If an AI implementation reduces delivery time by 50% but doubles the defect rate, the net value is not positive. The defects have to be caught, corrected, and the corrected work re-delivered. If defect correction costs are not measured, the time savings look clean even though they are not.

The error rate baseline needs to be measured in the same way both before and after. This means having a consistent definition of "error" — not a standard that shifts based on how much pressure the team is under, not a standard that changes because the AI implementation changed what errors look like. The same rubric, applied to the same type of output, at both measurement points.

What Counts as an Error

Defining errors precisely is harder than it sounds. In most operational contexts, there are multiple error types: errors caught before delivery (rework), errors caught after delivery by the customer or downstream system (defects), and errors that propagate silently through the system without being caught (the most expensive category).

For AI ROI measurement, I distinguish between these types because they have different costs. Rework costs labor time. Defects cost labor time plus customer relationship damage. Silent propagation costs are often unknowable until a system failure surfaces them.

The measurement should track all three types. Rework rate is usually available from task management systems that track revision cycles. Defect rate is usually available from customer support or QA systems. Silent propagation errors require active sampling — pulling random output samples and evaluating them against quality standards, looking for errors that were never flagged.

Adding a silent propagation sampling step to your measurement framework will almost always reveal errors that the other two categories miss. It will also give you a more realistic picture of the AI system's actual quality level.

Measurement Category 3: Capacity Expansion

Capacity Expansion measures what the organization was able to do that it could not do before, expressed in units of output — not inputs, not activity metrics, output.

This category matters because it captures a type of value that Time Recovery and Error Reduction do not. Time Recovery shows you the same work is happening faster. Error Reduction shows you the same work is happening better. Capacity Expansion shows you that new work is happening that was not possible before because the human capacity did not exist.

The cost reduction I reference — 60% cost reduction — comes from this category. The same volume of output that previously required a team of a certain size now requires a smaller team, because AI workflows handle the volume that additional contractors would have handled. Measured in terms of output per labor cost, the reduction is approximately 60%. Measured in terms of headcount required for the same output volume, it is similar.

Capacity Expansion is also the hardest category to measure honestly, because organizations often count capacity expansion that would have happened anyway through hiring. The measurement needs to control for this: it should reflect the incremental output enabled by AI, not total output growth during the measurement period.

The Activity Metric Trap

The most common Capacity Expansion measurement error is using activity metrics instead of output metrics. Activity metrics — number of AI queries run, number of documents processed, number of tasks assigned to AI — measure input to the AI system, not output from the business process.

An organization can run 10,000 AI queries per day and have no measurable capacity expansion if the queries are not producing work that would otherwise require human time. The relevant question is: what completed outputs did the AI enable that would not have been completed without it?

This question is harder to answer than activity metrics, which is why organizations default to activity metrics. But activity metrics will not survive scrutiny from a CFO or a board member. They are not ROI. They are usage statistics.

Common Gaming Patterns

Understanding the gaming patterns in AI ROI measurement makes it easier to avoid them in your own work and to evaluate the credibility of external claims.

The most common pattern is cherry-picking the measurement scope. An AI program that touches ten use cases will have variance in results across those use cases. Measuring only the strongest performers and presenting that as program-level ROI is selection bias. It is common, it is rarely acknowledged, and it produces ROI claims that do not survive program-level review.

The second pattern is baseline manipulation. The baseline period is the period before AI implementation during which performance is measured. Selecting a baseline period that includes known disruptions — a system migration, unusual seasonal demand, a staffing gap — makes the post-implementation period look better by comparison. Selecting a baseline that excludes those disruptions gives a more accurate comparison.

The third pattern is cost understatement. AI tool licensing costs are visible and typically included in ROI calculations. Process redesign labor, integration development, training time, ongoing maintenance, and the cost of errors during the ramp-up period are often excluded. They are real costs. Including them produces a lower ROI number, but a more honest one. An ROI calculation that would collapse if implementation costs were fully included is not an ROI calculation — it is a marketing document.

The fourth pattern is time horizon gaming. AI implementations frequently show strong near-term results from the novelty effect — teams pay closer attention to quality when a new system is in place. Measuring ROI at three months and extrapolating to an annual figure without accounting for settling effects produces optimistic projections. Measure at twelve months for a more stable picture.

Operational Evidence

The 40-70% delivery reduction figure in my AI work comes from measuring time-to-delivery on comparable work packages across multiple implementation contexts. The baseline was collected in the six to twelve weeks before AI workflow integration. The measurement period began after a four-to-six-week ramp-up window to allow for calibration and training. The comparison used median time-to-delivery, not mean, to reduce the effect of outlier work packages.

The 60% cost reduction figure comes from comparing contractor costs for equivalent output volumes before and after AI integration. The specific context is a content production operation where the volume of output per quarter was held constant while the number of external contractors required was reduced. The cost per unit of output dropped by approximately 60%.

Both figures are bounded to specific contexts and implementation types. They do not generalize to all AI implementations. Delivery reduction is highest when the AI handles the first-pass work that previously required skilled human time — drafting, initial analysis, first-round review. It is lower when the AI handles post-processing steps where human time was already minimal. Cost reduction is highest when the AI reduces contractor dependency; it is lower when the primary labor cost is internal salaried staff whose costs do not change with workload.

What makes these figures defensible is that they were produced using the three-category framework described here — wall-clock time measurement on comparable work packages, error rate tracking at both periods, and output-volume measurement against labor cost. If the measurement methodology had been self-reported time estimates and activity metrics, the figures would look different and would not be worth citing.

Where This Does Not Apply

This measurement framework assumes a repeatable operational workflow with measurable inputs and outputs. It is appropriate for production workflows: content creation, document review, customer support, data processing, code review. These workflows have consistent enough structure that baselines and comparisons are meaningful.

It does not apply well to exploratory or creative AI use. If you are using AI for strategic analysis, brainstorming, or generative work where output quality is highly context-dependent and judgment-intensive, the before/after comparison is difficult to construct because "comparable work packages" are hard to define. The value of AI in these use cases is real but requires different measurement approaches — often qualitative assessment rather than quantitative measurement.

It also does not apply to AI use cases where the output is novel by definition. If the AI is enabling a product or service that did not exist before, there is no baseline to compare against. In this case, the relevant measurement is product adoption and revenue impact, not AI-specific ROI.

The framework also requires baseline data. If your organization did not track delivery times, defect rates, or output volumes before AI implementation, you cannot run a rigorous before/after comparison. In this case, start collecting the baseline data now, even if an AI implementation is already underway. Partial measurement is better than no measurement, and the data you collect now becomes the baseline for evaluating future improvements.

The Principle

Honest AI ROI measurement is not a technical challenge. It is a discipline challenge. The technical work — defining comparable work packages, tracking timestamps, counting defects, measuring output volumes — is straightforward. The discipline is measuring things that might show disappointing results and reporting them anyway, because an organization that knows the truth about its AI investments can make better decisions than one operating on optimistic estimates.

The three categories — Time Recovery, Error Reduction, and Capacity Expansion — are not designed to produce the largest possible ROI number. They are designed to produce a true one. A true ROI number that shows a 25% improvement is more valuable than a gamed number showing 70%, because the true number can be defended, scaled from, and used as a basis for investment decisions. The gamed number eventually collapses under scrutiny and takes the AI program's credibility with it.

ShareTwitter / XLinkedIn

Explore more

← All Writing