Research, signal design, and decision systems

How can we audit a KPI for Goodhart’s Law (teams gaming the metric) and redesign the signal and incentives so the metric stays a trustworthy signal?

Lucía Ferrer
Lucía Ferrer
10 min read·

Answer

Treat a gamed KPI as a broken instrument, not a moral failure. Audit it by separating the true outcome you care about from the proxy you measure, then test how easily people can manipulate the number without improving reality. Redesign by pairing the KPI with a few guardrails, verifying it with lightweight audits and sampling, and changing incentives so the easiest path to the target is the right behavior.

Most KPI problems are not math problems. They are human problems with a spreadsheet costume.

When a KPI becomes a target, it stops being a good measure, which is the core of Goodhart’s Law. The fix is not to “motivate harder” or add ten more dashboards. The fix is to re establish what the KPI is meant to represent, map how it can be hacked, verify what is actually happening on the ground, and then reshape incentives so the metric remains an honest signal rather than a game to win. (If you have ever watched people speed up only when they see a speed camera, you already understand the plot.)

Clarify what the KPI is supposed to represent (intent vs proxy)

Option Best for What you gain What you risk Choose if
Conduct independent audits Verifying data accuracy and process adherence Trustworthy data, reduced gaming Undetected fraud or systemic errors Data integrity is paramount, especially with high-stakes metrics
Define the true outcome Preventing Goodhart's Law Clear, un-gameable objective Measuring the wrong thing You are setting a new KPI or re-evaluating an existing one
Map the 'metric attack surface' Proactively identifying manipulation points Visibility into how metrics can be gamed Blind spots in data integrity You suspect data manipulation or want to build robust metrics
Implement guardrail metrics Detecting unintended consequences Early warning of negative impacts Optimizing one metric at the expense of others Your primary metric is critical and has potential side effects
Limit the number of KPIs Focusing effort and reducing complexity Clear priorities, less metric fatigue Missing important signals, oversimplification Your team is overwhelmed by too many metrics or conflicting goals
Review incentives tied to metrics Aligning behavior with desired outcomes Motivated teams, ethical data practices Unintended gaming, demotivation You observe metric manipulation or unhealthy competition

Start with the question executives often skip: “What decision is this KPI supposed to improve?” A KPI is not the goal. It is a proxy for a goal.

Write the intent in one sentence, in plain language. Examples:

“First response time is meant to represent customers feeling quickly supported.”

“Sales calls per rep is meant to represent healthy pipeline creation.”

“Story points completed is meant to represent shipping customer value.”

Now state the failure mode: “What would it look like if the KPI improved while the true outcome got worse?” This single line is the fastest way to surface Goodhart risk. It also forces you to define the time horizon. Many metrics look great this week and terrible next quarter.

Practical tip: run a “proxy assumptions” check in a 30 minute session. Ask, “For this number to be valid, what must be true about behavior, data capture, and customer impact?” If you cannot list assumptions, you cannot defend the metric.

MIT Sloan’s writing on KPI design emphasizes that metrics should support learning and decisions, not just monitoring, and that poorly designed KPIs can drive unintended behavior when they become targets. That is the starting mindset shift. See [1]

Spot practical warning signs of Goodhart’s Law and gaming

Gaming rarely announces itself as cheating. It shows up as “process improvements,” “better prioritization,” or suspiciously perfect numbers.

Look for patterns that suggest the data is lying to you:

First, bunching near thresholds. If bonuses kick in at 95 percent compliance, you will see a weird amount of 95.1 percent.

Second, sudden distribution shape changes. A metric that used to vary naturally becomes smooth, flat, or oddly consistent.

Third, short term spikes followed by long term degradation. For example, support closes more tickets, but repeat contacts or churn rises.

Fourth, channel mix shifts. If one path is measured and another is not, work migrates to the unmeasured channel.

Fifth, reduced correlation with lagging outcomes. The KPI used to predict retention or revenue and now it does not.

Sixth, audit trail anomalies. More manual overrides, more “misc” categories, more backdated entries.

Common mistake: assuming gaming only happens when people are unethical. What is more common is that you created a cliff. People adapt rationally to survive the incentive, even if the organization pays the price later. The better response is to redesign the system, not to deliver a speech about values.

For a crisp overview of how metrics fail once targeted and why “honest metrics” can die under pressure, see [2] and [3]

Map the KPI’s manipulability (how it can be hacked)

Now do what security teams do: map the attack surface. You are not accusing people. You are stress testing the measurement.

Use a simple flow: inputs, process, measurement, reporting. Then ask: “Where can someone change the number without changing reality?” The result should be a short list of levers, who can pull them, and how detectable the manipulation is.

Here is a lightweight “metric attack surface” table you can build in one working session:

Practical tip: do a “two hour ride along” with the people closest to the work. Ask them, “If your job depended on hitting this number, what would you do?” You will get better intelligence than any dashboard can provide.

Engineering and product teams often see this with velocity metrics and output counts, where teams optimize the measured activity rather than the real outcome. A concrete software oriented discussion is here: [4]

Triangulate the KPI with counter metrics and guardrails

A single KPI tied to a reward is an invitation to local optimization. Triangulation is the antidote: one primary signal, a few complementary signals, and a couple of hard guardrails.

A practical set of five, kept intentionally small:

First, an outcome metric (usually lagging). Example: retention, renewals, repeat purchase.

Second, a leading indicator. Example: activation rate, qualified pipeline created.

Third, a quality guardrail. Example: defect rate, reopen rate, customer satisfaction.

Fourth, an efficiency metric. Example: cost to serve, cycle time.

Fifth, an integrity metric. Example: percent of records touched manually, percent of items reclassified.

The point is not to drown people in metrics. It is to make common hacks visible. If the KPI improves but the guardrail moves the wrong way, you have an early warning.

A useful heuristic from experienced teams is: “If it can be gamed, it must be paired.” MIT Sloan’s KPI guidance also points toward using multiple signals to reduce perverse incentives and to keep KPIs decision relevant. See [5]

Verify reality with sampling, audits, and ground truth checks

If you only ever look at aggregates, you will be fooled by averages. Ground truth comes from sampling.

Set up three checks that do not require a large analytics project.

First, random sampling. Pick a small number of cases each week and verify the KPI classification against the underlying record and customer outcome.

Second, stratified sampling for edge cases. Over sample the categories most likely to be abused, like exclusions, escalations, and “other.”

Third, shadow measurement. Run an independent calculation from source systems, or use a “mystery shopper” style check in customer facing processes.

In parallel, watch for statistical tells: bunching near thresholds, outlier clusters by team, and drift where the KPI relationship to outcomes changes over time.

This is where “Conduct independent audits” matters most: the credibility boost often outweighs the operational cost, especially for high stakes metrics.

For more on how metric gaming shows up inside product organizations and why it often looks like “success,” see [6]

After the table, explicitly call out 2–4 of these controls by name (1 line each):

Conduct independent audits: use lightweight sampling to validate the KPI against source records and outcomes.

Define the true outcome: write the real goal in plain language and document the “looks good but is bad” failure mode.

Map the 'metric attack surface': identify where timing, reclassification, and reporting choices can distort the number.

Implement guardrail metrics: add a small set of paired signals that expose the most likely hacks.

Redesign targets and incentives so behavior aligns with outcomes

If you want the metric to stay honest, avoid “single number pay.” People will climb whatever ladder you put in front of them, even if it is leaning against the wrong wall.

Three incentive redesign moves work reliably.

First, use ranges instead of cliffs. A hard threshold creates end of period contortions. A range, like 90 to 95, reduces the incentive to game timing and categorization.

Second, turn guardrails into hard constraints. Example: you can only “win” on speed if quality stays above a minimum.

Third, raise the level of optimization. Tie a meaningful portion of rewards to team level or customer level outcomes, so one group cannot offload pain onto another.

Practical tip: separate “learning metrics” from “evaluation metrics.” If every metric is tied to compensation, people stop telling you the truth. You need some numbers that are safe to improve through experimentation.

Incentive design is system change, not a motivational poster. When you change what is rewarded, you change what people notice, what they talk about, and what gets hidden. A helpful framing is here: [7] and a cautionary list of how rewards can train “smart” bad behavior is here: [8]

Harden the metric definition (reduce ambiguity and loopholes)

Many “gaming” incidents are actually definition leaks. If “resolved” is ambiguous, someone will resolve it in the most convenient way.

Harden the metric with a simple metric contract that lives next to the dashboard:

Meaning: what this KPI represents and what it does not.

Computation: data sources, formula, timing, and refresh cadence.

Inclusion and exclusion rules: what counts, what does not, and why.

Attribution: which team gets credit in shared workflows.

Deduplication: how repeats and re opens are handled.

Versioning and change control: who can change the definition and how the change is communicated.

Also add logging that makes manipulation visible, like counts of manual edits or category changes. The goal is not surveillance. It is making the metric stable enough to trust.

Implement governance: ownership, cadence, and escalation

A KPI without an owner becomes folklore. Assign roles and decision rights so the organization knows who keeps the signal honest.

At minimum:

An executive sponsor who cares about the real outcome, not just the number.

A metric owner responsible for definition, interpretation, and changes.

A data steward responsible for data quality and lineage.

An independent QA or audit function that can sample and challenge.

Set a cadence that matches risk.

Monthly: operational review of KPI plus guardrails, with a short section on integrity signals.

Quarterly: sampling audit and review of the attack surface list.

Annually: decide whether to keep, replace, or retire the KPI.

Create an escalation rule that everyone understands: if guardrails deteriorate, or integrity signals spike, you pause incentive impact and trigger a deeper audit. This avoids the worst behavior, which is “keep paying on a broken metric because changing it is awkward.”

Executive playbook: a 30 60 90 day KPI trust reset

This is the pragmatic sequence that resets trust without blowing up morale.

Day 0 to 30: Diagnose and de risk.

Define the true outcome and write the failure mode.

Map the metric attack surface with frontline input.

Run quick integrity checks: threshold bunching, reclassification rates, manual override counts.

Communicate intent carefully: “We are improving measurement so we can invest in what works,” not “We are hunting cheaters.”

Day 31 to 60: Triangulate and verify.

Add a small set of guardrails and one integrity metric.

Start random and stratified sampling audits.

Compare KPI movement to lagging outcomes and investigate breaks in correlation.

Decide which manipulations are loopholes to close versus behaviors that reveal a process problem to fix.

Day 61 to 90: Redesign incentives and lock the definition.

Shift targets from cliffs to ranges.

Make guardrails binding for rewards.

Publish the metric contract, assign the owner, and put change control in place.

Finalize the governance cadence and the escalation playbook.

One tasteful reality check: if you pay people to hit a number, do not be surprised when they get very good at hitting the number.

If you do just one thing first, do this: define the true outcome and the “looks good but is bad” failure mode, then map the metric attack surface with the people closest to the work. That combination will tell you whether you are dealing with a data problem, a process problem, an incentive problem, or all three, and it will keep you from rewarding the KPI equivalent of a cardboard trophy.

Sources


Last updated: 2026-03-25 | Calypso

Sources

  1. sloanreview.mit.edu — sloanreview.mit.edu
  2. medium.com — medium.com
  3. whydidithappen.com — whydidithappen.com
  4. codepulsehq.com — codepulsehq.com
  5. mitsloanme.com — mitsloanme.com
  6. medium.com — medium.com
  7. howtothink.ai — howtothink.ai
  8. medium.com — medium.com

Tags

your-data-is-lying-to-you-here-s-how-to-catch-it