Goodhart’s Law states: when a measure becomes a target, it ceases to be a good measure. This principle was originally applied to economic policy, but it applies with equal force to product engineering.
A team measured on bugs closed will close bugs without fixing the underlying problems, or will reclassify open bugs to reduce counts, or will write code that produces easily-closable small bugs rather than addressing root causes. The metric became a target. The metric stopped measuring what it was supposed to measure.
A team measured on features shipped will ship small, low-complexity features to maximise count. The metric rewards shipping over impact. A single feature that triples retention counts as one feature. Thirty minor UI tweaks that produce no measurable change also count as thirty features.
The design principle that resists Goodhart corruption is this: choose metrics where the only viable path to improvement is to improve the actual outcome. Retention is difficult to game because the only way to retain users is to deliver continuing value. Users who find no value leave. No amount of notification-spam or dark-pattern friction holds them indefinitely.
The HEART framework provides a structured approach to this. HEART stands for Happiness, Engagement, Adoption, Retention, and Task success. Each dimension measures a different aspect of user value:
Happiness measures user satisfaction, typically through surveys or sentiment analysis. It is the hardest to measure automatically and the most qualitative. Technically, it requires infrastructure for collecting and analysing survey responses at the right moment in the user flow.
Engagement measures the depth of interaction: how often users return, how much of the product they use per session, how many of the core features they activate. High engagement is a proxy for the user finding continuing value. Technically, it requires event tracking at the feature level.
Adoption measures the proportion of users reaching a defined milestone, either for a new feature or for first activation. A high adoption metric on a new feature is only meaningful if the users who adopted it retained. Technically, adoption requires funnel tracking and cohort analysis.
Retention measures the proportion of users returning after a defined interval. Retention is the strongest leading indicator of long-term value. Technically, it requires time-series analysis of user activity and cohort comparisons.
Task success measures whether users who attempt a specific action complete it. A high task success rate means the system is doing its job. A low one means friction exists somewhere between intent and completion. Technically, it requires instrumentation at the task level, not just at the page level.
Instrumentation is not a measurement afterthought. It is an architectural requirement. A system that does not emit events at the right granularity cannot produce HEART metrics. The observability of user behaviour must be designed in from the start.