A metric framework is a structured system for selecting, organising, defining, and governing the metrics an organisation uses to measure performance, track progress, and make decisions. Rather than accumulating metrics ad hoc — adding a new dashboard whenever a new question arises — a metric framework establishes a deliberate architecture: which metrics matter, how they relate to each other, who owns them, how they are calculated, and how they connect to strategic objectives. A well-designed metric framework transforms a scattered collection of data points into a coherent measurement system that drives aligned action across every level of an organisation.
The fundamental problem a metric framework solves is measurement proliferation. Without a framework, organisations accumulate metrics over time — each team, function, and initiative adds its own measurements — resulting in hundreds of disconnected data points, conflicting definitions, duplicated reporting, and an inability to distinguish the vital few signals from the trivial many. When a company has 200 metrics but no structure, it effectively has no metric system at all: executives cannot identify what matters most, teams optimise for local measures that conflict with company-level goals, and data debates consume more time than the insights they are meant to generate.
A metric framework imposes discipline on this problem by establishing a hierarchy, a taxonomy, and a governance model. It answers the questions that raw metrics cannot: Why is this metric tracked? What decision does it inform? What is the chain of causality connecting this operational measurement to a strategic outcome? Who is accountable when this metric moves? What is the authoritative definition used across all reporting? These structural answers are what transform metrics from measurement artefacts into management tools.
Core Components of a Metric Framework
Every metric framework, regardless of the specific methodology used, contains a set of foundational components that give it structure and operational utility. These components define what gets measured, how it is measured, why it matters, and who is responsible for acting on it. The absence of any one component creates a corresponding gap in the framework’s effectiveness.
| Component | Description | Key Questions Answered |
|---|---|---|
|
Metric Hierarchy
|
A tiered structure organising metrics from strategic to operational levels
|
Which metrics matter most? How do lower-level metrics connect to top-level goals?
|
|
Metric Definitions
|
Precise, unambiguous formulas and calculation rules for each metric
|
How exactly is this metric calculated? What data sources are used?
|
|
Ownership and Accountability
|
Named individuals or teams responsible for each metric’s performance
|
Who is accountable when this metric declines? Who has authority to act on it?
|
|
Targets and Benchmarks
|
Defined thresholds representing good, acceptable, and concerning performance
|
What does success look like? When should this metric trigger an intervention?
|
|
Data Sources and Lineage
|
Documentation of where each metric’s underlying data originates
|
Where does this number come from? How reliable is the underlying data?
|
|
Review Cadence
|
Defined frequency for reviewing, reporting, and acting on each metric
|
How often is this metric reviewed? Who sees it and in what format?
|
|
Causal Relationships
|
Mapping of how metrics influence each other (leading vs. lagging)
|
Which metrics predict others? What levers drive change in outcome metrics?
|
The Metric Hierarchy
The metric hierarchy is the structural backbone of any metric framework. It organises metrics into tiers based on their relationship to strategic outcomes, their audience, and their level of aggregation. At the top sits the single most important metric — often called the North Star Metric or primary success metric — that represents the core value the organisation creates. Below it are a small number of strategic metrics that collectively describe overall business health. Below those sit departmental and functional metrics, and at the base sit the granular operational and diagnostic metrics that explain movements in the layers above.
| Tier | Name | Typical Count | Audience | Purpose |
|---|---|---|---|---|
|
Tier 1
|
North Star / Primary Metric
|
1
|
Entire organisation
|
Defines the single most important measure of organisational success
|
|
Tier 2
|
Strategic / Executive Metrics
|
5–10
|
Board, C-suite, senior leadership
|
Comprehensive health check across all major business dimensions
|
|
Tier 3
|
Departmental / Functional Metrics
|
10–30
|
Department heads, team leads
|
Functional performance and contribution to strategic metrics
|
|
Tier 4
|
Operational / Diagnostic Metrics
|
30–100+
|
Teams, individual contributors
|
The North Star Metric
The North Star Metric (NSM) is the single metric that most accurately represents the core value the product or organisation delivers to customers. It sits at the apex of the metric hierarchy and serves as the ultimate arbiter of whether the organisation is moving in the right direction. Unlike financial metrics such as revenue or profit — which are outcomes of value delivery rather than direct measures of it — the North Star Metric attempts to capture the experience of customer value itself. For Airbnb, the North Star is nights booked; for Spotify, it is time spent listening; for a SaaS business, it might be weekly active users engaging with the product’s core workflow.
The power of the North Star Metric lies in its unifying function: when an organisation rallies around a single top-level metric, all teams can orient their work toward moving the same number. Product teams ask whether new features will increase the NSM. Marketing teams ask whether campaigns bring in users who contribute to the NSM. Customer success teams ask whether their interventions protect the NSM by keeping customers engaged. This alignment, when achieved, dramatically reduces the internal coordination cost of prioritisation debates.
North Star Metric Examples by Business Type:
E-commerce: Gross Merchandise Volume (GMV) or Orders per Month
SaaS / B2B: Weekly Active Users (WAU) using core feature
or Net Revenue Retention (NRR)
Consumer App: Daily Active Users (DAU) or Session Length
Marketplace: Transactions Completed or Gross Bookings
Media / Content: Time Spent or Content Consumed per User
Financial Services: Assets Under Management (AUM) or Loans Originated
Healthcare SaaS: Patient Outcomes Improved or Appointments Completed
Metric Framework Typologies
Several named metric frameworks have been developed and adopted across the technology, product management, and business strategy communities. Each offers a different lens for organising and selecting metrics, and organisations frequently combine elements from multiple frameworks to suit their specific context. Understanding the major typologies allows practitioners to draw on the most appropriate structure for their measurement problem.
AARRR (Pirate Metrics)
Developed by venture capitalist Dave McClure, the AARRR framework — nicknamed “Pirate Metrics” for its acronym — organises metrics around the five stages of the customer lifecycle: Acquisition, Activation, Retention, Referral, and Revenue. Each stage has its own set of metrics, and the framework makes the causal chain from first customer touch to monetisation explicit. AARRR is particularly popular in early-stage startups and growth teams because it maps directly to the funnel mechanics most relevant during the growth phase of a business.
AARRR Framework:
A — Acquisition: How do users find you?
Metrics: CAC, channel conversion rate, organic vs. paid traffic mix
A — Activation: Do users experience your product's core value?
Metrics: Time-to-first-value, onboarding completion rate, day-1 retention
R — Retention: Do users come back?
Metrics: DAU/MAU, churn rate, NRR, cohort retention curves
R — Referral: Do users tell others?
Metrics: NPS, referral rate, viral coefficient (K-factor)
R — Revenue: Do users pay?
Metrics: MRR, ARR, ARPU, LTV, LTV:CAC ratio
HEART Framework (Google)
Developed by Google’s research team, the HEART framework organises metrics around five dimensions of user experience quality: Happiness, Engagement, Adoption, Retention, and Task Success. Unlike AARRR — which follows the customer journey chronologically — HEART evaluates the quality of the user experience across multiple dimensions simultaneously. It is particularly useful for product teams evaluating the impact of UX changes, new feature launches, or redesigns where the goal is to improve experience quality rather than simply move funnel conversion numbers.
| Dimension | Description | Example Metrics |
|---|---|---|
|
Happiness
|
User satisfaction and sentiment
|
CSAT score, NPS, app store rating
|
|
Engagement
|
Depth and frequency of user interaction
|
Sessions per user per week, features used per session
|
|
Adoption
|
New users or features gaining traction
|
|
|
Retention
|
Users returning over time
|
|
|
Task Success
|
Users completing intended actions efficiently
|
Balanced Scorecard
Developed by Robert Kaplan and David Norton in the early 1990s, the Balanced Scorecard organises metrics across four strategic perspectives: Financial, Customer, Internal Processes, and Learning and Growth. Its core insight was that financial metrics alone are insufficient for managing a business — they are lagging indicators of past decisions, and a company that tracks only financial results will always be reacting to history rather than managing its future. The Balanced Scorecard adds three forward-looking perspectives that capture the operational and capability conditions that produce financial outcomes, creating a more complete and balanced picture of organisational performance.
| Perspective | Core Question | Example Metrics |
|---|---|---|
|
Financial
|
How do we look to shareholders?
|
Revenue growth, ROE, EBITDA margin, FCF
|
|
Customer
|
How do customers see us?
|
|
|
Internal Processes
|
What must we excel at?
|
On-time delivery, defect rate, cycle time, OEE
|
|
Learning and Growth
|
Can we continue to improve?
|
Employee engagement score, training hours, innovation pipeline
|
Input-Output-Outcome Framework
The Input-Output-Outcome framework distinguishes between three types of metrics based on their position in the causal chain of value creation. Input metrics measure the resources and effort invested. Output metrics measure the activities and deliverables produced by those inputs. Outcome metrics measure the actual change in customer or business results produced by those outputs. This framework is particularly useful for separating execution measurement (did we do the work?) from impact measurement (did the work produce the intended result?), a distinction that many organisations blur when they mistake activity tracking for performance management.
Input → Output → Outcome Chain Example:
INPUT: Engineering headcount allocated to feature development
(Metric: Developer hours invested; Marketing spend)
OUTPUT: Features shipped; campaigns launched; content published
(Metric: Features released per quarter; ads served)
OUTCOME: Change in customer behaviour or business results
(Metric: Feature adoption rate; revenue generated; NRR improvement)
Key Principle:
Optimising inputs without tracking outcomes creates activity without impact.
Organisations must measure all three levels to manage the full causal chain.
Leading vs. Lagging Metrics in a Framework
A well-designed metric framework explicitly maps the leading and lagging relationships between metrics. Lagging metrics confirm what has already happened — revenue, profit, customer count, churn — and are the ultimate measures of business success. Leading metrics predict what is likely to happen — pipeline coverage, product engagement scores, employee satisfaction — and are the levers that management can pull in advance of outcomes deteriorating. Without this mapping, organisations manage by looking in the rearview mirror: they discover problems only after the financial results already reflect them, by which point intervention is expensive and delayed.
| Lagging Metric | Leading Predictors | Intervention Window |
|---|---|---|
|
Monthly Churn Rate
|
Product login frequency, support ticket volume, NPS decline, feature adoption rate
|
30–90 days before churn event
|
|
Revenue Growth
|
Sales pipeline value, lead conversion rate, trial-to-paid conversion
|
30–60 days before revenue realisation
|
|
Employee Turnover
|
60–180 days before resignation
|
|
|
Customer Acquisition
|
Website traffic, MQL volume, demo request rate, content engagement
|
14–45 days before conversion
|
|
System Reliability (Uptime)
|
Error rate trends, deployment frequency, MTTR, test coverage
|
Hours to days before incident
|
Metric Definition Standards
One of the most practically valuable outputs of a metric framework is a metric dictionary or data catalogue: a document that records the authoritative definition of every metric the organisation tracks. Without this, the same metric name refers to different calculations in different teams — “active users” means monthly to the marketing team, weekly to the product team, and daily to the growth team. These definition divergences produce the most common and damaging failure mode in organisational measurement: data debates. When leadership spends executive meeting time arguing about which number is correct instead of deciding what to do about the business, the metric framework has failed.
| Field | Description | Example Entry |
|---|---|---|
|
Metric Name
|
Official name used across all systems
|
Monthly Recurring Revenue (MRR)
|
|
Definition
|
Plain-language description of what is measured
|
Total normalised monthly value of all active subscription contracts
|
|
Formula
|
Precise calculation rule
|
Sum of (Annual Contract Value / 12) for all active subscriptions at month end
|
|
Data Source
|
Authoritative system of record
|
Salesforce CRM — Opportunity object, Closed Won stage
|
|
Owner
|
Accountable individual or team
|
Head of Finance / Revenue Operations
|
|
Review Cadence
|
How often formally reviewed
|
Weekly (internal); Monthly (board reporting)
|
|
Inclusions / Exclusions
|
Edge cases and boundary conditions
|
Includes multi-year contracts normalised to monthly; excludes one-time professional services fees
|
Metric Framework Design Principles
Effective metric frameworks are guided by a set of design principles that prevent the most common measurement pathologies. These principles are not abstract ideals — each one is the direct solution to a specific failure mode that organisations encounter when their measurement systems grow without structure or discipline.
| Principle | Description | Failure Mode It Prevents |
|---|---|---|
|
Fewer is more
|
Track only metrics that drive decisions; ruthlessly eliminate vanity metrics
|
Metric proliferation — hundreds of metrics, none prioritised
|
|
Outcome over activity
|
Prioritise outcome metrics over input and output metrics
|
Organisations measuring effort instead of impact
|
|
Causal clarity
|
Map the mechanisms connecting metrics — how does A affect B?
|
Optimising metrics in isolation without understanding interactions
|
|
Single source of truth
|
One authoritative definition and data source per metric
|
Data debates consuming decision-making time
|
|
Ownership accountability
|
Every metric has one named owner with authority to act
|
Metrics monitored by everyone and owned by no one
|
|
Actionability
|
Every tracked metric must connect to a decision or action
|
Reporting metrics no one acts on — measurement theatre
|
|
Balance
|
Pair each optimisation metric with a counter-metric to prevent gaming
|
Teams optimising one metric while inadvertently destroying another
|
Counter-Metrics and Metric Pairing
One of the most important and frequently overlooked design principles in metric frameworks is the use of counter-metrics — a paired metric that guards against the unintended consequences of optimising for a primary metric in isolation. Every metric optimisation creates pressure to achieve the number by any means available, including means that produce the number while destroying value elsewhere. Counter-metrics make these trade-offs visible and discourage gaming. A product team tasked with increasing daily active users might achieve the number by sending aggressive push notifications — temporarily inflating DAU while destroying user satisfaction and increasing uninstall rates. Pairing DAU with a satisfaction or retention counter-metric makes this trade-off immediately visible.
| Primary Metric | Counter-Metric | Gaming Behaviour Prevented |
|---|---|---|
|
Daily Active Users (DAU)
|
User satisfaction score / Uninstall rate
|
Aggressive notifications inflating sessions without genuine engagement
|
|
Sales Contracts Closed
|
Customer retention rate / NRR
|
Signing poor-fit customers to hit quota; high subsequent churn
|
|
Bug Resolution Time
|
Closing tickets without fixing root cause to hit time target
|
|
|
Response Time (Support)
|
First Contact Resolution (FCR) rate
|
Responding quickly with unhelpful answers to hit SLA
|
|
Content Published
|
Engagement rate / Time on page
|
Publishing low-quality content volume to hit publishing targets
|
Metric Framework Governance
Governance is the operational system that keeps a metric framework alive, accurate, and aligned with organisational priorities over time. Without governance, metric frameworks decay: definitions drift, new metrics are added without review, ownership becomes unclear, and the framework gradually returns to the fragmented state it was designed to replace. Governance defines who can add, modify, or retire metrics; how frequently the framework is reviewed; what process is required to change a metric definition; and how metric quality and data integrity are monitored on an ongoing basis.
| Governance Element | Description |
|---|---|
|
Metric Review Board
|
Cross-functional group (Finance, Product, Data, Operations) that approves new metrics, definition changes, and retirements
|
|
Annual Framework Review
|
Full review of all Tier 1 and Tier 2 metrics against current strategic priorities; retire metrics that no longer drive decisions
|
|
Definition Change Process
|
Formal process for updating metric definitions, including impact assessment on historical data comparability
|
|
Data Quality Monitoring
|
Automated checks on data freshness, completeness, and consistency for all framework metrics
|
|
Metric Ownership Registry
|
Maintained list of all metrics with current owners, last reviewed date, and data source documentation
|
Metric Framework vs. Dashboard
A metric framework and a dashboard are frequently confused, but they represent fundamentally different things. A dashboard is a visualisation tool — it displays metrics in a format designed for monitoring and review. A metric framework is the decision architecture that determines which metrics belong on the dashboard, how they are defined, why they matter, and what actions they should trigger. A dashboard without a framework is a collection of charts; a framework without a dashboard is a document. Both are needed: the framework provides the structure and meaning; the dashboard provides the visibility and accessibility. Many organisations invest heavily in dashboard tooling while neglecting framework design — producing beautifully designed displays of the wrong metrics, inconsistently defined, with no clear connection to strategic priorities.
Metric Framework in Investor and ESG Context
Investors in growth-stage and public companies assess the maturity of a metric framework as a proxy for management quality and operational discipline. A management team that can clearly articulate their North Star Metric, explain the causal chain connecting operational metrics to financial outcomes, and demonstrate consistent metric definitions across reporting periods signals that the business is managed with rigour rather than intuition. In due diligence processes, inconsistent or undefined metrics — different ARR figures on different slides, unexplained changes in how churn is calculated quarter-over-quarter — are significant red flags that indicate either poor data governance or deliberate metric manipulation.
In the ESG domain, the metric framework concept is being applied to sustainability reporting with increasing rigour. Frameworks such as GRI (Global Reporting Initiative), SASB (Sustainability Accounting Standards Board), and TCFD (Task Force on Climate-related Financial Disclosures) are essentially metric frameworks for non-financial performance: they define which ESG metrics matter, how they should be calculated, what data should be disclosed, and how performance should be contextualised against industry benchmarks. As ESG reporting requirements move from voluntary to mandatory in many jurisdictions, organisations that have already built robust ESG metric frameworks will have a significant compliance and credibility advantage over those that have treated sustainability metrics as discretionary communications exercises.
Common Metric Framework Pitfalls
Even well-intentioned metric framework initiatives frequently fail or produce limited value due to a predictable set of implementation errors. Understanding these pitfalls in advance allows organisations to design their frameworks with the structural safeguards needed to avoid them. The most destructive pitfall is Goodhart’s Law — the principle that when a measure becomes a target, it ceases to be a good measure. As soon as individuals and teams are held accountable for a metric, they will find ways to move the metric that may or may not reflect genuine improvement in the underlying condition being measured. This is not a failure of integrity; it is a natural response to measurement pressure that must be anticipated and designed around.
| Pitfall | Description | Prevention |
|---|---|---|
|
Goodhart’s Law
|
Metrics become targets and lose validity as genuine performance measures
|
Use counter-metrics; rotate metrics periodically; separate measurement from incentives
|
|
Vanity Metrics
|
Tracking impressive-looking numbers that don’t connect to business outcomes
|
Apply the “so what?” test: what decision does this metric inform?
|
|
Metric Proliferation
|
Continuously adding metrics without retiring old ones
|
Enforce a one-in-one-out policy; require business case for new metrics
|
|
Definition Drift
|
Metric definitions change over time without documentation
|
Formal change management process; version-controlled metric dictionary
|
|
Data-Decision Gap
|
Metrics are tracked and reported but not acted upon
|
Every metric must have a named owner with a defined action playbook
|
|
Misaligned Incentives
|
Team metrics optimised locally at the expense of company-level outcomes
|
Ensure team metrics are derived from and aligned to company-level framework
|
Related Terms
- KPI (Key Performance Indicator) — The individual metrics that populate a metric framework; KPIs are the building blocks that a framework organises into a coherent measurement system
- OKR (Objectives and Key Results) — A goal-setting framework that uses metrics as Key Results; OKRs and metric frameworks are complementary — OKRs define where to go, the metric framework defines what to track along the way
- North Star Metric — The single top-tier metric in a hierarchy framework representing core customer value; sits at Tier 1 of the metric hierarchy
- Leading Indicator — A forward-looking metric that predicts future outcomes; a critical component of any balanced metric framework
- Lagging Indicator — A retrospective metric confirming past outcomes; financial KPIs are typically lagging indicators
- Balanced Scorecard — A metric framework typology organising performance across Financial, Customer, Internal Process, and Learning and Growth perspectives
- AARRR (Pirate Metrics) — A metric framework organising metrics across the Acquisition, Activation, Retention, Referral, and Revenue stages of the customer lifecycle
- HEART Framework — Google’s metric framework for evaluating user experience quality across Happiness, Engagement, Adoption, Retention, and Task Success
- Goodhart’s Law — The principle that when a measure becomes a target, it ceases to be a good measure; a foundational design constraint for metric frameworks
- Vanity Metric — A metric that looks impressive but does not connect to meaningful business decisions; a metric framework’s primary purpose is to eliminate vanity metrics from reporting
- Counter-Metric — A paired metric that guards against the unintended consequences of optimising a primary metric in isolation; essential for preventing metric gaming
- Data Governance — The policies and processes for managing data quality, ownership, and consistency; the operational foundation on which a metric framework depends
External Resources
- Amplitude — The North Star Playbook:
 Building a metric framework around a single defining success metric - Google re:
Work — Setting Goals with OKRs:  The relationship between OKRs and underlying metric frameworks - Harvard Business Review — The Balanced Scorecard:
 Measures That Drive Performance (Kaplan and Norton) - McKinsey and Company — Is Your Company Measuring What Matters:
 Strategic metric selection and framework design - Global Reporting Initiative (GRI) Standards — The leading ESG metric framework for sustainability performance reporting
Disclaimer
The information provided in this article is intended for educational and informational purposes only. Metric framework concepts, typologies, design principles, and governance recommendations discussed herein reflect general industry conventions, widely cited management literature, and publicly available practitioner guidance as of the time of writing. Specific frameworks referenced — including AARRR, HEART, Balanced Scorecard, and others — are the intellectual property of their respective originators and are described here for educational purposes. Implementation approaches vary significantly by organisation size, industry, business model, and maturity stage. Nothing in this article constitutes management consulting, data strategy, legal, financial, or professional advice. Readers should conduct independent research and consult qualified professionals before designing or implementing metric frameworks within their organisations. Uninformed Investors makes no representation as to the accuracy, completeness, or timeliness of the information contained herein.
Metric Framework definition is complete. The article covers: core components, metric hierarchy (4 tiers), North Star Metric concept and examples, major framework typologies (AARRR, HEART, Balanced Scorecard, Input-Output-Outcome), leading vs. lagging metric mapping, metric definition standards and data dictionary structure, design principles, counter-metrics and pairing, governance model, framework vs. dashboard distinction, investor and ESG context, common pitfalls including Goodhart’s Law, and all related terms.