“Accounting is not a neutral record of economic events. It is a technology that makes certain kinds of facts visible and renders others invisible.”
— Peter Miller & Ted O'Leary, on what measurement infrastructure does and who controls it
Why Good Metrics Don't Become Standard
The three preceding papers in this series have established that alternatives to the engagement metric exist, that they can be operationally defined, and that they measure constructs — voluntary return, session satisfaction, attentional completion, cognitive sovereignty, collective epistemic health — that matter in ways the engagement metric does not. This would seem to be sufficient for adoption: if better metrics are available, why aren't they used?
The answer is not ignorance. The limitations of the engagement metric are documented in internal platform research (as the Haugen disclosures make clear), in academic literature (hundreds of published studies), and in regulatory records (FTC complaints, Congressional hearings, European regulatory investigations). The people who manage these metrics know what they do and do not measure. The engagement metric persists not because no one has proposed alternatives but because the institutional actors who would have to adopt alternatives have structural interests in the current regime that make adoption of alternatives costly and non-adoption costless.
Metric regime change — the replacement of one dominant measurement standard with another — is a political and institutional process, not a technical one. It requires overcoming the interests of actors who benefit from the current regime, coordinating on a replacement among actors who have incentives to defect from any common standard, and investing in the infrastructure that makes the alternative measurable at scale. None of these problems are solved by demonstrating that the alternative metrics are better. They are solved by the same mechanisms that have historically produced metric regime changes in other domains: regulatory mandate, professional body standardization, and the strategic behavior of early adopters who benefit from differentiation.
The Four Institutional Actors
The engagement metric regime is maintained by four institutional actors whose interests, individually and in interaction, make voluntary metric regime change improbable in the absence of external constraint.
Actor 1: Digital Advertising Platforms
Google, Meta, Amazon, TikTok, and their competitors are the primary generators and beneficiaries of engagement metrics. The engagement metric is the basis on which they sell advertising — it is their revenue model's core claim to advertisers. A platform that adopts alternative metrics signals to advertisers that engagement is no longer its primary optimization target, reducing the basis for the advertising premium its inventory commands. The financial interest in the engagement metric regime is direct and large: Meta's $117B 2023 advertising revenue is predicated on engagement metric performance as its primary value proposition to advertisers.
Actor 2: The Advertising Industry
Programmatic advertising buyers — agencies, brand advertisers, and the demand-side platforms that manage programmatic purchasing — have built their targeting, optimization, and reporting infrastructure around engagement metrics. Click-through rates, impression counts, engagement rates, and conversion tracking are the currencies of the programmatic advertising market. Alternative metrics would require new measurement infrastructure, new optimization algorithms, and new reporting frameworks. The coordination problem is severe: individual advertisers cannot shift to alternative metrics without their competitors also shifting, because the metrics serve competitive comparison as well as absolute performance assessment.
Actor 3: Digital Media Research and Academic Publishing
Academic research on digital media relies on engagement metrics as both dependent variables (measuring the effects of interventions on engagement) and independent variables (measuring exposure to digital media by engagement level). A metric regime change would require revising the dependent variable specifications across thousands of published studies and the independent variable specifications for ongoing longitudinal research. The academic community has an institutional interest in the current metrics not because the metrics are good but because they are standardized — and standardization is a precondition for comparability across studies.
Actor 4: Platform Regulators
Regulatory agencies that have begun to apply digital platform oversight — the FTC, the EU's Digital Services Act enforcement machinery, the UK Information Commissioner's Office — have built their enforcement tools around the existing metric regime. GDPR enforcement tracks consent rates and data processing volumes. DSA enforcement tracks illegal content removal rates and algorithmic transparency disclosures. None of the existing regulatory frameworks mandate cognitive welfare metrics, because cognitive welfare metrics are not part of the regulatory reporting infrastructure that regulators have built their processes around. Regulators have institutional interest in the metrics they currently use, because shifting to alternative metrics requires rebuilding enforcement processes, training enforcement staff, and drafting new guidance.
What Each Has to Lose
| Institutional Actor | Current Metric Benefit | Alternative Metric Cost | Defection Incentive |
|---|---|---|---|
| Advertising Platforms | Revenue justified by engagement scale; no accountability for welfare externalities | Engagement decline under welfare optimization; reduced advertising CPMs; accountability for welfare outcomes | Very high — direct revenue threat |
| Advertising Industry | Standardized optimization infrastructure; comparable performance across platforms | Infrastructure rebuild cost; loss of cross-platform comparability during transition | High — coordination problem; first mover disadvantaged |
| Academic Research | Cross-study comparability; replication of existing study designs | Measurement infrastructure investment; existing literature partially deprecated | Moderate — individual researchers disadvantaged by changing standards |
| Regulators | Enforcement processes built on existing metrics; clear causal chain from metric to harm | New guidance required; enforcement staff retraining; ambiguity in early implementation | Low — regulators respond to legislative mandate, not market incentives |
Precedents for Metric Regime Change
The engagement metric is not the first dominant measurement standard to produce systematic harm while resisting voluntary replacement. The history of metric regime change in other domains provides both models for how change happens and realistic assessments of how long it takes.
Generally Accepted Accounting Principles (GAAP)
Before GAAP standardization in the 1930s and 1940s, companies used accounting metrics of their own construction, which they optimized to project desired financial images to investors. The Securities Exchange Act of 1934 and the subsequent delegation of accounting standard-setting to the Financial Accounting Standards Board established mandatory standardized metrics with independent audit requirements. The transition took decades and required legislative mandate, not voluntary adoption. The parallel to cognitive welfare metrics is direct: the engagement metric is a company-controlled measure optimized to project desired performance images to advertisers and investors; the alternative metrics proposed in this series are designed to function as standardized measures with independent audit requirements.
Nutrition Labeling
Food companies voluntarily disclosed nutritional information before the Nutrition Labeling and Education Act (1990), but the disclosures were inconsistent, self-serving, and not comparable across products. The NLEA mandated standardized disclosure — specific nutrients, standardized serving sizes, required placement on packaging. The voluntary disclosure era produced noise; the mandatory standardized disclosure era produced actionable information. The parallel is exact: platforms voluntarily report engagement metrics that are self-serving and not comparable across platforms; mandatory welfare metric disclosure with standardized definitions would produce actionable information.
Carbon Reporting
The transition from voluntary to mandatory greenhouse gas reporting in the EU (Corporate Sustainability Reporting Directive, effective 2024 for large companies) provides the most recent precedent. Voluntary carbon disclosure (the CDP, the Task Force on Climate-related Financial Disclosures) was adopted by large companies as a reputational and investor relations instrument from 2000 onward; the CSRD mandated standardized disclosure across all large EU companies regardless of voluntary commitment. The mandatory standardization resolved the comparability problem that voluntary disclosure could not: mandatory standards with third-party audit requirements produce comparable data; voluntary disclosure produces incomparable self-reports.
The Disanalogy
A critical difference separates these precedents from the cognitive welfare metric case. GAAP mandated disclosure of financial quantities companies were already measuring internally. Nutrition labeling mandated disclosure of nutritional properties that food science could already quantify. Carbon reporting mandated disclosure of emissions that atmospheric chemistry could already calculate. In each case, the regulatory mandate required disclosure of an already-measurable quantity — the innovation was mandatory standardized reporting, not the development of a new measurement instrument. The Cognitive Sovereignty Index proposed in this series does not yet exist as a validated psychometric instrument. The precedents demonstrate that mandatory disclosure works once the metric exists. They do not demonstrate that regulatory mandates can conjure the metric into existence. The measurement instrument must be developed, validated, and demonstrated to be reliable before the regulatory analogy fully applies.
The Four Required Interventions
The Institutional Lock cannot be broken by persuasion, voluntary commitment, or market competition alone. The precedents in Section IV uniformly show that metric regime changes require external mandate from an actor whose interests are not aligned with the current metric regime. The following four interventions are designed to work together as a package; any one of them alone is insufficient to produce the full metric regime change that cognitive sovereignty governance requires.
Intervention 1: Regulatory Mandate for Welfare Metric Reporting
Digital platforms above a jurisdictional user threshold (proposed: 1 million monthly active users in the jurisdiction) must report standardized cognitive welfare metrics — voluntary return rate, session satisfaction, attentional completion rate, post-session well-being — in annual public disclosures, audited by independent third parties appointed by the regulatory authority. This is the analog of the CSRD carbon disclosure mandate: it creates a reporting obligation that applies regardless of voluntary commitment, with standardized definitions that make disclosures comparable across platforms, and with third-party audit requirements that make disclosures credible.
The Legal Architecture series (LA-001 through LA-005) describes the statutory anatomy this mandate requires. The intervention is not novel — it is the application of an established regulatory tool (standardized mandatory disclosure with independent audit) to a new domain (cognitive welfare metrics).
Intervention 2: Advertiser Welfare Standards
The advertising industry — agencies, brand advertisers, and programmatic buying platforms — can require welfare metric performance as a precondition for advertising placement, analogous to the brand safety standards that major advertisers already apply (preventing ads from appearing adjacent to extremist content, for example). An advertiser welfare standard would require that platforms seeking advertising from brand-safety-committed advertisers disclose their welfare metric performance and meet minimum thresholds for voluntary return rate and post-session well-being scores.
This intervention works through market incentives rather than regulatory mandate, but it depends on coordination among major advertisers — the same coordination problem that afflicts the advertising industry's relationship to any standard. The World Federation of Advertisers, the Association of National Advertisers, and the Interactive Advertising Bureau are the appropriate coordination venues. The parallel is the Global Alliance for Responsible Media (GARM), which established brand safety standards that became industry-wide when major advertisers committed collectively.
Intervention 3: Academic Journal Disclosure Requirements
Academic journals publishing research on digital media effects should require disclosure of the engagement metrics used in the research — both as independent variables (measuring digital media exposure) and as dependent variables (measuring platform engagement outcomes) — and should require discussion of the limitations of those metrics for the welfare constructs the research addresses. This intervention reforms the academic literature infrastructure that currently treats engagement metrics as neutral proxies for outcomes they do not reliably measure.
The model is the APA's CONSORT guidelines for clinical trials: standardized reporting requirements that make methodological choices visible and auditable. Journal-level disclosure requirements do not mandate metric change; they create visibility that enables scholarly critique and shapes the research agenda toward metric development.
Intervention 4: Public Welfare Metric Reporting Obligation
Platforms above the jurisdictional threshold should be required to publish quarterly public welfare metric reports — standardized, in machine-readable format, accessible to researchers and the public — analogous to the quarterly financial reports that public companies are required to file with securities regulators. The welfare metric reports would include the platform-level voluntary return rate, session satisfaction, and post-session well-being scores, broken down by user demographic where privacy-preserving disclosure is feasible, and the Cognitive Sovereignty Index population scores for the platform's user base where that instrument has been validated.
Public reporting serves three functions that private regulatory reporting does not: it enables civil society monitoring, it creates reputational accountability (a platform whose welfare metrics decline visibly faces public pressure that it does not face from private regulatory disclosure), and it creates the data infrastructure that MR-003's attentional commons measurement requires.
The Pathway to Standardization
The four interventions interact in sequence. Academic journal disclosure requirements (Intervention 3) establish the research infrastructure that validates the alternative metrics and produces the evidence base that regulators and advertisers need to justify the mandate. Advertiser welfare standards (Intervention 2) create early-adopter incentives that demonstrate the feasibility of welfare metric adoption before regulatory mandate. Regulatory mandate (Intervention 1) resolves the coordination problem by requiring all covered platforms to adopt standardized metrics regardless of voluntary commitment. Public reporting requirements (Intervention 4) create the accountability infrastructure that makes the regulatory mandate enforceable and creates the data for attentional commons monitoring.
The realistic timeline, calibrated against the carbon reporting precedent, is a decade from regulatory mandate to functional standardization across major platforms. The CSRD was adopted in 2022; large EU companies are required to comply by 2025; standardized comparable data will not be available until 2027 or later. The equivalent cognitive welfare mandate, if adopted in the EU or UK in 2027, would produce standardized comparable data by 2030–2032.
This is not slow by the standards of metric regime change — GAAP took two decades; nutrition labeling took a decade from mandate to stable implementation. But it is slow relative to the pace of harm: the engagement metric has been actively degrading the attentional commons since approximately 2012, and standardized measurement of that degradation will not be available until the next decade at the earliest. This is the argument for beginning the legislative work now, not after further documentation of harm: the Institutional Lock does not break faster because the evidence of harm accumulates. It breaks when a legislature acts.
The gaming objection to mandatory welfare metric reporting holds that platforms will optimize for metric scores rather than the underlying welfare outcomes — producing the same Goodhart's Law inversion that corrupted the engagement metric. This is a genuine risk. The carbon reporting analogy is instructive: scope 3 emissions reporting is widely acknowledged to be gamed through creative accounting for supply chain emissions. Voluntary return rate could be gamed by suppressing push notifications while deploying email and SMS alternatives; post-session well-being scores could be gamed by timing prompts at positive emotional moments.
The response is threefold. First, gaming is detectable: third-party audit requirements with platform data access can identify systematic gaming by comparing the metrics against behavioral telemetry. Second, gaming is self-limiting when the gaming itself becomes visible: the engagement metric was gamed openly and invisibly for a decade precisely because no external audit existed; welfare metric gaming under a mandatory disclosure regime will be visible to auditors and researchers. Third, the alternative to imperfect mandatory disclosure is no disclosure — which produces no accountability whatsoever. An imperfect accountability regime is better than no accountability regime, provided the imperfections are known and addressed iteratively. This is the lesson of financial accounting: GAAP is full of known limitations and gaming vectors; it is nonetheless vastly better than the pre-GAAP era of unconstrained accounting discretion.
What the Reformation Demands
The Measurement Reformation is the series that closes the argument that the Measurement Crisis opened. The Measurement Crisis (MC-001 through MC-006) documented how the engagement metric became a target, how targets replace measures, and how measurement inversion corrupts everything downstream. The Measurement Reformation documents what the replacement looks like — not as a wish but as a specification: four alternative metrics defined and validated (MR-001), a composite index proposed and calibrated (MR-002), population-level collective indicators developed (MR-003), and the institutional pathway to adoption mapped (this paper).
The reformation demands what reformations always demand: actors with authority over the institutional regime who are willing to use it. In the domain of cognitive welfare metrics, those actors are legislatures and regulatory bodies. The Legal Architecture series has documented what the statutory anatomy of cognitive sovereignty law requires. The Measurement Reformation series has documented what that law needs to mandate: not just platform behavior, but platform measurement — specific, standardized, audited, public welfare metrics that make visible the harm the engagement metric has rendered invisible.
The Institutional Lock is real. It will not be broken by the quality of the alternative metrics alone, by the evidence of harm alone, or by the volume of academic critique alone. It will be broken when a legislature, somewhere, acts on the evidence that the four prior sagas have assembled, the legal architecture this saga has specified, the design standards this saga has named, and the measurement infrastructure this series has proposed. That is the task that the Institute's research program exists to inform.
The HEXAD series that follows will specify the practical implementation layer — how individual people, institutions, and organizations translate the policy prescriptions of Saga V into lived practice. The measurement infrastructure proposed here is one input to that implementation: a way of knowing, at the individual and collective level, whether the restoration is working.
Selected Sources
- Miller, P., & O'Leary, T. (1987). Accounting and the construction of the governable person. Accounting, Organizations and Society, 12(3), 235–265.
- Securities Exchange Act of 1934. Pub. L. No. 73-291, 48 Stat. 881 (1934). (GAAP mandate precedent)
- Nutrition Labeling and Education Act of 1990. Pub. L. No. 101-535, 104 Stat. 2353 (1990).
- European Commission. (2022). Corporate Sustainability Reporting Directive. Directive 2022/2464/EU.
- World Federation of Advertisers. (2020). Global Alliance for Responsible Media (GARM): Brand Safety Floor + Suitability Framework.
- Goodhart, C. (1975). Problems of monetary management: The UK experience. Papers in Monetary Economics, Reserve Bank of Australia.
- Haugen, F. (2021). Facebook internal research documents. Submitted to the U.S. Securities and Exchange Commission.
The Institute for Cognitive Sovereignty. (2026). The Measurement Reformation [ICS-2026-MR-004]. The Institute for Cognitive Sovereignty. https://cognitivesovereignty.institute/measurement-reformation/the-measurement-reformation