The Observation
The three series of Saga II โ the Accountability Gap, Engineered Incompetence, and the Capability Crisis โ were developed as independent research programs, each following evidence in a distinct institutional domain. The Accountability Gap examined the governance vacuum surrounding lethal autonomous weapons systems. Engineered Incompetence examined how scientific measurement instruments become captured by the institutions they are designed to evaluate. The Capability Crisis examined the structural collapse of cognitive and professional competence across multiple sectors of the workforce.
These three domains appear to have nothing obvious in common. Military AI governance is a problem of international law and arms control. Instrument capture in institutional science is a problem of epistemology and research incentives. Workforce competence degradation is a problem of education, organizational culture, and cognitive development. Different institutions, different literatures, different proposed remedies.
And yet when the three bodies of evidence are placed side by side and the question is asked not "what is failing" but "how is it failing," the answer is identical in all three cases. The mechanism of failure is the same. The architecture that produced the failure is the same. The reason the failure is self-sustaining is the same.
What they share is the removal of productive friction โ and the consequent elimination of the self-correction mechanisms that productive friction enabled. This paper names that shared architecture and establishes why three independent collapses that appear to address different problems are, at the level of mechanism, one collapse in three registers.
The First Collapse: The Accountability Vacuum
The Accountability Gap documents a specific and precise governance failure: the deployment of increasingly autonomous lethal systems in the absence of any legal framework assigning accountability for deaths those systems cause. This is not a gap in the sense of a missing regulation awaiting drafting. It is a structural void โ a space in which no legal actor exists who can be held responsible for a lethal autonomous decision.
The void was not created by malice. It was created by the removal of human decision-making from the lethal decision loop for reasons that appeared locally rational at each step of the removal. Human operators are slower than autonomous systems. Humans fatigue. Humans introduce inconsistency. From a pure operational-efficiency standpoint, every argument favors removing the human from the loop. Each removal was approved by institutional review processes that evaluated the local efficiency gain without evaluating the systemic accountability consequence.
The Legal Review process required by Additional Protocol I of the Geneva Conventions applies when a new weapon is "studied, developed, acquired or adopted." No major military program has failed Legal Review on grounds of inadequate human accountability over a lethal decision in over two decades of autonomous weapons development. The review process has become a rubber stamp โ not because reviewers are corrupt, but because the accountability question has been removed from the review criteria by incremental redefinition of what "human in the loop" means.
The friction that was removed was accountability itself โ the requirement that a human being be identifiable who made the lethal decision and can be held responsible for it. When accountability is present, it creates friction: decisions slow down, approvals are harder to obtain, documentation is required, commanders are cautious. When accountability is removed, decisions accelerate, approvals become automatic, documentation requirements atrophy, and caution disappears. The system becomes more efficient in the short term and incapable of self-correction in the long term. There is no mechanism for the system to notice when it is causing unjustifiable harm, because the actor who would have noticed โ the accountable human โ has been removed.
The Second Collapse: Instrument Capture
Engineered Incompetence documents a different kind of institutional failure: the process by which the measurement instruments designed to hold scientific institutions accountable are captured by those institutions and converted from oversight tools into performance management tools. The result is that the institutions being evaluated begin to optimize for the metric rather than for the underlying thing the metric was designed to measure โ and the metric improves as the underlying thing deteriorates.
Peer review was designed as a friction mechanism. Requiring independent expert evaluation before publication imposes delay, requires justification, forces engagement with contrary evidence, and creates the possibility that work will be rejected or substantially revised. These are all costs. They are also what makes peer review useful as an error-correction mechanism. A paper that passes frictionless review has demonstrated only that it was submitted. A paper that survives adversarial review has demonstrated something about its quality.
Replication rates in psychology (36% in the Open Science Collaboration's 2015 study), cancer biology (roughly 11% of landmark findings replicated in the Reproducibility Project: Cancer Biology), and preclinical pharmacology (as low as 20-25% across multiple independent assessments) reflect the systematic capture of peer review by publication incentive structures. The friction of replication โ the requirement that an independent lab produce the same result before a finding is considered established โ has been almost entirely removed from normal scientific practice.
The capture happened incrementally, through the same locally rational decisions that removed accountability in the military domain. Each step of the process was improved in isolation: submission timelines shortened, reviewer pools expanded to reduce bottlenecks, impact factor metrics introduced to reward productivity, funding tied to publication counts to increase output. Every individual improvement made the system more efficient. Collectively, they removed the friction that peer review depended on to function as an error-correction mechanism. The system became more productive and less accurate simultaneously. There is now no reliable mechanism by which institutional science self-corrects, because the corrective mechanism โ adversarial review, replication, and rejection of bad work โ has been systematically devalued.
The Third Collapse: The Readiness Collapse
The Capability Crisis documents the systematic degradation of cognitive and professional competence across multiple workforce sectors and educational pipelines. The mechanism is specific: the removal of productive difficulty from educational, training, and professional development processes โ in the name of accessibility, equity, retention, and completion metrics โ has produced workers who cannot perform the demanding cognitive tasks their job titles describe.
Difficulty is a friction mechanism. The struggle required to acquire a genuine skill โ the frustration, failure, revision, and extended effort that mastery demands โ is not an unfortunate side effect of learning. It is the process by which durable neural pathways are established, by which skills transfer beyond the original learning context, by which genuine competence is distinguished from performance. Removing difficulty to improve completion rates produces completers who have not acquired the skills the credential represents.
The National Assessment of Educational Progress (NAEP) โ the "Nation's Report Card" โ shows that the proportion of US 8th graders performing at or above proficiency in mathematics fell from 34% in 2019 to 26% in 2022, even as grade point averages continued to rise. The gap between credential performance (GPAs, graduation rates, completion metrics) and competence performance (NAEP, PISA, licensing exam pass rates) has widened consistently for three decades. The friction-removal interventions designed to improve completion metrics have successfully improved completion metrics while degrading the competencies those metrics were designed to represent.
The military domain provides the starkest illustration of the downstream consequences. Recruiting shortfalls reaching 41,000 in fiscal year 2023 for the US Army โ the worst since the end of the Vietnam-era draft โ occurred not because of declining interest in military service but because the cognitive and physical standards required for service could not be met by a sufficient proportion of the eligible population. The pipeline had been depleted not at the entry point but upstream, in the educational and developmental systems that were supposed to prepare people for demanding institutions. The Capability Crisis is what happens when the institutions that produce competence remove the difficulty that makes competence possible.
The Root Architecture
The three collapses describe different institutions, different failure modes, and different domains of consequence. The Accountability Gap produces unaccountable lethal decisions. Instrument capture produces unreliable scientific knowledge. The Readiness Collapse produces a workforce incapable of performing the tasks civilization requires of it. These are genuinely distinct problems.
But the architecture that produced them is the same in each case. In each domain, a mechanism existed that imposed friction on the system's outputs. The friction was costly in the short term โ it slowed approvals, delayed publications, reduced graduation rates, increased the effort required to operate. The friction was removed or attenuated in the name of efficiency, throughput, accessibility, and measurable performance improvement. The removal succeeded in improving the short-term metrics. It simultaneously removed the system's capacity for self-correction, because the friction was the self-correction mechanism.
Friction removed: human accountability requirement in lethal decision loop. Efficiency gain: faster, more scalable autonomous operations. Self-correction lost: no actor who can recognize unjustifiable harm and alter behavior.
Friction removed: adversarial peer review, replication requirement, publication of negative results. Efficiency gain: higher output volume, shorter timelines, better publication metrics. Self-correction lost: no reliable mechanism to distinguish true findings from artifacts.
Friction removed: difficulty, failure tolerance, demanding standards in education and training. Efficiency gain: higher completion rates, better graduation metrics, lower attrition. Self-correction lost: no reliable mechanism to distinguish genuine competence from credential acquisition.
The shared architecture is not metaphorical. In each case, an error-correction loop existed, was identified as a source of inefficiency, and was atrophied through a series of individually justifiable decisions. In each case, the atrophy was invisible in the short term because the efficiency metrics improved while the thing the efficiency metrics were supposed to measure continued to deteriorate. In each case, the detection of the atrophy required measurement instruments that were themselves subject to the same friction-removal dynamic โ and were therefore unreliable at precisely the moment they were needed.
What Productive Friction Is
The concept requires precision. Friction is not suffering for its own sake. Not all friction is productive. The argument is not that systems should be made more difficult arbitrarily or that hardship improves outcomes regardless of its form. The argument is specific: there is a class of friction that is structurally necessary for self-correction, and removing it destroys the system's capacity to identify and repair its own errors.
Productive friction has a precise definition: it is resistance that forces a system to encounter evidence that its outputs are wrong before those outputs become irreversible. Accountability friction in the military domain forces decision-makers to encounter the question of whether a lethal decision is justifiable before it is executed. Review friction in science forces researchers to encounter the question of whether their findings are reproducible before they enter the literature. Difficulty friction in education forces learners to encounter the question of whether they have actually acquired a skill before they are credentialed as having done so.
"Productive friction is not an obstacle to the system's purpose. It is the mechanism by which the system can tell whether it is achieving its purpose. Remove it and the system becomes maximally efficient at producing outputs that may have nothing to do with the outcomes it was designed to produce."
The distinction between productive and unproductive friction is important because the arguments used to remove productive friction are always made in the name of legitimate values โ efficiency, accessibility, speed, equity, scale. These are not bad values. The problem is that applying them without distinguishing between frictions that produce waste and frictions that produce self-correction removes both kinds. A well-designed friction-reduction intervention would identify which friction is waste (redundant approvals, irrelevant prerequisites, administrative burden with no evaluative function) and preserve the friction that serves as error correction. What actually happens, consistently, is that all friction is treated as equivalent waste, and the efficiency improvements are captured while the error-correction capacity silently disappears.
The Removal Mechanism
In all three domains, the removal of productive friction followed the same sequence. First, the friction is identified as an inefficiency โ a bottleneck, a barrier, a source of attrition. Second, metrics are introduced to measure the cost of the friction: approval timelines, publication delays, dropout rates. Third, institutional incentives are restructured to reward the reduction of those metrics. Fourth, the friction is reduced, the metrics improve, and the improvement is treated as evidence that the friction's removal was beneficial. Fifth, the error-correction capacity that the friction provided continues to deteriorate, but the metrics that would capture that deterioration are the same metrics that were already compromised by the removal โ so the deterioration does not appear in the data until it reaches crisis proportions.
The sequence is identical in military procurement review, in the transformation of peer review under publish-or-perish incentives, and in the grade inflation and standards reduction of the K-12 and higher education systems. The specific actors, specific institutions, and specific historical moments differ. The logic is invariant.
The critical structural feature of this sequence is that it produces no internal signal of distress until very late. Efficiency metrics are improving. Institutional leaders can present rising output numbers, shorter timelines, and better completion rates. The error-correction function is atrophying silently in the background, producing no metric that deteriorates visibly until the errors it would have caught begin to generate consequences in the external world โ autonomous systems causing unjustifiable harm, unreplicable science producing failed drug trials, an underprepared workforce unable to staff critical systems.
Systems That Cannot Self-Correct
The defining feature of the Institutional Void is not that the three institutions have failed. It is that they have lost the capacity to recognize and repair their failures. The void in the title is not an absence of institutions. It is an absence of the corrective function that institutions are supposed to perform.
A system that can self-correct has three properties. It has mechanisms that generate feedback about the quality of its outputs โ reviews, evaluations, test results, accountability structures. It has actors who receive that feedback and are empowered to act on it โ reviewers, accountability holders, assessors. And it has processes by which the feedback translates into behavioral change โ revision, rejection, reassignment, regulatory response.
In each of the three collapsed domains, all three properties have been degraded. The feedback mechanisms have been captured or attenuated โ review processes that no longer generate adversarial feedback, accountability structures that no longer hold anyone responsible, assessment processes that no longer measure the underlying competency. The actors who receive the feedback have been restructured out of authority โ autonomous systems that replace accountable humans, publication pipelines that no longer require replication, credentialing processes that no longer require demonstrated difficulty. And the processes for translating feedback into change have been severed โ legal reviews that rubber-stamp rather than constrain, funding structures that reward output regardless of quality, institutional cultures that treat failure as a retention and recruitment problem rather than an error-correction signal.
The Institutional Void is therefore not a temporary malfunction. It is a structural state in which the three domains cannot identify their own failures, identify no actors to hold responsible for those failures, and have no processes for translating identified failures into corrections even in principle. This is what it means for an institution to have lost its self-correction capacity: not that it has failed to correct itself, but that the mechanism by which it would have done so has been removed.
The Simultaneity Problem
The three collapses are not merely structurally similar. They are temporally correlated. The critical period of friction removal in all three domains concentrates between roughly 1985 and 2015, with accelerating atrophy in the 2000s and 2010s. This is not coincidental. The simultaneity reflects a shared external cause: the generalization of productivity and efficiency ideology across institutional domains that had previously been at least partially insulated from it.
The New Public Management movement of the 1980s and 1990s introduced business productivity frameworks โ output metrics, throughput optimization, efficiency measurement โ into public institutions that had previously operated under different logic. Research universities adopted impact factor and citation count metrics. Military procurement adopted business-case analysis frameworks. Educational institutions adopted completion and retention metrics as primary accountability measures. The same framework was applied to all three simultaneously, and all three responded by optimizing for the framework's metrics โ which meant identifying friction as waste and removing it.
The Research Excellence Framework in the UK (introduced 1986, accelerated 1992), the Government Performance and Results Act in the US (1993), and No Child Left Behind (2001) represent the same productivity ideology applied to research, government, and education respectively within a fifteen-year window. Each introduced output metrics that rewarded friction reduction. Each produced the predictable combination of improving metrics and degrading underlying function.
The simultaneity matters for intervention strategy. If the three collapses had independent causes, they could be addressed independently. Because they share a common causal structure applied simultaneously, the remedies that would interrupt the shared architecture need to be applied simultaneously across domains. A reform of scientific review processes that leaves military accountability structures and educational standards in their current state will be partially re-captured by the broader institutional culture that continues to treat all friction as waste.
Cross-Domain Amplification
The three collapses do not merely share a root architecture. They amplify each other through specific cross-domain channels, making the combined collapse more severe and more durable than any of the three in isolation.
The most important cross-amplification runs from the Capability Crisis to both other domains. Engineered Incompetence documents the capture of scientific measurement โ but scientific measurement is conducted by scientists who were trained in the same educational system that the Capability Crisis has documented as degraded. A workforce with reduced critical thinking capacity, reduced statistical literacy, and reduced tolerance for difficulty will produce different peer review, different replication standards, and different scientific norms than the workforce that designed those institutions. The Capability Crisis is one of the inputs into the Engineered Incompetence dynamic, not merely a parallel problem.
The same cross-amplification applies in the military domain. The Accountability Gap's review processes โ whatever remains of them โ are staffed by legal and policy professionals who were trained in the same degraded institutions the Capability Crisis documents. The quality of Legal Review for autonomous weapons systems is partly a function of whether the lawyers and ethicists conducting it have been trained to tolerate the cognitive difficulty of genuinely adversarial evaluation, or trained in institutions that rewarded them for meeting completion metrics. The pipeline matters.
-
CC โ EI: The scientific workforce produced by degraded educational standards applies weaker statistical rigor, less critical review, and lower replication expectations than the workforce produced by more demanding standards. The Capability Crisis is an upstream input into instrument capture.
The researchers who normalize p-hacking were trained in institutions that had already removed the friction of demanding methodology courses.
-
EI โ AG: Unreliable institutional science fails to produce the empirical basis needed for accountability regulation. Autonomous weapons policy depends on reliable empirical data about system performance, failure modes, and harm patterns โ data produced by the same scientific institutions whose reliability Engineered Incompetence documents as compromised.
The evidence base for autonomous weapons regulation is produced by science that cannot reliably replicate its own findings.
-
AG โ CC: The removal of human accountability from high-stakes decisions reduces the stakes experienced by the workforce that makes those decisions, attenuating the seriousness that demanding professional development requires. When no individual is accountable for outcomes, the training that prepares individuals to accept accountability becomes harder to justify and sustain.
The seriousness of training is partly a function of the seriousness of the consequences. Remove the consequences and training standards follow.
Why Single-Domain Solutions Fail
Reforms are underway in each of the three domains. Open science initiatives are attempting to restore replication requirements and pre-registration of hypotheses. Defense policy discussions are producing proposals for meaningful human control standards for autonomous weapons. Educational reform movements are attempting to restore academic rigor and resist grade inflation. These efforts are real, evidence-based, and staffed by people who understand their respective domains well.
They are all subject to the same structural problem: they are attempting to restore productive friction within a single domain while the broader institutional culture that systematically treats friction as waste remains intact across all three domains. The shared root architecture continues to generate pressure toward friction removal in each domain, and the cross-domain amplification dynamics continue to supply impaired inputs from the other domains even when one domain partially repairs itself.
Replication requirements in science will be partially undercut by a workforce whose training in methodology and statistical reasoning has been degraded by the Capability Crisis. Autonomous weapons accountability standards will be partially undercut by a scientific evidence base that Engineered Incompetence has made unreliable. Educational rigor restoration will be partially undercut by an accountability culture in the institutions that employ graduates which has removed the demand signal that makes rigorous training valuable.
The void cannot be filled one domain at a time. It requires coordinated restoration of productive friction across all three domains simultaneously โ and it requires naming the shared architecture that removed it, rather than treating three domain-specific failures as three independent problems with independent solutions.
The Compound Collapse
The Institutional Void names what the three series of Saga II collectively document: a compound collapse of institutional self-correction capacity across the three domains most critical to civilizational resilience. Military governance, scientific knowledge production, and workforce competence are not peripheral functions. They are the three primary systems by which a civilization manages existential risks, generates reliable knowledge about the world, and maintains the human capacity required to operate complex institutions. All three have lost their self-correction mechanisms through the same architectural decision.
The compound nature of the collapse matters for the same reason the compound nature of the Capture Loop matters: the interactions between the three domains make the combined collapse more severe than the sum of the individual collapses. A civilization with unreliable science, unaccountable lethal autonomous systems, and an underprepared workforce is not three times as vulnerable as a civilization with any one of those problems. It is qualitatively differently vulnerable, because the three problems interact in ways that make each harder to address in the presence of the others.
"The void is not an absence of institutions. It is an absence of the corrective function that institutions are supposed to perform. The buildings are still standing. The processes still run. The metrics still improve. But the mechanism by which the system notices it is wrong and changes โ that has been removed."
The recovery architecture for the Institutional Void is not obscure. The productive friction that was removed can be restored: genuine accountability structures in lethal autonomous decision-making, genuine adversarial review and replication requirements in science, genuine difficulty standards in education and professional training. These are not technically impossible. They are politically and economically costly, because the efficiency gains from friction removal have been captured by institutions that will resist their reversal.
What the Institutional Void establishes is that the three domains are one problem, and restoring any one of them in isolation will be partially undermined by the others. This is the strategic insight that the individual series cannot provide: the problem is not three failures in three domains. It is one failure in three registers. The response must be commensurate with the problem's actual scope.