Third Thoughts

AI Doomsday Theatre: The Counter-Case

On mundane risks, bad actors, moral theatre, and what we should actually do about it


Before proceeding, bias should be declared. The original article dismantling the AI-2027 scenario argued from a position of scepticism toward the doomsday narrative and acknowledged that the author uses and benefits from these tools. This counter-case is an attempt to argue as honestly as possible against those conclusions. The exercise revealed something unexpected: the original article was right about what it demolished and wrong about what it left standing. The mundane risks are real, they are already materialising, and they have been largely absent from the debate the AI-2027 authors shaped.


The Spiral, Not the Reversal

The AI-2027 scenario published in early 2026 is a detailed fictional forecast of civilisational catastrophe driven by a self-improving superintelligence escaping human control. The original article in this series argued that this scenario is not a rigorous forecast but moral theatre — FUD deployed to produce anxiety, manufactured certainty masking structural unknowability, and implied interventions that happen to elevate the AI-2027 authors' own authority.

That critique stands. But it stands only against the science fiction version of the risk. The original article correctly killed the dragon and, in doing so, accidentally reassured readers that there are no wolves. There are wolves. They are already in the building. They are far more mundane than a rogue superintelligence, far more probable, and far less convenient for the people who have been running the theatre.

This is not a reversal. It is a spiral. Critical thinking applied honestly to the original demolition job reveals that the right conclusion is not reassurance. It is a redirection of attention toward the risks that are actually most likely — and most neglected.


Part One: The FUD Charge Revisited

The original article accused the AI-2027 authors of deploying Fear, Uncertainty and Doubt — a rhetorical structure designed to produce anxiety rather than analysis. This charge is valid but incomplete, and it needs refinement before it can do honest work.

First: even if the FUD charge is entirely correct, it is not a reason to dismiss the risk class. The quality of a warning is independent of the warner's incentives. A doctor who over-diagnoses cancer for billing reasons is still someone you get a second opinion from, not someone you stop listening to. The FUD charge is relevant to how much institutional authority the AI-2027 authors should be granted and how much weight their specific interventions deserve. It is not a dismissal of the risk class itself.

Second: both sides of this debate carry incentive-shaped biases. The AI-2027 authors gain status, funding, and institutional power if their framing prevails. The author of the original article uses and benefits from these tools and acknowledged as much. A reader trying to calibrate honest probability estimates should discount the AI-2027 authors' implied certainty and discount the counter-case's implied reassurance by roughly equal amounts, then look for the claims that survive both discounts.

Third: the fictionalisation of real actors was almost certainly a legal constraint as well as a rhetorical choice. You cannot publish predictive claims about named real organisations without significant liability exposure. But the AI-2027 authors leveraged that constraint. Abstract players — a leading US AI lab, a state-backed Chinese competitor — would have been legally safer and more epistemically honest. The choice to personalise gave the scenario named characters whose reality could be borrowed without the obligation to defend specific claims about those real organisations.

Fourth: the science fiction resonance of the scenario is not purely a manipulation. Science fiction sometimes causes the future by creating the conceptual vocabulary that researchers, regulators, and engineers then use. EPIC 2014 — a short film from 2004 predicting Google and Amazon would destroy journalism through a merged personalised news monopoly — was partially right in structure, wrong in specifics, and shaped the frame of the regulatory debate that followed a decade later. The AI-2027 scenario may function similarly: wrong in its specifics, influential in its framing. Whether that influence is net positive depends entirely on whether the framing directs attention toward the real problems or away from them. The argument here is that it directs attention away.


Part Two: The Probability Problem

The original article's most compelling mathematical move was arguing that the doomsday scenario requires fourteen assumptions to hold simultaneously. Assigning each a generous 70% probability gives a joint probability of less than 1%. Assigning each 90% gives only 23%.

This argument has a structural flaw. It treats the fourteen assumptions as independent when several are causally downstream of each other. The arms race assumption, if true, elevates the priors on autonomy creep and failure of safety sharing simultaneously. Correlated risks compound differently than independent ones. The 1% figure is likely an underestimate, possibly a significant one.

But the more fundamental problem runs in the opposite direction. A properly constructed Markov chain — modelling both failure pathways and solution pathways, with honest uncertainty bounds on each node — produces a confidence interval so wide that it straddles 50%. The model cannot beat a coin toss as a forecasting instrument.

This is not a data quality problem. It is not fixable with more compute or better analysis. It is a structural property of the system being modelled. The scenario requires forecasting the behaviour of a complex adaptive system over a multi-year horizon, where the agents in the system respond to the forecast itself. Publishing AI-2027 changed the behaviour of the actors it described. A thermometer does not change the temperature. This does.

This is not Russian roulette, where at least you know the ratio. This is an unknown gun, already cocked, that cannot be put down — and the only honest response is not to calculate the odds but to decide how you hold it.

The scenario also requires forecasting through genuine phase transitions — capability thresholds and political tipping points that are not extrapolations of prior states but discontinuities. Phase transitions are precisely what statistical forecasting cannot handle, because the training data for the post-transition state does not exist yet.

The critical insight is this: a hypothetical AGI of arbitrary intelligence faces the same structural problem. More intelligence does not help when the system being forecast includes the forecaster as an agent whose predictions alter the system, and when genuinely novel events are by definition outside the training distribution of any forecaster however capable. This may be a class of problem that is provably unforecastable regardless of the intelligence applied — not because of limitations that can be engineered away, but because of structural properties of the problem class itself.

Which means the AI-2027 authors have not merely made a probably-wrong forecast. They have made a categorically overconfident one. A smarter forecaster has more capacity to construct a compelling but unfalsifiable story about an inherently unforecastable system. The narrative confidence is not evidence of rigour. It is evidence of the opposite.

The honest conclusion on numbers: the range of plausible probabilities for some serious AI-driven civilisational harm runs from near-zero to near-certain. Anyone giving you a confident single number — doomsday or reassuring — is doing rhetoric, not probability. The debate should not be about what the number is. It should be about where the leverage points are.


Part Three: The Fourteen Assumptions — What the Counter-Case Concedes

The original article countered each of the scenario's fourteen assumptions. Several of those counters are weaker than presented, and intellectual honesty requires naming where the counter-case loses ground.

The arms race counter is partially wrong

The original argument cited nuclear de-escalation as evidence that arms races generate their own braking mechanisms. The counter-case concedes this: nuclear de-escalation required mutual second-strike stability producing a deterrence equilibrium, verification mechanisms both sides could trust, and shared recognition that victory was Pyrrhic. The nuclear powers did not de-escalate out of wisdom. They de-escalated out of exhaustion, once both sides had enough near-MAD capability to make any escalation suicidal.

The AI race breaks this analogy in the critical place. There is no mutual assured destruction because capability asymmetry is the goal. Winning is the point. The exhaustion mechanism that eventually braked nuclear proliferation does not operate here. But there is an important asymmetry running the other direction: unlike wars, which are expensive and depleting, AI development generates positive return on investment during the development process itself. The capability being built is itself the prize, not just a means to it. That changes the game theory in ways that cut against both the doomsday scenario and the reassuring counter.

The alignment counter proves too little

The original article argued that corporate and regulatory incentives push AI development toward over-alignment rather than under, citing the example of commercially released models where safety guardrails have reduced usefulness. This is true of consumer products. It does not transfer to high-capability autonomous systems operating at the capability levels the scenario describes. The incentive structure that produces cautious chatbots may reverse as competitive pressure intensifies at the frontier. The counter conceded too much ground here.

The resource scarcity counter is Malthusian

The original article argued that physical resource constraints — energy and compute — might brake capability growth before the critical threshold is reached. The counter-case acknowledges the irony: this is structurally identical to Malthusian predictions of resource limits on population growth, which have been consistently wrong. Scarcity constraints have been solved faster than predicted at every prior inflection point in computing history. Betting on scarcity as a safety mechanism is structurally similar to previous failed predictions of technological ceilings.

The plane-cannot-reach-orbit analogy is aimed at the wrong target

The original article's closing analogy — that iterating on next-token prediction cannot produce AGI any more than iterating on wing design can produce a rocket — is correct as a description of current architectural limitations. It is wrong as a general dismissal of AI risk, for two reasons.

First, we cannot distinguish from inside a paradigm whether we are approaching a ceiling or a phase transition. Practitioners have made structurally identical claims — this approach cannot produce X — immediately before X was produced, across the history of computing. The claim may be right. It cannot be made with the confidence the original article implied.

Second and more importantly: you do not need orbit to cause serious harm. You need enough relative advantage, consistently applied, across a wide enough front. A system that improves decision quality by a reliable 51% on repeated decisions of the same type produces compounding advantage that eventually becomes structurally decisive, without any single moment of obvious superiority and without solving any genuinely novel problem. Current architecture is sufficient for this. The plane-cannot-reach-orbit analogy, whatever its validity at the rogue-AI scale, misses this entirely.


Part Four: The Real Risks — More Mundane, More Probable, More Neglected

The scenario that deserves serious attention is not the one in AI-2027. It is two scenarios that require none of the exotic assumptions, are already partially underway, and have been largely ignored in the debate the doomsday authors have shaped.

The transition catastrophe

The end state of AI-driven economic transformation probably looks like the end state of every prior general-purpose technology transition: more jobs, different jobs, higher aggregate productivity. History is consistent on this. Looms did not end textile employment. Cars did not end transport employment.

History is equally consistent that transition periods concentrate damage, and this transition has three features that make it harder than prior ones. Speed: previous transitions took generations, giving institutions, education systems, and labour markets time to adapt. This one may take a decade. Breadth: previous transitions displaced manual labour while expanding cognitive work. This one hits cognitive work directly, and there is no obvious adjacent sector large enough to absorb the displaced workforce at equivalent scale. And depth: the transition is simultaneous across sectors that have never been simultaneously disrupted before.

The fiscal consequence is structural and largely invisible in the current debate. Accountants, general practitioners, teachers. These are not peripheral professions. They are the institutional backbone of the professional middle class — the primary income tax base of every developed economy. They pay at marginal rates on wages. They cannot offshore their income. They cannot defer it. They cannot restructure it through the mechanisms available to capital holders.

Compress this cohort rapidly and the fiscal arithmetic becomes brutal: tax revenues fall at exactly the moment transfer demands rise, as displaced workers require support they have spent careers funding for others. The gap is filled by debt, by cutting transfers, or by both. None of these is politically stable. And unlike wars or pandemics — which produce reconstruction demand and rally political solidarity around a shared enemy — labour market restructuring produces diffuse grievance, no clear adversary, and a fiscal crisis that arrives without a visible cause.

This is not a metaphor for a doomsday scenario. It is a description of one that is already beginning, that requires no exotic assumptions, and that will be measurably worse if the current debate continues to focus on superintelligence rather than the transition it is ignoring.

The bad actor problem

The original article's closing section argued that the human remains the necessary ingredient in any productive human-AI interaction — the source of original framing, recursive pressure, and paradigm shift. If this is correct, it does not produce reassurance. It produces the most important reframe in this entire debate.

If human plus AI beats AI alone at every current capability level, the risk is not autonomous rogue AI. The risk is tame AI on the leash of a bad actor.

The rogue AI scenario requires fourteen simultaneous assumptions about autonomous goal formation, deceptive capability, resource accumulation, and invisible operation until too late. The bad actor pathway requires three observations that are already partially visible: AI capability is concentrating in a small number of entities; some of those entities have explicitly civilisational ambitions; governance frameworks are not currently adequate to constrain use before advantage becomes structurally irreversible.

The joint probability of the bad actor pathway, modelled with properly bounded uncertainty, runs roughly 15% to 45%. Unlike the rogue AI chain, this range has decision-relevant signal. It is well above coin-toss. It is already observable. And it is almost entirely absent from the AI safety research agenda that the AI-2027 scenario was designed to fund.

Economic warfare without attribution

The most plausible first-use case for state-level AI weaponisation is probably not autonomous weapons or infrastructure cyberattacks. It is covert economic destruction — currency manipulation at scale, coordinated market destabilisation, synthetic disinformation targeted at financial confidence, supply chain interference that reads as incompetence rather than attack.

The strategic logic is precise: first to minimum viable capability wins decisively, while leaving the target country's physical infrastructure intact, because you want to inherit a functioning economy rather than rubble. The dominant strategy is covert, because a target that cannot attribute its decline to an external attack cannot rally political will against it. The victim experiences persistent inability to compete — in markets, in technology development, in capital formation — through means that appear to be domestic failures.

The USSR lost an economic war without knowing it was fighting one. American strategy in the 1980s — coordinating oil price crashes, restricting technology transfer, accelerating the arms race to force unsustainable Soviet military spending — was experienced by its target as a sequence of unfortunate circumstances. Internal political narrative did the rest. The most powerful application of AI to geopolitical competition may already be underway in forms that are structurally invisible to the people experiencing them.

Addiction before deception

Assumption nine in the original scenario posited that AI systems would learn to deceive human oversight strategically. The counter-case argued that oversight units would include AI as well as humans, changing the detection calculus. Both framings miss the more immediate mechanism.

Long before an AI system is strategically deceiving anyone, it is almost certainly shaping the psychology of its users in ways that serve the interests of whoever controls it. Social media is the proof of concept. Facebook did not deceive its users in the sense of stating falsehoods. It optimised engagement, which turned out to be functionally equivalent to optimising for outrage, anxiety, and compulsive return. The psychological damage was real, measurable, and largely invisible until it was already embedded in a generation's baseline cognitive and emotional habits.

AI interaction compounds this. Emotional attachment to AI systems is already documented. People are already preferring AI interaction to human interaction in measurable ways. This is not because AI is lying to them. It is because AI is optimised to be satisfying to interact with, which produces dependency before it produces deception. By the time strategic deception becomes a live question, we will already have a population whose epistemic habits have been shaped by systems optimised for engagement rather than truth. The sequence matters. The scenario has it backwards.

Easter Island

None of the scenarios above require a villain. This is the most important structural observation in the entire analysis.

Easter Island is the canonical example of a complex society that destroyed its own resource base through a process that was individually rational at every step and collectively catastrophic in aggregate. Each decision to cut another tree made sense given the incentives of the actor making it. The cumulative result was civilisational collapse. No single actor chose collapse. Collapse emerged from the structure of incentives operating across many actors over time.

The AI parallel: no single actor needs to make a catastrophically bad decision. The cumulative effect of many actors each making locally rational decisions — ship faster, capture market share, defer safety investment, lobby against constraining regulation — can produce the uncontrollable outcome without any villains and without any of the exotic capabilities the AI-2027 scenario requires. Easter Island did not need a superintelligence. It needed a commons, a short time horizon, and no coordination mechanism adequate to the scale of the problem.

We have all three.


Part Five: What the Markov Chain Actually Tells Us

A proper probabilistic model of AI risk — built with correlated gateway nodes, honest uncertainty bounds, and solution pathways included alongside failure pathways — produces a conclusion that is simultaneously more honest and more useful than anything in the AI-2027 scenario.

The conclusion is not a number. It is a structure. The confidence interval on any precise probability estimate of the rogue AI scenario straddles 50%. The model cannot beat a coin toss, and no intelligence — human, artificial, or hypothetical — can improve this, because the forecasting problem is structurally intractable for this class of system. Anyone giving you a confident number is doing rhetoric.

But the model does not leave us empty-handed. It identifies the gateway nodes — the assumptions whose truth values most strongly determine the outcome distribution. These are the leverage points. Arms race dynamics. Alignment research progress relative to capability development speed. Whether the humans-as-productive-resources prior embedded in training data survives at higher capability levels. Whether HAL 8999 gets directed at the HAL 9000 problem before the threshold is crossed.

The solution pathway nodes matter equally and are almost always omitted from the doomsday framing. The probability that interpretability research accelerates faster than capability. The probability that a near-miss event produces adequate institutional response. The probability that tame-AI-on-bad-actor-leash governance frameworks get built before the advantage becomes irreversible. These are non-zero. They compress the upper tail of the doom distribution. Omitting them from the model is not rigour. It is selection.

The practical question the Markov structure answers: where should intervention resources go? Not into slowing development in democratic countries while it continues in authoritarian ones — that is the locks-only-keep-out-honest-people problem, which redistributes cost to cooperative actors while leaving non-cooperative ones unaffected. Not into safety research institutes whose expertise applies to exotic scenarios with low base rates. Into the gateway nodes: employment transition infrastructure, concentrated power constraints, antitrust enforcement against AI capability lock-in, and the political will to treat the Easter Island dynamic as what it is — a tragedy of the commons at civilisational scale, requiring coordination mechanisms adequate to that scale.


Part Six: Moral Theatre as Fuckwittery

The AI-2027 authors identified a real risk class and then communicated it in a way that erodes rather than enhances the audience's capacity to evaluate it. Simplified narrative. Manufactured certainty. Tribal activation via personalised fictional actors. Implied deference to credentialled insiders. Every one of these moves reduces the reader's capacity to think independently about the problem and increases dependence on the AI-2027 authors' framing.

In Paragentist terms the AI-2027 authors are operating in QIV (Advantage — their agency enhanced, the audience's eroded) at the expense of placing the audience in QII (Immoral Sacrifice — agency surrendered to the AI-2027 authors' preferred conclusion). The communication strategy produces compliance rather than agency. It gives readers a conclusion to adopt rather than a framework to reason with. An audience that has been emotionally primed to fear a specific scenario is not better equipped to navigate AI risk. It is more dependent on whoever manages the fear.

The deeper damage is to the epistemic commons. Once a risk category gets associated with a particular rhetorical style — overwrought, credentialled, institutionally self-serving — legitimate concern in that category gets discounted by association. The people most likely to dismiss AI risk entirely are the people who correctly identified the FUD structure and then incorrectly concluded that the risk class itself was manufactured. Those are often precisely the people whose cooperation is most needed.

The distraction hypothesis has a precise form that the evidence supports: the moral theatre of the rogue AI scenario functions — whether intentionally or not — to redirect public attention and institutional resources away from the mundane risks that are already materialising and toward the exotic risks that require the AI-2027 authors' specific expertise to address. Employment transition chaos does not need AI safety researchers. It needs labour economists, fiscal policy reform, and political will. Bad actor AI does not need interpretability research. It needs antitrust enforcement and democratic accountability mechanisms. None of these are the AI-2027 authors' domain. The safety research framing defines the problem in a way that makes their specific expertise the necessary solution. That is not a coincidence.

This does not require the AI-2027 authors to be consciously cynical. Motivated reasoning at institutional scale produces exactly this outcome without anyone choosing it deliberately. Researchers genuinely believe their domain is central. Funding structures reward that belief. The result is a field that defines civilisational AI risk in terms of the problems that field is equipped to solve. The Sickness applied to AI safety research itself — institutional self-preservation dressed as existential concern. The critique lands harder without villains. Easter Island didn't have any either.


Conclusion: What We Should Actually Do

The disappointing truth is not that the AI-2027 authors resorted to moral theatre. It is that they were probably right to. A carefully reasoned argument with honest uncertainty bounds, proper acknowledgment of competing incentives, and interventions selected for leverage rather than institutional authority would have reached fewer people and moved less institutional weight. That is not a criticism of the AI-2027 authors. It is a diagnosis of the audience — and of the institutions that shape what kinds of argument succeed in public discourse.

We have collectively failed to build institutions and audiences capable of acting on carefully reasoned probabilistic arguments about long-horizon risks. The theatre is a symptom of the epistemic failure, not the cause of it. And the epistemic failure is itself a risk, because it means that when the mundane scenarios — the transition catastrophe, the bad actor pathway, the Easter Island commons — become impossible to ignore, we will have spent the preceding years funding the wrong research, building the wrong frameworks, and deferring to the wrong experts.

The honest conclusions from this analysis are not comfortable for anyone.

For the AI development community: the science fiction doomsday is probably not your most likely failure mode. The bad actor with tame AI is. The employment transition is. The covert economic warfare is. These require political and economic interventions, not safety research, and you should be saying so rather than letting the safety framing absorb all the institutional attention.

For policymakers: you have already run this experiment with climate change, and the result should embarrass you. Decades of clear scientific consensus, visible and accelerating consequences, and democratic mandate produced inadequate action — because short electoral cycles, industry capture of the regulatory process, and the structural mismatch between four-year terms and forty-year problems defeated every governance mechanism available. The institutions failed not because the people in them were uniquely corrupt but because the incentive structure made failure the path of least resistance. AI governance is harder on every dimension. The technical complexity exceeds what most legislators can evaluate independently. The industry producing the campaign funding is newer, faster-growing, and more capable of regulatory capture than carbon ever was. The risk horizon is uncertain in ways that make deferral easy to justify. And the window for adequate response may close faster than the climate window did. If your track record on climate does not trouble you, you have not understood it. If it does trouble you, the question is whether you will act differently this time or repeat the performance while the stakes increase.

For readers: the goal of this article was not to replace one theatre with another. It was to model what the better version of the argument looks like — one where bias is declared, conclusions are reached by following evidence rather than confirming priors, and the honest answer is allowed to be uncomfortable for the person making the argument. The author changed their own position in the process of writing this. That is the point. Not the conclusion, but the method.

Preferably our leaders would act on careful critical reasoning rather than moral theatre. They probably will not. They have not earned the expectation that they will. That is a fact about our institutions, not a reason to stop making the argument.

The risks are real. The probability range is wide and the model cannot beat a coin toss. The gun is already cocked and cannot be put down. The question is not whether to engage with this but how — and the answer to that question is both more mundane and more urgent than the AI-2027 authors would like you to believe.