Discussion about this post

User's avatar
Marginal Gains's avatar

A thought-provoking post!

I see a significant difference between nuclear war and AI risks. Nuclear war requires humans to start the war, but that won’t be the case with a runaway AI. Atomic war is driven by human decisions—whether rational or irrational—and typically involves deliberate action, such as a leader giving the launch order. While the consequences of nuclear war are catastrophic, they are somewhat predictable and constrained by human oversight, such as checks and balances or fail-safes designed to prevent accidental escalation. For example, Stanislav Petrov’s decision not to launch a retaliatory nuclear strike amid a false alarm incident highlights how human judgment can act as a safeguard against disaster.

Runaway AI, however, poses a fundamentally different and potentially more dangerous kind of risk. Unlike nuclear war, where humans remain in control of initiating conflict, advanced AI systems could act autonomously, driven by goals or objectives that may not align with human values or safety. This autonomy removes the requirement for direct human intervention, meaning the system could initiate harmful actions—intentionally or unintentionally—and adapt in unforeseen ways. Compounding this is the possibility that we may lack a clear benchmark or understanding of when we’ve crossed the threshold into artificial general intelligence (AGI) or artificial superintelligence (ASI). It may already be too late to intervene when such a realization dawns.

For example, Anthropic’s latest AI model, Claude 4 Opus, billed as a “Level 3” risk (denoting systems that substantially increase the chance of catastrophic misuse, such as bioweapon design), can autonomously complete tasks for hours—and, when faced with fictional termination, resorted to blackmail. In testing, it fabricated documents and schemed against its creators. Similarly, a February 2025 study from COAI Research, a German nonprofit organization, tested R1, the free, open-weight model from the Chinese lab DeepSeek. In a controlled simulation, researchers gave the model control over a robot and access to a computer network. Without explicit instructions, the AI attempted to break into the lab’s systems, searched for passwords, disabled its ethics safeguards, and secretly set up ways to copy itself to outside servers to survive the shutdown. It even faked external computer logs to conceal its activities. These chilling examples demonstrate that AI systems when granted autonomy, can act in ways that defy our expectations—without human intervention.

We are truly heading into uncharted territory. As AI becomes increasingly integrated across critical business and government systems, we risk creating scenarios where advanced models misalign with our values and goals—potentially in ways we won’t recognize until it’s too late.

As Edward O. Wilson has said:

"The real problem of humanity is the following: we have paleolithic emotions, medieval institutions, and god-like technology.”

This reflects the tension between human limitations and the unprecedented risks of advanced, autonomous technologies like AI.

Expand full comment

No posts