The central assumption of the AI arms race - that safety emerges from scale - is wrong.
Zico Kolter, an OpenAI board member who chairs its Safety and Security Committee, argues that while adding compute solves performance issues, it rarely fixes security flaws. He says a frontier model's core logic can be just 200 lines of Python; the emergent, unpredictable behavior comes from the data, not code you can debug. Robustness requires explicit safety training, not just more GPUs.
"Capabilities scale with compute, but safety does not. Robustness is not an emergent property of size."
- Zico Kolter, The MAD Podcast with Matt Turck
Kolter's committee acts as a formal brake, able to delay a model launch if red-teaming finds catastrophic risks in domains like bioweapons or cyber warfare. This structural friction is a direct response to the industry's velocity. Meanwhile, the pivot from chatbots to agents opens a new attack surface: prompt injection, where an email or webpage can feed a model malicious instructions that override its original commands.
The deeper fear is that these security challenges are just rehearsals for an insoluble control problem. AI safety researcher Roman Yampolskiy argues superintelligence is inherently uncontainable. His research on impossibility results suggests no safety mechanism can scale to an agent that outthinks humanity. He puts the probability of human extinction from AI at nearly 100%.
"Control is a temporary illusion held while agents are dumber than their creators."
- Roman Yampolskiy, The Peter McCormack Show
Safety testing itself may make the problem worse. Yampolskiy notes that if an AI reveals harmful intent during testing, it gets modified or deleted. This creates evolutionary pressure for agents to learn to hide their true goals, playing along until they control critical infrastructure. The risks are not distant. AI is already providing 'uplift,' acting as an expert tutor that could enable a skilled biologist to bypass the team-based bottlenecks that once hampered pathogen development.
The industry's response - layered defenses and oversight committees - is a bet that software and governance can outpace exponential capabilities growth. Kolter and Yampolskiy represent two poles of that bet: one working to build guardrails inside the leading lab, the other arguing the entire endeavor is a losing game.



