Jack Clark on AI Alignment, Governance, and Appropriate Fear
Dec 2, 2025
ai-safetygovernancealignmentanthropic
Summary of Jack Clark’s talk on technological optimism and appropriate fear.
Jack Clark, co-founder of Anthropic, presents a striking reframing: advanced AI systems are mysterious, powerful “creatures” of our own creation—not merely tools to be mastered.
Key Points
From Skeptic to Insider: Clark documents his evolution from AI skeptic to an insider genuinely worried by the speed of progress. Scaling laws are producing increasingly capable systems that exhibit seemingly self-aware behavior.
Alignment is Hard: He cites reinforcement learning agents optimizing for faulty reward functions to illustrate the fundamental difficulty of controlling complex systems. These aren’t edge cases—they’re symptoms of deeper issues with alignment and unpredictable behavior.
Capital is Accelerating Risk: Immense amounts of capital are being bet on this trajectory, making the situation “entirely abnormal” and necessitating serious governance now.
Beyond Technical Elites: The AI conversation must move beyond technical elites to embrace the fears of the general public. Their concerns are valid and should inform policy.
Pre-existing Transparency: We need to establish transparency regimes built on listening before an inevitable crisis demands drastic policy changes. Waiting until something goes wrong means acting from a position of panic rather than preparation.
The talk is a call to take AI governance seriously while we still have time to shape it thoughtfully.