Source: Stratechery by Ben Thompson
Anthropic's decision to withhold its new model on safety grounds invites legitimate skepticism about competitive incentives dressed as caution. But the underlying problem is structural: if the company's concerns are genuine, the industry lacks adequate governance to manage increasingly dangerous capabilities. Anthropic is announcing that capabilities now exist that even their creators won't release—a threshold previous AI safety debates only theorized about. It exposes the inadequacy of both corporate self-regulation and current government oversight. Either Anthropic is exaggerating risks to sustain its safety narrative, or the AI industry has already produced systems it cannot safely deploy, and no one has a plan for what follows.