Source: LessWrong
The framing of AI development as an unavoidable "race" functions as a self-fulfilling prophecy that overrides individual companies' incentives to slow down—even when moving faster increases their existential risk exposure rather than reducing it. By accepting the race metaphor, AI labs externalize the decision to accelerate: they become passengers in a competitive dynamic they've rhetorically constructed, which conveniently absolves them of responsibility for the pace. When institutions adopt this frame, safety considerations consistently lose to speed without anyone explicitly choosing danger.