Comment by Jan Leike

Former head of alignment at OpenAI; now VP of safety at Anthropic
Alignment is not solved, but it increasingly looks solvable. [...] Since I first wrote about it in 2022, pretraining has continued improving and reinforcement learning has become much more significant — and our techniques are keeping pace. Unverified source (2026)
Like Share on X 3h ago
Policy proposals and claims
replying to Jan Leike