Jan Leike, a number one AI researcher who earlier this month resigned from OpenAI earlier than publicly criticizing the corporate’s strategy to AI security, has joined OpenAI rival Anthropic to guide a brand new “superalignment” group.
In a put up on X, Leike stated that his group at Anthropic will give attention to numerous facets of AI security and safety, particularly “scalable oversight,” “weak-to-strong generalization” and automatic alignment analysis.
A supply acquainted with the matter tells cryptonoiz that Leike will report on to Jared Kaplan, Anthropic’s chief science officer, and that Anthropic researchers at the moment engaged on scalable oversight — strategies to manage large-scale AI’s conduct in predictable and fascinating methods — will transfer to report back to Leike as Leike’s group spins up.
In some ways, Leike’s group sounds related in mission to OpenAI’s lately dissolved Superalignment group. The Superalignment group, which Leike co-led, had the bold aim of fixing the core technical challenges of controlling superintelligent AI within the subsequent 4 years, however usually discovered itself hamstrung by OpenAI’s management.
Anthropic has usually tried to place itself as extra safety-focused than OpenAI.
Anthropic’s CEO, Dario Amodei, was as soon as the VP of analysis at OpenAI and reportedly cut up with OpenAI after a disagreement over the corporate’s course — particularly OpenAI’s rising industrial focus. Amodei introduced with him a variety of ex-OpenAI staff to launch Anthropic, together with OpenAI’s former coverage lead Jack Clark.