OpenAI’s Former Chief Scientist Launches Safe Superintelligence Inc.
In a move that could reshape the AI industry, Ilya Sutskever, former co-founder and chief scientist of OpenAI, has announced the launch of a new company focused solely on AI safety. The startup, named Safe Superintelligence Inc. (SSI), aims to create a safe and powerful AI system.
One Goal, One Product
SSI is built with a clear mission in mind. As Sutskever stated in a Wednesday post, SSI has “one goal and one product” which is to develop an AI that is both safe and efficient. This singular focus helps SSI to sidestep the distractions and pressures often faced by AI teams at giants like OpenAI, Google, and Microsoft.
"We find that external pressures can often sidetrack progress. Our singular focus lets us avoid distractions from management overhead or product cycles," Sutskever explains. "Our business model ensures that safety, security, and progress are insulated from short-term commercial pressures. This way, we can scale in peace."
The Team Behind SSI
SSI is not just the brainchild of Sutskever. He is co-founding the company with Daniel Gross, a former AI lead at Apple, and Daniel Levy, who previously worked as a technical staff member at OpenAI. This experienced team is expected to bring significant advancements in safe AI development.
Background and Context
The creation of SSI follows a series of events at OpenAI. Last year, Sutskever led a push to remove OpenAI's CEO, Sam Altman. Following his departure from OpenAI in May, there were rumblings of a new venture.
The timing appears fortuitous, as other key figures from OpenAI cited safety concerns during their departures. AI researcher Jan Leike and policy researcher Gretchen Krueger both pointed to safety processes taking a backseat to more marketable products.
Focus on Safe Superintelligence
While OpenAI continues to grow its partnerships with tech giants like Apple and Microsoft, SSI is taking a different route. During an interview, Sutskever emphasized that SSI’s first and only product for the foreseeable future will be safe superintelligence. "We will not do anything else until then," he elaborated.
Conclusion
As the AI landscape continues to evolve, Safe Superintelligence Inc. promises to carve out a niche focused on prioritizing AI safety without the commercial distractions. This approach might well set a new standard for how AI innovations are pursued responsibly.
Stay tuned for further updates as SSI progresses toward its ambitious goal of creating a safe and powerful AI system.