OpenAI's former chief scientist just raised $1bn for a new firm aimed at developing responsible AI
AI researcher Ilya Sutskever is the latest former OpenAI staffer to start their own firm focused on AI safety


OpenAI's former chief scientist has raised $1 billion for his new firm to develop safe artificial intelligence systems.
Ilya Sutskever co-founded Safe Superintelligence (SSI) in June after departing OpenAI in May following a failed attempt to oust CEO Sam Altman in November 2023, which was initially backed by Sutskever.
Reports suggest that the investment values SSI at $5 billion. The investors include Andreessen Horowitz, Sequoia Capital, DST Global, and SV Angel, SSI said, as well as NFDG, which is co-run by Daniel Gross, an SSI co-founder and the CEO.
So far, SSI has ten employees in Palo Alto, California and Tel Aviv, Israel. According to Reuters, SSI plans to spend the investment hiring top AI engineers and researchers, as well as on the necessary processing power. Both staff and computing are costly when it comes to developing AI.
Sutskever initially backed the efforts to oust Altman, which appeared to largely focus on the tension between AI safety and shipping usable AI products. However, amidst the chaos that ensued at the AI giant he swiftly u-turned,
"I deeply regret my participation in the board's actions. I never intended to harm OpenAI,” Sutskever said in a statement posted to X.
Sutskever announced his departure from OpenAI in May, saying at the time he was "confident that OpenAI will build AGI that is both safe and beneficial".
Get the ITPro daily newsletter
Sign up today and you will receive a free copy of our Future Focus 2025 report - the leading guidance on AI, cybersecurity and other IT challenges as per 700+ senior executives
But weeks later, in mid June, he announced the launch of safety-focused SSI, alongside Gross, who previously worked on AI at Apple, and Daniel Levy, also formerly of OpenAI.
Sutskever previously worked with "godfather of AI" Geoffrey Hinton, who stepped down from Google in May 2023 in order to more openly talk about the risks of artificial general intelligence (AGI) and super-intelligent AI.
SSI isn't the first company to emerge from OpenAI with a focus on safer AI. In 2021, Dario Amodei and his sister Daniela Amodei founded Anthropic to create safer AI after leaving the firm, with both reportedly concerned about the direction of the company.
Safe Superintelligence’s plans
SSI publicized its launch via a single website page with plain text on white background.
RELATED WHITEPAPER
"We have started the world’s first straight-shot SSI lab, with one goal and one product: a safe superintelligence," the company said at the time.
"We approach safety and capabilities in tandem, as technical problems to be solved through revolutionary engineering and scientific breakthroughs," the statement says. "We plan to advance capabilities as fast as possible while making sure our safety always remains ahead."
Gross said in an interview with Reuters not to expect a product for years — a contrast to companies like OpenAI that are pushing out marketable versions of AI to fund wider work on AGI.
"It's important for us to be surrounded by investors who understand, respect and support our mission, which is to make a straight shot to safe superintelligence and in particular to spend a couple of years doing R&D on our product before bringing it to market," Gross told Reuters.
Freelance journalist Nicole Kobie first started writing for ITPro in 2007, with bylines in New Scientist, Wired, PC Pro and many more.
Nicole the author of a book about the history of technology, The Long History of the Future.
-
Bigger salaries, more burnout: Is the CISO role in crisis?
In-depth CISOs are more stressed than ever before – but why is this and what can be done?
By Kate O'Flaherty Published
-
Cheap cyber crime kits can be bought on the dark web for less than $25
News Research from NordVPN shows phishing kits are now widely available on the dark web and via messaging apps like Telegram, and are often selling for less than $25.
By Emma Woollacott Published
-
OpenAI woos UK government amid consultation on AI training and copyright
News OpenAI is fighting back against the UK government's proposals on how to handle AI training and copyright.
By Emma Woollacott Published
-
DeepSeek and Anthropic have a long way to go to catch ChatGPT: OpenAI's flagship chatbot is still far and away the most popular AI tool in offices globally
News ChatGPT remains the most popular AI tool among office workers globally, research shows, despite a rising number of competitor options available to users.
By Ross Kelly Published
-
‘DIY’ agent platforms are big tech’s latest gambit to drive AI adoption
Analysis The rise of 'DIY' agentic AI development platforms could enable big tech providers to drive AI adoption rates.
By George Fitzmaurice Published
-
OpenAI wants to simplify how developers build AI agents
News OpenAI is releasing a set of tools and APIs designed to simplify agentic AI development in enterprises, the firm has revealed.
By George Fitzmaurice Published
-
Elon Musk’s $97 billion flustered OpenAI – now it’s introducing rules to ward off future interest
News OpenAI is considering restructuring the board of its non-profit arm to ward off unwanted bids after Elon Musk offered $97.4bn for the company.
By Nicole Kobie Published
-
Sam Altman says ‘no thank you’ to Musk's $97bn bid for OpenAI
News OpenAI has rejected a $97.4 billion buyout bid by a consortium led by Elon Musk.
By Nicole Kobie Published
-
DeepSeek flips the script
ITPro Podcast The Chinese startup's efficiency gains could undermine compute demands from the biggest names in tech
By Rory Bathgate Published
-
SoftBank could take major stake in OpenAI
News Reports suggest the firm is planning to increase its stake in the ChatGPT maker
By Emma Woollacott Published