Log in

OpenAI founder Sutskever sets up new AI company devoted to "safe superintelligence"

Posted 6/20/24

The Associated Press (AP) — Ilya Sutskever, one of the founders of OpenAI who was involved in a failed effort to push out CEO Sam Altman, said he's starting a safety-focused artificial intelligence …

You must be a member to read this story.

Join our family of readers for as little as $5 per month and support local, unbiased journalism.


Already have an account? Log in to continue.

Current print subscribers can create a free account by clicking here

Otherwise, follow the link below to join.

To Our Valued Readers –

Visitors to our website will be limited to five stories per month unless they opt to subscribe. The five stories do not include our exclusive content written by our journalists.

For $6.99, less than 20 cents a day, digital subscribers will receive unlimited access to YourValley.net, including exclusive content from our newsroom and access to our Daily Independent e-edition.

Our commitment to balanced, fair reporting and local coverage provides insight and perspective not found anywhere else.

Your financial commitment will help to preserve the kind of honest journalism produced by our reporters and editors. We trust you agree that independent journalism is an essential component of our democracy. Please click here to subscribe.

Sincerely,
Charlene Bisson, Publisher, Independent Newsmedia

Please log in to continue

Log in
I am anchor

OpenAI founder Sutskever sets up new AI company devoted to "safe superintelligence"

Posted

The Associated Press (AP) — Ilya Sutskever, one of the founders of OpenAI who was involved in a failed effort to push out CEO Sam Altman, said he's starting a safety-focused artificial intelligence company.

Sutskever, a respected AI researcher who left the ChatGPT maker last month, said in a social media post Wednesday that he's created Safe Superintelligence Inc. with two co-founders. The company's only goal and focus is safely developing "superintelligence” - a reference to AI systems that are smarter than humans.

The company vowed not to be distracted by “management overhead or product cycles,” and under its business model, work on safety and security would be “insulated from short-term commercial pressures,” Sutskever and his co-founders Daniel Gross and Daniel Levy said in a prepared statement.

The three said Safe Superintelligence is an American company with roots in Palo Alto, California, and Tel Aviv, “where we have deep roots and the ability to recruit top technical talent.”

Sutskever was part of a group that made an unsuccessful attempt last year to oust Altman. The boardroom shakeup, which Sutskever later said he regretted, also led to a period of internal turmoil centered on whether leaders at OpenAI were prioritizing business opportunities over AI safety.

At OpenAI, Sutskever jointly led a team focused on safely developing better-than-human AI known as artificial general intelligence, or AGI. When he left OpenAI, he said that he had plans for a “very personally meaningful” project, but offered no details.

Sutskever said that it was his choice to leave OpenAI.

Days after his departure, his team co-leader Jan Leike also resigned and leveled criticism at OpenAI for letting safety “take a backseat to shiny products.” OpenAI later announced the formation of a safety and security committee, but it's been filled mainly with company insiders.

___

The Associated Press and OpenAI have a licensing and technology agreement that allows OpenAI access to part of the AP’s text archives.