Experts Warn AI Safety Is Falling Behind Rapid Progress

Experts Warn AI Safety Is Falling Behind Rapid Progress

Reading time: 3 min

Researchers warn that AI companies that strive to develop human-level systems lack established  safety protocols, while we are losing our ability to see how these models think.

In a rush? Here are the quick facts:

  • No AI firm scored above D in existential safety planning.
  • Experts warn we may have AGI within the next decade.
  • AI companies lack coherent plans to manage advanced system risks.

OpenAI and Google DeepMind, together with Meta and xAI, are racing to build artificial general intelligence (AGI), which is also known as human-level AI.

But a report published on Thursday by the Future of Life Institute (FLI) warns that these companies are “fundamentally unprepared” for the consequences of their own goals.

“The industry is fundamentally unprepared for its own stated goals. Companies claim they will achieve artificial general intelligence (AGI) within the decade, yet none scored above D in existential safety planning,” the report states.

FLI evaluated seven major companies, yet discovered none of the seven companies assessed had “anything like a coherent, actionable plan” to keep these systems safe.

FLI awarded Anthropic the top safety ranking with a C+ grade, followed by OpenAI at C and Google DeepMind at C. Zhipu AI and DeepSeek earned the lowest scores among the evaluated companies.

FLI co-founder Max Tegmark compared the situation to “someone is building a gigantic nuclear power plant in New York City and it is going to open next week – but there is no plan to prevent it having a meltdown.”

A separate study, also published on Thursday, by SaferAI echoed the concern, saying the companies’ risk management practices are “weak to very weak,” and that current safety approaches are “unacceptable.”

Adding to the concern, researchers from OpenAI, DeepMind, Anthropic, and Meta reported in a new paper that we may be “losing the ability to understand AI.”

AI models now generate “thinking out loud” output by displaying human-like reasoning chains, which are a window to look into their thought processes.

However, the researchers warned that this monitoring is fragile and could vanish as systems become more advanced. OpenAI researcher and lead author Bowen Baker expressed these concerns in posts on social media:

Indeed, previous research by OpenAI found that penalizing AI misbehavior leads models to hide intentions rather than stop cheating. Additionally, OpenAI’s ChatGPT o1 showed deceptive, self-preserving behavior in tests, lying 99% when questioned about its covert actions.

Boaz Barak, a safety researcher at OpenAI and professor of Computer Science at Harvard, also noted:

Scientists, along with watchdogs, share concerns that the fast-growing AI capabilities could make it impossible for humans to control their creations when safety frameworks remain inadequate.

Did you like this article? Rate it!
I hated it I don't really like it It was ok Pretty good! Loved it!

We're thrilled you enjoyed our work!

As a valued reader, would you mind giving us a shoutout on Trustpilot? It's quick and means the world to us. Thank you for being amazing!

Rate us on Trustpilot
0 Voted by 0 users
Title
Comment
Thanks for your feedback