AI poses human extinction risk, Sam Altman and other tech leaders warn

The Microsoft Bing App is seen running on an iPhone in this photo illustration on May 30, 2023 in Warsaw, Poland. (Photo by Jaap Arriens/NurPhoto via Getty Images)

Jaap Arriens | Nurphoto | GettyImages

Artificial intelligence may lead to human extinction and reduce the risks associated with the technology stated should be a global priority, industry experts and tech leaders in an open letter.

“Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war,” the statement read on Tuesday.

related investing news


Sam Altman, CEO of ChatGPT-maker OpenAI, as well as executives from GoogleDeepMind’s AI arm Microsoft were among those who supported and signed the short statement from the Center for AI Safety.

The technology has gathered pace in recent months after chatbot ChatGPT was released for public use in November and subsequently went viral. In just two months after its launch, it reached 100 million users. ChatGPT has amazed researchers and the general public with its ability to generate humanlike responses to users’ prompts, suggesting that AI could replace jobs and imitate humans.

The statement Tuesday said that there has been increasing discussion about a “broad spectrum of important and urgent risks from AI.”

Read more about tech and crypto from CNBC Pro

But it said it could be “difficult to voice concerns about some of advanced AI’s most severe risks” and had the aim of overcoming this obstacle and opening up the discussions.

ChatGPT has arguably sparked much more awareness and adoption of AI as major firms around the world have raced to develop competitive products and capabilities.

Altman had admitted in March that he is a “little bit scared” of AI as he worried that authoritarian governments would develop the technology. Other tech leaders such as Tesla’s Elon Musk and former Google CEO Eric Schmidt has warned about the risks AI poses to society.

In an open letter in March, Musk, Apple co-founder Steve Wozniak and several tech leaders urged AI labs to stop training systems to be more powerful than GPT-4 — which is OpenAI’s latest large language model. They also called for a six-month pause on such advanced development.

AI hits the stock market in a big way as earnings give clarity

“Contemporary AI systems are now becoming human-competitive at general tasks,” said the letter.

“Should we automate away all the jobs, including the fulfilling ones? Should we develop nonhuman minds that might eventually outnumber, outsmart, obsolete and replace us? Should we risk losing control of our civilization?” the letter asked.

Last week, Schmidt also separately warned about the “existential risks” associated with AI as the technology advances.

Denial of responsibility! is an automatic aggregator of the all world’s media. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials, please contact us by email – The content will be deleted within 24 hours.
Andrew Naughtie

News reporter and author at @websalespromo