The Microsoft Bing App is seen running on an iPhone in this photo illustration on 30 May, 2023 in Warsaw, Poland. (Photo by Jaap Arriens/NurPhoto via Getty Images)
Jaap Arriens | Nurphoto | Getty Images
Artificial intelligence may lead to human extinction and reducing the risks associated with the technology should be a global priority, industry experts and tech leaders stated in an
3 hours ago
2 days ago
Sam Altman, CEO of ChatGPT-maker OpenAI, as well as executives from Google‘s AI arm DeepMind and Microsoft were among those who supported and signed the short statement from the Center for AI Safety.
The technology has gathered pace in recent months after chatbot ChatGPT was released for public use in November and subsequently went viral. In just two months after its launch, it reached 100 million users. ChatGPT has amazed researchers and the general public with its ability to generate humanlike responses to users’ prompts, suggesting that AI could replace jobs and imitate humans.
The statement Tuesday said that there has been increasing discussion about a “broad spectrum of important and urgent risks from AI.”
Read more about tech and crypto from CNBC Pro
But it said it can be “difficult to voice concerns about some of advanced AI’s most severe risks” and had the aim of overcoming this obstacle and opening up the discussions.
ChatGPT has arguably sparked much more awareness and adoption of AI as major firms around the world have raced to develop rival products and capabilities.
Altman had admitted in March that he is a “little bit scared” of AI as he worries that authoritarian governments would develop the technology. Other tech leaders such as Tesla’s Elon Musk and formerGoogle CEO Eric Schmidt have cautioned about the risks AI poses to society.
“Contemporary AI systems are now becoming human-competitive at general tasks,” said the letter.
“Should we automate away all the jobs, including the fulfilling ones? Should we develop nonhuman minds that might eventually outnumber, outsmart, obsolete and replace us? Should we risk loss of control of our civilization?” the letter asked.