Sopa Images | Lightrocket | Getty Images
“Contemporary AI systems are now becoming human-competitive at general tasks, and we must ask ourselves: Should we let machines flood our information channels with propaganda and untruth?” the letter read.
“Should we automate away all the jobs, including the fulfilling ones? Should we develop nonhuman minds that might eventually outnumber, outsmart, obsolete and replace us? Should we risk loss of control of our civilization?”
The letter added: “Such decisions must not be delegated to unelected tech leaders.”
The Future of Life Institute is a nonprofit organization based in Cambridge, Massachusetts, that campaigns for the responsible and ethical development of artificial intelligence. Its founders include MIT cosmologist Max Tegmark and Skype co-founder Jaan Tallinn.
The institute has previously gotten the likes of Musk and Google-owned AI lab DeepMind to promise never to develop lethal autonomous weapons systems.
The institute said it was calling on all AI labs to “immediately pause for at least 6 months the training of AI systems more powerful than GPT-4.”
GPT-4, which was released earlier this month, is thought to be far more advanced than its predecessor GPT-3.
“If such a pause cannot be enacted quickly, governments should step in and institute a moratorium,” it added.
ChatGPT, the viral AI chatbot, has stunned researchers with its ability to produce humanlike responses to user prompts. By January, ChatGPT had amassed 100 million monthly active users only two months into its launch, making it the fastest-growing consumer application in history.
The technology is trained on huge amounts of data from the internet, and has been used to create everything from poetry in the style of William Shakespeare to drafting legal opinions on court cases.
But AI ethicists have also raised concerns with potential abuses of the technology, such as plagiarism and misinformation.
Musk has previously said he thinks AI represents one of the “biggest risks” to civilization.
The Tesla and SpaceX CEO is also one of the co-founders of OpenAI, though he left OpenAI’s board in 2018 and no longer holds a stake in the company.
He has criticized the organization a number of times recently, saying he believes it is diverging from its original purpose.
Regulators are also racing to get a handle on AI tools as the technology is advancing at a rapid pace. On Wednesday, the U.K. government published a white paper on AI, deferring to different regulators to supervise the use of AI tools in their respective sectors by applying existing laws.