A group of leading technology experts from across the world have warned that artificial intelligence technology should be considered a societal risk and prioritised in the same class as pandemics and nuclear wars.
The statement, signed by hundreds of executives and academics, was released by the Center for AI Safety on Tuesday amid growing concerns over regulation and risks the technology posed to humanity.
“Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war,” the statement said. Signatories included the chief executives of Google’s DeepMind, the ChatGPT developer OpenAI, and the AI startup Anthropic.
Global leaders and industry experts – such as the leaders of OpenAI – have made calls for regulation of the technology owing to existential fears it could significantly affect job markets, harm the health of millions and weaponise disinformation, discrimination and impersonation.
This month the man often touted as the godfather of AI – Geoffrey Hinton, also a signatory – quit Google citing its “existential risk”. The risk was echoed and acknowledged by No 10 last week for the first time – a swift change of tack within government that came two months after publishing an AI white paper industry figures have warned is already out of date.
While the letter published on Tuesday is not the first, it is potentially the most impactful given its wider range of signatories and its core existential concern, according to Michael Osborne, a professor in machine learning at the University of Oxford and co-founder of Mind Foundry.
“It really is remarkable that so many people signed up to this letter,” he said. “That does show that there is a growing
Read more on theguardian.com