Stay up to date with notifications from The Independent

Notifications can be managed in browser preferences.

AI can help generate synthetic viruses and spark pandemics, warns former Google executive

Google DeepMind co-founder discusses ‘darkest scenario’ where people ‘will experiment with pathogens’

Vishwam Sankaran
Thursday 07 September 2023 08:20 BST
Comments
Related video- ChatGPT: Schools put new policies in place

Synthetic viruses could be generated through the misuse of artificial intelligence and potentially spark pandemics, a former Google executive and AI expert has warned.

Google DeepMind co-founder Mustafa Suleyman expressed concern that the use of AI to engineer pathogens to cause more harm may lead to a scenario like a pandemic.

“The darkest scenario is that people will experiment with pathogens, engineered synthetic pathogens that might end up accidentally or intentionally being more transmissible or more lethal,” he said in a recent episode of a podcast.

Similar to how there are restrictions in place to prevent people from easily accessing pathogenic microbes like anthrax, Mr Suleyman has called for the means to restrict access to advanced AI technology and software that runs such models.

“That’s where we need containment. We have to limit access to the tools and the know-how to carry out that kind of experimentation,” he said in The Diary of a CEO podcast.

“We can’t let just anyone have access to them. We need to limit who can use the AI software, the cloud systems, and even some of the biological material,” the Google DeepMind co-founder said.

“And of course on the biology side it means restricting access to some of the substances,” he said, adding that AI development needs to be approached with a “precautionary principle”.

Mr Suleyman’s statements echo concerns raised in a recent study that even undergraduates with no relevant background in biology can detail suggestions for bio-weapons from AI systems.

Researchers, including those from the Massachusetts Institute of Technology, found chatbots can suggest “four potential pandemic pathogens” within an hour and explain how they can be generated from synthetic DNA.

The research found chatbots also “supplied the names of DNA synthesis companies unlikely to screen orders, identified detailed protocols and how to troubleshoot them, and recommended that anyone lacking the skills to perform reverse genetics engage a core facility or contract research organization”.

Such large language models (LLMs), like ChatGPT, “will make pandemic-class agents widely accessible as soon as they are credibly identified, even to people with little or no laboratory training,” the study said.

The study, whose authors included MIT bio risk expert Kevin Esvelt, called for “non-proliferation measures”.

Such measures could include “pre-release evaluations of LLMs by third parties, curating training datasets to remove harmful concepts, and verifiably screening all DNA generated by synthesis providers or used by contract research organizations and robotic ‘cloud laboratories’ to engineer organisms or viruses”.

Join our commenting forum

Join thought-provoking conversations, follow other Independent readers and see their replies

Comments

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in