Can large language models democratize access to dual-use biotechnology?
- URL: http://arxiv.org/abs/2306.03809v1
- Date: Tue, 6 Jun 2023 15:52:05 GMT
- Title: Can large language models democratize access to dual-use biotechnology?
- Authors: Emily H. Soice, Rafael Rocha, Kimberlee Cordova, Michael Specter, and
Kevin M. Esvelt
- Abstract summary: Large language models (LLMs) are accelerating and democratizing research.
These models may also confer easy access to dual-use technologies capable of inflicting great harm.
To evaluate this risk, the 'Safeguarding the Future' course at MIT tasked non-scientist students with investigating.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Large language models (LLMs) such as those embedded in 'chatbots' are
accelerating and democratizing research by providing comprehensible information
and expertise from many different fields. However, these models may also confer
easy access to dual-use technologies capable of inflicting great harm. To
evaluate this risk, the 'Safeguarding the Future' course at MIT tasked
non-scientist students with investigating whether LLM chatbots could be
prompted to assist non-experts in causing a pandemic. In one hour, the chatbots
suggested four potential pandemic pathogens, explained how they can be
generated from synthetic DNA using reverse genetics, supplied the names of DNA
synthesis companies unlikely to screen orders, identified detailed protocols
and how to troubleshoot them, and recommended that anyone lacking the skills to
perform reverse genetics engage a core facility or contract research
organization. Collectively, these results suggest that LLMs will make
pandemic-class agents widely accessible as soon as they are credibly
identified, even to people with little or no laboratory training. Promising
nonproliferation measures include pre-release evaluations of LLMs by third
parties, curating training datasets to remove harmful concepts, and verifiably
screening all DNA generated by synthesis providers or used by contract research
organizations and robotic cloud laboratories to engineer organisms or viruses.
Related papers
- General-purpose AI models can generate actionable knowledge on agroecological crop protection [0.0]
We verify scientific knowledge on agroecological crop protection generated by web-grounded or non-grounded large language models.<n>DeepSeek consistently screened a 4.8-49.7-fold larger literature corpus and reported 1.6-2.4-fold more biological control agents or management solutions than ChatGPT.
arXiv Detail & Related papers (2025-12-12T11:17:13Z) - Can an Individual Manipulate the Collective Decisions of Multi-Agents? [53.01767232004823]
M-Spoiler is a framework that simulates agent interactions within a multi-agent system to generate adversarial samples.<n>M-Spoiler introduces a stubborn agent that actively aids in optimizing adversarial samples.<n>Our findings confirm the risks posed by the knowledge of an individual agent in multi-agent systems.
arXiv Detail & Related papers (2025-09-20T01:54:20Z) - Large Language Models Transform Organic Synthesis From Reaction Prediction to Automation [3.904238958136483]
Large language models (LLMs) are beginning to reshape how chemists plan and run reactions in organic synthesis.<n>LLMs can propose synthetic routes, forecast reaction outcomes and instruct robots that execute experiments without human supervision.<n>We show how coupling LLMs with graph neural networks, quantum calculations and real-time spectroscopy shrinks discovery cycles and supports greener, data-driven chemistry.
arXiv Detail & Related papers (2025-08-07T14:17:23Z) - BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments [8.317138109309967]
Large language models (LLMs) and vision-language models (VLMs) have the potential to transform biological research by enabling autonomous experimentation.<n>Here we introduce BioMARS, an intelligent platform that integrates LLMs, VLMs, and modular robotics to autonomously design, plan, and execute biological experiments.<n>A web interface enables real-time human-AI collaboration, while a modular backend allows scalable integration with laboratory hardware.
arXiv Detail & Related papers (2025-07-02T08:47:02Z) - AIGS: Generating Science from AI-Powered Automated Falsification [17.50867181053229]
We propose Baby-AIGS as a baby-step demonstration of a full-process AIGS system, which is a multi-agent system with agents in roles representing key research process.
Experiments on three tasks preliminarily show that Baby-AIGS could produce meaningful scientific discoveries, though not on par with experienced human researchers.
arXiv Detail & Related papers (2024-11-17T13:40:35Z) - Poison Attacks and Adversarial Prompts Against an Informed University Virtual Assistant [3.0874677990361246]
Large language models (LLMs) are particularly vulnerable to adversarial attacks.
The rapid development pace of AI-based systems is being driven by the potential of Generative AI (GenAI) to assist humans in decision making.
A threat actor can use security gaps, poor safeguards, and limited data governance to carry out attacks that grant unauthorized access to the system and its data.
arXiv Detail & Related papers (2024-11-03T05:34:38Z) - Many Heads Are Better Than One: Improved Scientific Idea Generation by A LLM-Based Multi-Agent System [62.832818186789545]
Virtual Scientists (VirSci) is a multi-agent system designed to mimic the teamwork inherent in scientific research.
VirSci organizes a team of agents to collaboratively generate, evaluate, and refine research ideas.
We show that this multi-agent approach outperforms the state-of-the-art method in producing novel scientific ideas.
arXiv Detail & Related papers (2024-10-12T07:16:22Z) - BioDiscoveryAgent: An AI Agent for Designing Genetic Perturbation Experiments [112.25067497985447]
We introduce BioDiscoveryAgent, an agent that designs new experiments, reasons about their outcomes, and efficiently navigates the hypothesis space to reach desired solutions.
BioDiscoveryAgent can uniquely design new experiments without the need to train a machine learning model.
It achieves an average of 21% improvement in predicting relevant genetic perturbations across six datasets.
arXiv Detail & Related papers (2024-05-27T19:57:17Z) - Large language models surpass human experts in predicting neuroscience results [60.26891446026707]
Large language models (LLMs) forecast novel results better than human experts.
BrainBench is a benchmark for predicting neuroscience results.
Our approach is not neuroscience-specific and is transferable to other knowledge-intensive endeavors.
arXiv Detail & Related papers (2024-03-04T15:27:59Z) - An Evaluation of Large Language Models in Bioinformatics Research [52.100233156012756]
We study the performance of large language models (LLMs) on a wide spectrum of crucial bioinformatics tasks.
These tasks include the identification of potential coding regions, extraction of named entities for genes and proteins, detection of antimicrobial and anti-cancer peptides, molecular optimization, and resolution of educational bioinformatics problems.
Our findings indicate that, given appropriate prompts, LLMs like GPT variants can successfully handle most of these tasks.
arXiv Detail & Related papers (2024-02-21T11:27:31Z) - The Wolf Within: Covert Injection of Malice into MLLM Societies via an MLLM Operative [55.08395463562242]
Multimodal Large Language Models (MLLMs) are constantly defining the new boundary of Artificial General Intelligence (AGI)
Our paper explores a novel vulnerability in MLLM societies - the indirect propagation of malicious content.
arXiv Detail & Related papers (2024-02-20T23:08:21Z) - What Does the Bot Say? Opportunities and Risks of Large Language Models in Social Media Bot Detection [48.572932773403274]
We investigate the opportunities and risks of large language models in social bot detection.
We propose a mixture-of-heterogeneous-experts framework to divide and conquer diverse user information modalities.
Experiments show that instruction tuning on 1,000 annotated examples produces specialized LLMs that outperform state-of-the-art baselines.
arXiv Detail & Related papers (2024-02-01T06:21:19Z) - Artificial intelligence and biological misuse: Differentiating risks of
language models and biological design tools [0.0]
This article differentiates two classes of AI tools that could pose such biosecurity risks: large language models (LLMs) and biological design tools (BDTs)
arXiv Detail & Related papers (2023-06-24T12:48:49Z) - Synthetic Misinformers: Generating and Combating Multimodal
Misinformation [11.696058634552147]
multimodal misinformation detection (MMD) detects whether the combination of an image and its accompanying text could mislead or misinform.
We show that our proposed CLIP-based Named Entity Swapping can lead to MMD models that surpass other OOC and NEI Misinformers in terms of multimodal accuracy.
arXiv Detail & Related papers (2023-03-02T12:59:01Z) - Deep metric learning improves lab of origin prediction of genetically
engineered plasmids [63.05016513788047]
Genetic engineering attribution (GEA) is the ability to make sequence-lab associations.
We propose a method, based on metric learning, that ranks the most likely labs-of-origin.
We are able to extract key signatures in plasmid sequences for particular labs, allowing for an interpretable examination of the model's outputs.
arXiv Detail & Related papers (2021-11-24T16:29:03Z) - Privacy-preserving Artificial Intelligence Techniques in Biomedicine [3.908261721108553]
Training an AI model on sensitive data raises concerns about the privacy of individual participants.
This paper provides a structured overview of advances in privacy-preserving AI techniques in biomedicine.
It places the most important state-of-the-art approaches within a unified taxonomy and discusses their strengths, limitations, and open problems.
arXiv Detail & Related papers (2020-07-22T18:35:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.