Key Takeaways
Microsoft researchers have uncovered a troubling weakness in the global biosecurity infrastructure designed to prevent the creation of dangerous biological weapons, revealing that artificial intelligence can be weaponized to bypass safety systems meant to detect deadly toxins and pathogens.
In a peer-reviewed study published in the journal Science on October 2, a team led by Eric Horvitz, Microsoft's Chief Scientific Officer, demonstrated how generative AI tools could redesign known toxic proteins to slip past commercial DNA screening software while potentially maintaining their lethal properties.
The research, which Microsoft calls the "Paraphrase Project," represents what the company describes as a biological "zero-day vulnerability"—borrowing terminology from cybersecurity to describe a previously unknown flaw that defenders had no opportunity to address.
The biological arms race accelerates
The vulnerability exploits a fundamental limitation in how DNA synthesis companies screen orders.
Current biosecurity software compares customer requests against databases of known dangerous sequences, flagging close matches for review.
Using several AI protein design models, including Microsoft's own EvoDiff platform, researchers found they could alter toxic protein structures just enough to evade detection while computer models predicted the proteins would retain their harmful function.
"The diversified proteins essentially flew through the screening techniques," Horvitz told Nature magazine.
Working in collaboration with the International Gene Synthesis Consortium and major DNA manufacturers, including Twist Bioscience and Integrated DNA Technologies, Microsoft conducted the research entirely digitally, never producing any physical toxic proteins.
Before publication, the team alerted the U.S. government and screening software developers, who deployed patches to their systems.
However, Adam Clore, director of technology research and development at Integrated DNA Technologies and a study co-author, emphasized the limitations of the current fix.
"The patch is incomplete, and the state of the art is changing," Clore said. "But this isn't a one-and-done thing. It's the start of even more testing. We're in something of an arms race."
Even after the patch was applied, the improved screening tools still failed to flag approximately 3% of malicious sequences in follow-up testing, according to Nature's coverage of the research.
Industry and government response
The timing of the discovery comes amid growing concern about AI's dual-use potential in biotechnology.
The same generative protein design tools driving breakthrough medical research—including those developed by 2024 Nobel Prize winner David Baker at the University of Washington could theoretically be repurposed for harmful applications.
Emily Leproust, CEO and co-founder of Twist Bioscience, acknowledged the challenge. "For known proteins and sequences, industry best practices for biosecurity screening are robust and highly effective.
However, as AI capabilities evolve, screening practices must evolve just as quickly," Leproust said. "Recognizing the critical role that we play in the advancement of our customers' research, together with Microsoft, we are looking around the corner to identify and guide industry next steps for impactful drug discovery while advancing science responsibly."
Dean Ball, a fellow at the Foundation for American Innovation, argued the findings demonstrate an urgent need for stronger oversight.
"This finding, combined with rapid advances in AI-enabled biological modeling, demonstrates the clear and urgent need for enhanced nucleic acid synthesis screening procedures coupled with a reliable enforcement and verification mechanism," Ball said.
DNA order screening is already considered a cornerstone of U.S. biosecurity policy.
In May 2024, President Trump signed an executive order calling for a comprehensive overhaul of biological research safety measures, though new federal guidelines have yet to be released.
The defense point debate
The research has ignited fierce debate about where biosecurity defenses should be concentrated. Horvitz said Microsoft set out to answer two key questions: whether AI protein design tools could be used to redesign toxins while preserving lethality, and whether screening systems could be patched to detect such threats.
"Thanks to the study and efforts of dedicated collaborators, we can now say yes," Horvitz stated in Microsoft's official blog post about the research.
However, not all experts are convinced that DNA synthesis screening represents the optimal defense strategy.
Michael Cohen, an AI safety researcher at the University of California, Berkeley, criticized both the research methodology and the reliance on commercial DNA manufacturers as a choke point.
"The challenge appears weak, and their patched tools fail a lot," Cohen said. "There seems to be an unwillingness to admit that sometime soon, we're going to have to retreat from this supposed choke point."
Cohen advocates for building biosecurity protections directly into AI systems themselves, limiting the type of information these models can provide rather than depending on downstream screening by DNA vendors.
Practical defense or false security?
Clore defended the focus on DNA synthesis screening, noting the industry's concentrated structure makes it a practical point of intervention.
Only a handful of major companies dominate DNA synthesis in the United States, and they already maintain close working relationships with government agencies.
"You can't put that genie back in the bottle," Clore said, referring to the widespread availability of AI technology. "If you have the resources to try to trick us into making a DNA sequence, you can probably train a large language model."
James Diggans, head of policy and biosecurity at Twist Bioscience and chair of the board at the International Gene Synthesis Consortium, provided some reassurance about the current threat level.
According to NPR's coverage, Twist has had to refer orders to law enforcement fewer than five times in the past decade.
Looking ahead
The Microsoft study deliberately withheld certain details to prevent misuse, including which specific toxic proteins were used in testing.
However, the researchers noted that dangerous proteins like ricin—derived from castor beans—and infectious prions linked to mad cow disease are already well-documented threats in scientific literature.
As AI continues to revolutionize drug discovery and protein design, the race between innovation and security shows no signs of slowing.
Read more: