On this page
DNA sequence
©Gio_tto
Sandpit
Uni­ver­si­ty of MainzNatural Sci­en­ces

AI in biology

How can ar­ti­fi­ci­al in­tel­li­gence be used in biology without risking misuse?

DNA, RNA, and pro­te­ins play a key role in con­trol­ling the mole­cu­lar pro­ces­ses in our cells. Cutting-edge AI models are capable of pre­dic­ting the func­tions of bio­lo­gi­cal se­quen­ces like these and de­si­gning new ones, in­clu­ding AI-ge­ne­ra­ted viruses. At a sandpit event or­ga­ni­zed by Wübben Stif­tung Wis­sen­schaft, experts dis­cus­sed stra­te­gies for pre­ven­ting misuse of the tech­no­lo­gy.

Sci­en­tists at Stan­ford Uni­ver­si­ty de­mons­tra­ted in Sep­tem­ber 2025 what ge­ne­ra­ti­ve AI models are already capable of. They pro­du­ced genomes of func­tio­n­ing bac­te­rio­pha­ges, some of which had very low si­mi­la­ri­ty to their natural role models. Bac­te­rio­pha­ges are viruses that attack bac­te­ria. In the lab, the syn­the­tic virus suc­cess­ful­ly de­s­troy­ed Esche­ri­chia coli bac­te­ria, showing that ge­ne­ra­ti­ve AI models hold great po­ten­ti­al for medical ap­p­li­ca­ti­ons. For in­stan­ce, AI-ge­ne­ra­ted bac­te­rio­pha­ges could offer a way of figh­t­ing multi-re­sistant bac­te­ria. And protein struc­tures ge­ne­ra­ted by similar AI models could serve as a basis for the­ra­peu­tic an­ti­bo­dies – to treat cancer, for in­stan­ce.

We don't want to conjure up any night­ma­rish sce­n­a­ri­os; we want to assess the po­ten­ti­al impacts of ge­ne­ra­ti­ve AI models in the life sci­en­ces with a cri­ti­cal eye.

Maximilian Sprang

But how do we prevent these AI models from being used to design bio­lo­gi­cal se­quen­ces that could be harmful? If we give our ima­gi­na­ti­on free rein, we can think of count­less sce­n­a­ri­os in which this tech­no­lo­gy could be misused: A ma­li­cious agent could design a novel virus for use as a bio­lo­gi­cal weapon, for example, or a pa­tho­gen that only attacks rice plant va­rie­ties that grow in a hostile neigh­bo­ring country.

“We don't want to conjure up any night­ma­rish sce­n­a­ri­os; we want to assess the po­ten­ti­al impacts of ge­ne­ra­ti­ve AI models in the life sci­en­ces with a cri­ti­cal eye,” says Ma­xi­mi­li­an Sprang, a bio­in­for­ma­ti­ci­an at Jo­han­nes Gu­ten­berg Uni­ver­si­ty Mainz. In No­vem­ber 2025 he brought leading experts in the fields of com­pu­ter science, biology, ethics, po­li­tics, and ad­ja­cent sub­jec­ts tog­e­ther with re­p­re­sen­ta­ti­ves of the phar­maceu­ti­cals in­dus­try to high­light the risks that could arise from ge­ne­ra­ti­ve AI, as well as the op­por­tu­nities. The focus was on dual-use sce­n­a­ri­os, in which AI tech­no­lo­gy can be used both for life-saving me­di­ci­ne and deadly bio­lo­gi­cal weapons. “The problem is that the models are im­pro­ving at a fast pace and can be used for both good and pro­ble­ma­tic ap­p­li­ca­ti­ons,” says Sprang.

The sandpit princip­le: Mul­ti­ple di­sci­pli­nes and pro­duc­tive misun­derstan­dings

participants from the sandpit
©WSW

Ma­xi­mi­li­an Sprang (center) with par­ti­ci­pants from the sandpit

For three days, the 24 re­se­ar­chers gathe­red in In­gel­heim had the op­por­tu­ni­ty to share in­for­ma­ti­on on the current state of AI use in the life sci­en­ces, discuss dif­fe­rent per­spec­tives, and develop pro­po­sals for dealing with po­ten­ti­al risks. One of the aims was to for­mu­la­te risk-pre­ven­ti­on stra­te­gies, draw up gui­de­li­nes, and forge stron­ger links between tech­ni­cal experts and po­li­cy­ma­kers to make ge­ne­ra­ti­ve AI use safer. “Biology sup­plies us with the most highly de­ve­lo­ped ‘hard­ware’ in the uni­ver­se, which we can use for future health­ca­re and the welfare of our planet,” says Marc Güell, one of the par­ti­ci­pants, who is a syn­the­tic bio­lo­gist at Pompeu Fabra Uni­ver­si­ty in Bar­ce­lo­na. “But we must take care to use the tech­no­lo­gy in the best way.”

The dis­cus­sion took place in a sandpit, an event format that Wübben Stif­tung Wis­sen­schaft uses for in-depth ex­plo­ra­ti­on of current re­se­arch ques­ti­ons in in­ter­di­sci­pli­na­ry teams. “AI in the life sci­en­ces is such a mul­ti­fa­ce­ted topic that it is im­pos­si­ble to grasp with a silo ap­proach,” says Sprang. “Thin­king through the con­se­quen­ces re­qui­res know­ledge from a wide range of dif­fe­rent fields.”

The sandpit ap­proach is based on three dia­lo­gue formats: short, thought-pro­vo­king “seed talks”, round-table dis­cus­sions, and in­ter­ac­tive work­shops. The seed talks by experts from com­pu­ter science, bio­tech­no­lo­gy, and po­li­tics brought all the par­ti­ci­pants up to speed with the current state of play in the speakers’ re­spec­tive fields. The round-table dis­cus­sions in various groups of dif­fe­rent sizes fa­ci­li­ta­ted an ex­chan­ge of views on the chal­len­ges faced in this context, while the in­ter­ac­tive work­shops – based on design thin­king – pro­du­ced ideas for safer use of AI. “The small groups and their chan­ging con­fi­gu­ra­ti­ons re­sul­ted in a clash of di­sci­pli­nes and per­spec­tives, which ine­vi­ta­b­ly led to misun­derstan­dings, but these were pro­duc­tive rather than ob­st­ruc­tive,” says Rosae María Martín Peña, one of the par­ti­ci­pants, who is a postdoc at the Centre for Ethics and Law in the Life Sci­en­ces at Leibniz Uni­ver­si­ty Han­no­ver.

Safety gaps that must be closed

Re­gu­la­ti­ons and pro­ce­du­res for en­su­ring the safety of newly created bio­lo­gi­cal se­quen­ces already exist. For in­stan­ce, there are de­tec­tion pro­to­cols that can at least prevent the in­vol­un­ta­ry ge­nera­ti­on of harmful se­quen­ces. They compare DNA that is due to be created ar­ti­fi­ci­al­ly with known harmful se­quen­ces and sound the alarm if the si­mi­la­ri­ty is too great. “But there are safety gaps, for in­stan­ce if AI is used to exploit the gray zone between known DNA se­quen­ces and new part-se­quen­ces, and the new se­quence falls through the cracks,” says Sprang. “It is not yet pos­si­ble to make brand new DNA se­quen­ces, but simply re­con­fi­gu­ring exis­ting DNA could be enough to evade exis­ting safe­guards and produce dan­ge­rous se­quen­ces.”

One im­portant dis­cus­sion con­cer­ned risk mi­ni­mi­za­ti­on when using AI models to help trans­la­te bio­me­di­cal re­se­arch into cli­ni­cal prac­tice. A lack of pre­cisi­on or a dis­tor­ti­on in the un­der­ly­ing data can make even well-in­ten­tio­ned use sce­n­a­ri­os pro­ble­ma­tic and, in the worst case, could harm pa­ti­ents. So we have to ask what data the AI models are using to produce their results. “The models are very data-hungry and most of the data they feed on still relates to Cau­ca­si­an men, which will distort the pre­dic­tions,” says Sprang. In the same way, de­mo­gra­phic and so­cio­eco­no­mic dif­fe­ren­ces are often not given suf­fi­ci­ent con­si­de­ra­ti­on. Poverty, for in­stan­ce, is a si­gni­fi­cant health factor. “If these aspects are ignored, it can lead to faulty con­clu­si­ons.”

To close these kinds of safety gaps, we need ethical re­gu­la­ti­ons that go beyond tech­ni­cal safety mea­su­res and cover the al­lo­ca­ti­on of re­spon­si­bi­li­ties, data pro­tec­tion at po­pu­la­ti­on level, and robust over­sight me­cha­nisms, as well as in­no­va­ti­on.

Rosae Martín Peña

This raises a key ques­ti­on: Who is re­spon­si­ble if so­me­thing goes wrong? For example, if the wrong tre­at­ment is pre­scri­bed at a hos­pi­tal on the basis of mis­di­rec­ted AI models. In most cases, re­spon­si­bi­li­ty is shared between mul­ti­ple actors, and the lia­bi­li­ty si­tua­ti­on is unclear. “To close these kinds of safety gaps, we need ethical re­gu­la­ti­ons that go beyond tech­ni­cal safety mea­su­res and cover the al­lo­ca­ti­on of re­spon­si­bi­li­ties, data pro­tec­tion at po­pu­la­ti­on level, and robust over­sight me­cha­nisms, as well as in­no­va­ti­on,” says Martín Peña.

The sandpit con­ti­nues to have an impact as an active expert network

Teilnehmer des Sandpits
©WSW

The main aim of the sandpit was, in the first place, to draw up a white paper on ethical data use and the risks of dual-use sce­n­a­ri­os, in­clu­ding re­com­men­da­ti­ons for mi­ni­mi­zing these risks. “At the end of the sandpit, we worked in small groups to develop so­lu­ti­ons to very con­cre­te pro­blems that will serve as the basis for the white paper,” says Sprang. “We were all sur­pri­sed at how many good sug­ges­ti­ons were pro­du­ced in such a short time thanks to the design thin­king ap­proach.” The white paper, which will be sent to po­li­cy­ma­king bodies in summer 2026, is in­ten­ded to build a bridge between po­li­ti­cal and sci­en­ti­fic and tech­ni­cal actors and fa­ci­li­ta­te evi­dence-based re­gu­la­ti­on.

AI models promise nu­me­rous sen­si­ble use sce­n­a­ri­os for the future, but we ur­gent­ly need to examine what they can really do, the risks and the ob­sta­cles.

Maximilian Sprang

The sandpit could also lead to the crea­ti­on of a large Eu­ropean re­se­arch project that would develop AI models to design DNA se­quen­ces with in­trin­sic safe­guards against the risk of misuse. A number of funding ap­p­li­ca­ti­ons have already been sub­mit­ted or are in the plan­ning stage. Im­pro­ved me­cha­nisms to iden­ti­fy harmful ge­ne­ra­ted DNA se­quen­ces are another area the par­ti­ci­pants are working on. “AI models promise nu­me­rous sen­si­ble use sce­n­a­ri­os for the future, but we ur­gent­ly need to examine what they can really do, the risks and the ob­sta­cles,” says Sprang. “This is par­ti­cu­lar­ly im­portant in me­di­ci­ne, where misuse can quickly become life-threa­ten­ing.”

Maximilian Sprang
©Ma­xi­mi­li­an Sprang/privat

Ma­xi­mi­li­an Sprang has been leading a junior re­se­arch group at the Medical Center of Jo­han­nes Gu­ten­berg Uni­ver­si­ty Mainz since March 2025, where he com­bi­nes bio­in­for­ma­tics and AI to uncover pat­terns in bio­lo­gi­cal data and support trans­la­tio­nal re­se­arch in im­mu­no­lo­gy. At the end of 2025, he re­cei­ved the Ein­stein Foun­da­ti­on Early Career Award 2025 for his project “Erring Ri­go­rous­ly,” which aims to improve re­pro­du­ci­bi­li­ty and data re­lia­bi­li­ty in func­tio­n­al ge­no­mics.

From initial brain­stor­ming to re­se­arch project: The Wübben Foun­da­ti­on sand­pits are where brave new re­se­arch ideas are born. Click here for more in­for­ma­ti­on

Anne Schmieder
©Uni­ver­si­tät Leipzig

In conversation with: Anne Schmieder

«We can’t yet blindly rely on ge­ne­ra­ti­ve AI models, but the re­sour­ces they save make them an in­dis­pensable asset.»

What per­spec­tive did you con­tri­bu­te to the dis­cus­sion?
As a re­se­ar­cher in a protein design la­bo­ra­to­ry, the Schoeder Lab at Leipzig Uni­ver­si­ty, and as a member of the AI com­pe­tence center ScaDS.AI, I work with ge­ne­ra­ti­ve AI tools every day. I par­ti­ci­pa­ted in the sandpit to learn about dif­fe­rent per­spec­tives on how ge­ne­ra­ti­ve AI, or GenAI, is chan­ging the bio­sci­en­ces – and to take part in in­ter­di­sci­pli­na­ry dia­lo­gue. In my opinion, the grea­test po­ten­ti­al of GenAI is in the fields of phar­maceu­ti­cals, me­di­ci­ne, and che­mi­stry. The speed and ca­pa­ci­ty avail­ab­le now when running se­ar­ches for po­ten­ti­al drug can­di­da­tes have in­crea­sed ex­po­nen­ti­al­ly. Ex­pe­ri­men­tal va­li­da­ti­on in the lab is still es­sen­ti­al, but the early search process is much more ef­fi­ci­ent. Alt­hough I believe the be­ne­fits out­weigh the risks, the grea­test chal­len­ge is still in­ter­pre­ta­bi­li­ty. It is cur­r­ent­ly dif­fi­cult to un­der­stand fully how these models process input data and how to in­ter­pret the results with 100 percent cer­tain­ty. We can’t yet blindly rely on ge­ne­ra­ti­ve AI models, but the re­sour­ces they save make them an in­dis­pensable asset.

 

Why was the sandpit a good format for ad­van­cing the dis­cus­sion on ge­ne­ra­ti­ve AI?
The sandpit format was great for spar­king an ongoing dis­cus­sion. I par­ti­cu­lar­ly liked how the group work en­cou­ra­ged us to learn about de­ve­lop­ments outside our spe­ci­fic niches. The pro­fes­sio­nal mo­de­ra­ti­on was key. It not only kept us on track, but also enabled va­lu­able side dis­cus­sions, which led to strong con­nec­tions between the par­ti­ci­pants. Since dif­fe­rent groups were as­sem­bled for each task, we were con­stant­ly being con­fron­ted with new per­spec­tives. This dynamic struc­tu­re meant we were able to keep re­fi­ning our own per­spec­tives and im­pro­ving the quality of our joint results.

 

What do you hope the sandpit will achieve in the long term?
I hope the sandpit will focus awa­reness on this topic and help spread ex­per­ti­se among sci­en­tists. Our home in­sti­tu­ti­ons will profit di­rec­t­ly from this know­ledge trans­fer because we can better educate col­leagues who may not deal with AI or its po­ten­ti­al risks every day. In terms of the re­spon­si­ble use of GenAI, I hope future pro­jec­ts will place an even stron­ger em­pha­sis on bio­safe­ty and proac­tive risk as­sess­ment. It is vital that we discuss these impacts openly before any serious risks arise. For me per­so­nal­ly, the sandpit has already been a success: I have met in­credi­b­ly fa­sci­na­ting re­se­ar­chers, which has led to ex­ci­ting col­la­bo­ra­ti­ons.

Contact
Dr. Ma­xi­mi­li­an Sprang, Uni­ver­si­ty Medical Center of the Jo­han­nes Gu­ten­berg Uni­ver­si­ty Mainz, The Mayer Lab, ma­sprang@uni-mainz.de