Skip to content
The large language of life?

OpenAI starts offering a biology-tuned LLM

GPT-Rosalind is an LLM trained on biology workflows, available in closed access.

John Timmer | 37
A complex blue, glowing web of connections on a black background.
Biological systems have large webs of interactions that the human brain can struggle to process. Credit: Andriy Onufriyenko
Biological systems have large webs of interactions that the human brain can struggle to process. Credit: Andriy Onufriyenko
Story text

On Thursday, OpenAI announced it had developed a large language model specifically trained on common biology workflows. Called GPT-Rosalind after Rosalind Franklin, the model appears to differ from most science-focused models from major tech companies, which have generally taken a more generic approach that works for various fields.

In a press briefing, Yunyun Wang, OpenAI’s Life Sciences Product Lead, said the system was designed to tackle two major roadblocks faced by current biology researchers. One is the massive datasets created by decades of genome sequencing and protein biochemistry, which can be too much for any one researcher to take in. The second is that biology has many highly specialized subfields, each with its own techniques and jargon. So, for example, a geneticist who finds themselves working on a gene that’s active in brain cells might struggle to understand the immense neurobiological literature.

Wang said the company had taken an LLM and trained it on 50 of the most common biological workflows, as well as on how to access the major public databases of biological information. Further training has resulted in a system that can suggest likely biological pathways and prioritize potential drug targets. “We’re connecting genotype to phenotype through known pathways and regulatory mechanisms, infer likely structural or functional properties of proteins, and really leveraging this mechanistic understanding,” Wang said.

To address LLMs’ tendencies toward sycophancy and overenthusiasm, OpenAI says it has tuned the model to be more skeptical, so it’s more likely to tell you when something is a bad drug target. There was a lot of talk about GPT-Rosalind’s “reasoning” and “expert-level” abilities. We were told that the former was defined as being able to work through complex, multi-step processes, while the latter was derived from the model’s performance on a handful of benchmarks.

It’s unclear whether OpenAI has tackled the hallucination issue that has plagued a variety of LLMs and can also strike when the systems are prompted to explain the steps the company took to reach its conclusions. Given past experience, it’s likely we’ll see a mix of glowing reports about unexpected connections the AI finds, as well as instances where it produces obviously erroneous suggestions.

For the moment, however, the company is limiting access due to concerns about the model’s potential for harmful outputs if asked to do something like optimize a virus’s infectivity. Only US-based entities can apply to OpenAI’s trusted access deployment structure at the moment, and the company will limit who can use it. A more limited Life Sciences Research Plugin will be made generally available.

As noted above, a number of other companies have made science-focused agentic LLMs available, but those were much less focused than GPT-Rosalind, which is biology-specific. Until we start hearing reports on the effectiveness of this new model, it’s difficult to evaluate whether this focus improves its utility.

Photo of John Timmer
John Timmer Senior Science Editor
John is Ars Technica's science editor. He has a Bachelor of Arts in Biochemistry from Columbia University, and a Ph.D. in Molecular and Cell Biology from the University of California, Berkeley. When physically separated from his keyboard, he tends to seek out a bicycle, or a scenic location for communing with his hiking boots.
37 Comments