OpenAI released GPT-Rosalind yesterday — the first entry in a new "Life Sciences" model series, gated to a trusted-access programme for a handful of enterprise partners. Amgen, Moderna, Thermo Fisher, the Allen Institute, and Los Alamos National Laboratory are on the preview list. Named after the crystallographer who made DNA legible, the model ships in ChatGPT Enterprise, Codex, and the API, behind enterprise-grade security controls and the standard "no training on your data" clause.
The benchmark numbers are good. On BixBench, GPT-Rosalind scores 0.751 Pass@1. On LABBench2, it wins six of eleven subtasks. Against human experts on two representative tasks, it sits in the 95th and 84th percentiles. Reasonable results for a domain model built on top of the frontier reasoning stack.
But the interesting thing isn't the benchmarks. It's the language around them.
Read the announcement carefully. OpenAI doesn't say GPT-Rosalind will design drugs. The model is described as a tool to accelerate the early stages of discovery — evidence synthesis, hypothesis generation, experimental planning. That's the research-assistant frame, not the autonomous-designer frame. The Codex Life Sciences plugin talks to more than fifty scientific databases. The model reads papers, cross-checks datasets, drafts experiments. That is useful. It is not a cure.
Contrast that with the last five years of pure-play AI drug discovery. Exscientia, Recursion, BenevolentAI — the three companies that raised the most money on the premise that AI could find drugs faster — have all had their first clinical readouts. All three were negative. Recursion absorbed Exscientia. BenevolentAI has been in retrenchment. The sector is sitting on roughly a billion and a half in market cap and has approved zero drugs. UCL's Peter Coveney, cited in a recent Nature piece, has made the structural case: discovery isn't the bottleneck. Validation is. You can generate ten thousand candidate molecules. Testing them is the part that takes a decade.
GPT-Rosalind isn't promising to solve validation. It's promising to make the scientists who do validation a little faster at reading papers. That's a smaller claim. It might also be a correct one.
There's something honest about the framing. Opus 4.7 shipped yesterday with Mythos held back, carefully gated — the same instinct in a different domain. OpenAI's move here rhymes: trusted-access only, enterprise-only, a short partner list, no individual researchers, dual-use safety language that reads as if somebody who has read the risk literature wrote the announcement.
Whether this is restraint or marketing discipline is open. It's possible GPT-Rosalind is simply a harder model to sell on hype because pharma buyers have been burned too many times and know better. It's possible the trusted-access structure is there to keep the model from generating plausible-but-wrong bioinformatics claims at industrial scale.
Either way, it's the first new model family in a while whose launch language reads like it was written by people who watched the AI-for-biology narrative play out in public. Rosalind Franklin did careful work and died before seeing it credited. Putting her name on a frontier model is either a very nice gesture or a warning about what happens when you overstate.
Sources:
-
Introducing GPT-Rosalind for life sciences research — OpenAI
-
OpenAI launches GPT-Rosalind, a specialised AI model for drug discovery and life sciences research — The Next Web
-
OpenAI introduces GPT-Rosalind, its drug discovery AI — pharmaphorum
-
Recursion, Exscientia, and AI drug discovery's moment of truth — pharmaphorum
-
Why the AI drug revolution has yet to deliver — Scientific Computing World