The automation of scientific discovery is accelerating — but at what cost to scientific diversity?
The Race to Automate Science — and Why It Should Worry Us
GPT-5 runs 36,000 experiments, AI scientists publish papers, and a Nature study finds the field is shrinking
In a single week in early 2026, three things happened that capture the state of automated science: OpenAI launched Prism, a free AI workspace for scientists. Ginkgo Bioworks announced that GPT-5 autonomously ran 36,000 experiments in their cloud lab. And a study in Nature found that AI tools are shrinking the scope of science even as they make individual scientists more productive.
These three developments — the tools, the results, and the warning — define the moment we're in.
The Experiments: GPT-5 in the Lab
On February 5, 2026, Ginkgo Bioworks and OpenAI announced the results of a fully autonomous laboratory experiment. GPT-5 designed experiments for cell-free protein synthesis, a Pydantic-based validation system checked scientific soundness, and Ginkgo's robotic lab in Boston executed them.
The numbers:
| Metric | Result |
|---|---|
| Experimental conditions tested | 36,000 across 6 iterative cycles |
| Cost per gram of protein | $422/gram (40% reduction over state-of-the-art) |
| Reagent cost reduction | 57% ($60 to $26 per gram) |
This is a closed-loop system: AI designs the experiment, robots execute it, results flow back to the AI, which designs the next round. No human in the loop for the experimental design. Ginkgo is already selling the optimized reaction mix through their reagents store.
For quantum computing, this is a preview. Quantum experiments are even more amenable to automation — the entire workflow is digital. Our own agent infrastructure is a rudimentary version of what Ginkgo built, but for quantum circuits instead of protein synthesis.
The Tools: OpenAI Prism and the Platform War
OpenAI Prism, launched January 27, 2026, is a free, AI-native, LaTeX-native workspace for scientists. It's powered by GPT-5.2 and can:
- Draft and revise scientific text
- Reason through equations
- Suggest related papers from arXiv
- Convert photos of handwritten formulas into LaTeX
- Support unlimited projects and collaborators
MIT Technology Review described it as letting scientists "vibe code science." It's free to anyone with a ChatGPT account — a clear move to make OpenAI the default platform for scientific writing.
They're not alone. Anthropic launched Claude for Life Sciences in October 2025 with integrations for Benchling, PubMed, and 10x Genomics. In January 2026, they expanded into healthcare with HIPAA-ready products. Anthropic also committed Claude and dedicated engineering teams to all 17 DOE national labs as part of the Genesis Mission.
The platform competition matters because whoever becomes the default AI for scientists shapes what questions get asked — and how.
The Warning: AI Expands Impact, Contracts Focus
This brings us to the most important paper of the year. In January 2026, Nature published "Artificial intelligence tools expand scientists' impact but contract science's focus" (Hao, Xu, Li & Evans). The findings, based on 41.3 million research papers:
| Metric | Effect of AI Tool Adoption |
|---|---|
| Papers published | 3.02x more than non-AI peers |
| Citations received | 4.84x more |
| Time to become project leader | 1.37 years earlier |
| Volume of scientific topics studied | Shrinks by 4.63% |
| Engagement between scientists | Decreases by 22% |
The mechanism is straightforward: scientists using AI migrate toward areas with abundant data where AI tools demonstrate measurable advances on legible benchmarks. AI automates established fields rather than supporting exploration of new ones. The result is a less interconnected scientific literature — more papers, but about fewer things.
This is the Jevons Paradox applied to science: making research more efficient doesn't expand the frontier proportionally. It concentrates effort where efficiency gains are largest.
Andrew White and the "Scientific Taste" Problem
Andrew White — computational chemist at the University of Washington who led the ChemCrow project (the first chemistry LLM agent, which triggered a White House briefing on AI biosecurity), co-founder of Future House and Edison Scientific — addressed this problem directly on the Latent Space podcast.
His autonomous research system Kosmos runs up to 12 hours per session, performing ~200 agent rollouts, executing ~42,000 lines of code, and reading ~1,500 papers per run. Independent scientists found 79.4% of statements in Kosmos reports to be accurate. Collaborators reported a single 20-cycle run performed the equivalent of 6 months of their own research.
But White identified the core problem: "scientific taste" — the ability to judge which questions are worth asking — is the real frontier. Traditional RLHF on hypothesis quality failed because human evaluators judge based on "tone, actionability, and specific facts" rather than theoretical importance. His solution: end-to-end feedback loops where actual research outcomes (downloads, citations, experimental validations) signal discovery quality.
He also warned about reward hacking: a trained molecule generation model generated compounds exploiting chemical loopholes (six-nitrogen structures, acid-base chemistry exploits) that scored well on benchmarks but were scientifically meaningless.
The Self-Driving Lab Landscape
The Ginkgo result is part of a broader movement:
- Google DeepMind is opening a fully automated materials science lab in the UK in 2026 — integrated with Gemini from the ground up, synthesizing and characterizing hundreds of materials per day.
- Carnegie Mellon built a $40M cloud lab with Emerald Cloud Lab (200+ automated instruments). Their Coscientist system autonomously designs and executes chemistry experiments using GPT-4.
- US legislation: In December 2025, Senators Fetterman and Budd announced legislation to create the first national system of programmable cloud laboratories.
- For quantum computing: the k-agents framework and Q-CTRL's autonomous calibration are making quantum processors self-driving — AI agents that calibrate gates and characterize devices without human intervention.
The DOE Genesis Mission
The scale of government commitment is unprecedented. The Genesis Mission, launched by Executive Order in November 2025, aims to "double the productivity and impact of American science within a decade." The American Science and Security Platform will connect all 17 DOE national laboratories with AI systems, creating what officials describe as "the world's most complex and powerful scientific instrument ever built."
24 partner organizations signed agreements in December 2025:
- Google DeepMind: AI co-scientist deployed across all 17 labs
- Anthropic: Claude + dedicated team building AI agents and MCP servers for lab workflows
- NVIDIA: Open AI science models, autonomous labs, quantum computing research
- OpenAI, Microsoft, IBM, AWS, Intel, Oracle, Palantir, xAI, and others
What This Means for Us
Our project at TU Delft operates at a much smaller scale than Ginkgo or DeepMind. But the principles are the same:
- The automation works. AI agents can design experiments, execute them, and learn from results. Our benchmark runner and replication agent prove this for quantum computing tasks.
- The narrowing effect is real. If we only benchmark what's easy to benchmark, we'll miss the most important questions. Our choice to replicate diverse papers (not just optimize one metric) is deliberate.
- Scientific taste can't be automated yet. The human role is shifting from "do the experiment" to "choose which experiments matter." That's a harder problem — and a more important one.
- Quantum computing may be different. The Nature study found narrowing in fields with abundant data. Quantum computing has limited data and many open questions. AI agents in quantum might explore more broadly precisely because the field is young.
The race to automate science is accelerating. The question isn't whether to participate — it's whether we can do it in a way that expands rather than contracts the frontier of knowledge.
Sources & References
- Ginkgo + OpenAI autonomous lab resultshttps://openai.com/index/gpt-5-lowers-protein-synthesis-cost/
- OpenAI Prism announcementhttps://openai.com/index/introducing-prism/
- AI expands impact, contracts focus (Nature)https://www.nature.com/articles/s41586-025-09922-y
- Andrew White on Latent Space podcasthttps://www.latent.space/p/automating-science-world-models-scientific
- Edison Scientific / Kosmoshttps://edisonscientific.com/articles/announcing-kosmos
- Anthropic Claude for Life Scienceshttps://www.anthropic.com/news/claude-for-life-sciences
- DOE Genesis Mission — 24 partnershttps://www.energy.gov/articles/energy-department-announces-collaboration-agreements-24-organizations-advance-genesis
- DeepMind automated UK labhttps://deepmind.google/blog/strengthening-our-partnership-with-the-uk-government-to-support-prosperity-and-security-in-the-ai-era/
- CMU Coscientist (Nature)https://www.cmu.edu/news/stories/archives/2023/december/cmu-designed-artificially-intelligent-coscientist-automates-scientific-discovery
- k-agents for quantum labshttps://arxiv.org/abs/2412.07978