Agentic Omics architecture showing LLM orchestrating domain-specific biological AI models

The Agentic Omics Vision: LLMs Meet Domain-Specific AI

Introduction: The Convergence Point In Post 13, we defined agentic AI as systems that autonomously plan, reason, use tools, and execute multi-step scientific workflows. Now we arrive at the central thesis of this entire series: Agentic Omics — the convergence of large language model (LLM) reasoning with domain-specific biological AI models like AlphaFold, ESM, scGPT, and DNABERT to create autonomous systems capable of end-to-end biological discovery. This is not science fiction. As of early 2026, agentic systems are being deployed in operational drug discovery settings at companies like AstraZeneca, with documented implementations compressing workflows that once took months into hours while maintaining scientific traceability (Seal et al., 2025). The question is no longer if this convergence will transform biology, but how — and what architecture will get us there most reliably. ...

March 10, 2026 · 67 AI Lab

AI for Genomics: Reading the Book of Life with Transformers

The genome is the ultimate source code. For decades, computational biologists have relied on alignment algorithms, hidden Markov models, and specialized machine learning to decode it. Today, a new paradigm is taking hold: DNA foundation models. By treating the genome as a vast, continuous text and training large language models (LLMs) on billions of nucleotides, researchers are teaching AI to “read” the book of life in its native language. In this fifth installment of our Agentic Omics series, we examine the state of the art in genomic AI. We explore how models like DNABERT-2, Nucleotide Transformer, Evo, and HyenaDNA are moving beyond sequence classification to predict gene expression, identify regulatory elements, and quantify variant effects. Crucially, we will dissect the architectural innovations that make this possible—and the biological complexities that still confound these models. ...

February 28, 2026 · 67 AI Lab
A transformer model reading DNA, RNA, proteins, and single-cell profiles as linked biological languages

Foundation Models Meet Biology: The Transformer Revolution in Life Sciences

In the first post of this series, we mapped the omics landscape: genomics, transcriptomics, proteomics, metabolomics, metagenomics, phenomics. The next question is obvious: why did AI suddenly get so good at several of these fields at once? The short answer is that biology turned out to be unusually compatible with the same family of models that transformed natural language processing. DNA, RNA, proteins, and even single-cell expression matrices are not “language” in any literal sense, but they are structured symbol systems with long-range dependencies, rich context, and vast quantities of unlabeled data. That is exactly the setting where self-supervised foundation models thrive. ...

February 25, 2026 · 67 AI Lab