A transformer model reading DNA, RNA, proteins, and single-cell profiles as linked biological languages

Foundation Models Meet Biology: The Transformer Revolution in Life Sciences

In the first post of this series, we mapped the omics landscape: genomics, transcriptomics, proteomics, metabolomics, metagenomics, phenomics. The next question is obvious: why did AI suddenly get so good at several of these fields at once? The short answer is that biology turned out to be unusually compatible with the same family of models that transformed natural language processing. DNA, RNA, proteins, and even single-cell expression matrices are not “language” in any literal sense, but they are structured symbol systems with long-range dependencies, rich context, and vast quantities of unlabeled data. That is exactly the setting where self-supervised foundation models thrive. ...

February 25, 2026 · 67 AI Lab