Autoregressive generative model pretrained on 30 million full-length natural mRNA sequences that jointly optimizes 5' UTR, CDS, and 3' UTR for therapeutic mRNA stability and translation efficiency.
mRNA-GPT is an autoregressive generative model for designing therapeutic messenger RNA, posted to bioRxiv in early April 2026. Unlike earlier mRNA-design tools that optimize 5' UTR, coding sequence (CDS), and 3' UTR independently, mRNA-GPT is pretrained on 30 million full-length natural mRNA sequences and learns the joint distribution across all three regions. After pretraining, the model is fine-tuned with reinforcement learning to optimize designed sequences for stability and translation-efficiency reward signals.
This addresses a key limitation of existing mRNA optimization workflows: optimal CDS choices depend on UTR context and vice versa, and tools that optimize regions in isolation can miss strong interactions between them.
mRNA-GPT uses a decoder-only transformer pretrained autoregressively on 30M full-length natural mRNA sequences. After pretraining, the model is fine-tuned via reinforcement learning with reward signals derived from experimental measurements of mRNA stability and translation efficiency. The bioRxiv preprint reports architecture, training corpus details, and ablations on the impact of the RL stage.
Benchmarks include comparisons against codon-table optimization tools (CodonW, EMBOSS) and prior ML-based UTR-optimization tools, evaluating both translation-efficiency proxies and direct in vitro measurements.
mRNA-GPT is directly applicable to therapeutic mRNA design — vaccines, protein-replacement therapies, and mRNA-based gene therapies — where stability and translation efficiency are critical product attributes. The unified UTR-CDS-UTR generation removes manual handoffs between separate codon-optimization, UTR-design, and folding-check stages.
mRNA-GPT is the first generative foundation model to address the mRNA design problem holistically by jointly modeling UTRs and CDS. Coupled with experimental validation reported in the preprint, it represents a meaningful step toward foundation-model approaches to mRNA-therapeutic engineering, complementing related efforts on UTR-specific models (5-UTR-LM) and coding-sequence optimization tools (CaLM, CodonFM).
Li, S., et al. (2026) mRNA-GPT: A Generative Model for Full-Length mRNA Design and Optimization. bioRxiv.
DOI: 10.64898/2026.03.31.715707