Long-context generative RNA foundation model trained on 114 million full-length RNA sequences, supporting de novo design of tRNAs, aptamers, CRISPR guide RNAs, mRNAs, and circular RNAs.
EVA (Evolutionary Versatile Architect) is a generative RNA foundation model from the GENTEL Lab at the Chinese Academy of Sciences, posted to bioRxiv in March 2026. EVA is trained on OpenRNA v1, a curated corpus of 114 million full-length RNA sequences spanning all domains of life, and is the first unified architecture targeting the breadth of functional RNA design tasks: transfer RNAs, aptamers, CRISPR guide RNAs, messenger RNAs, and circular RNAs.
Unlike earlier RNA language models such as RNA-FM or Uni-RNA, which are primarily encoders for downstream classification or regression, EVA is generative — it samples RNA sequences from learned distributions conditioned on task-specific context. The model achieves state-of-the-art results on 7 of 9 public RNA design benchmarks, outperforming RfamGen, GenerRNA, and conventional structure-guided design tools.
EVA uses a decoder-only transformer architecture trained autoregressively on the OpenRNA v1 corpus. Sequence tokenization operates at the nucleotide level. The training objective is standard next-token prediction; conditional generation is supported via prefix prompting with class tokens or structural constraints. The preprint describes ablations on context length, training data filtering, and class-balancing strategies.
Benchmarks reported in the preprint include tRNA acceptor-stem design, theophylline aptamer generation, CRISPR-Cas9 guide RNA on-target activity, mRNA codon optimization, and circular-RNA scaffolding. EVA outperforms prior task-specific tools on 7 of 9 evaluated benchmarks.
EVA is suited for synthetic biology and RNA therapeutics groups that need to design functional RNAs without committing to a separate task-specific tool per RNA class. In therapeutic mRNA design, it provides a generative alternative to rule-based codon optimization. In aptamer engineering, it can propose candidate sequences that meet structural and binding constraints. For CRISPR applications, it offers guide-RNA design with predicted on-target activity informed by evolutionary context.
EVA is the first RNA foundation model to span the breadth of functional RNA design as a generative system rather than as a downstream encoder. By demonstrating SOTA on 7 of 9 tasks within a single unified model, it argues for foundation-model approaches in RNA design analogous to the protein-design trajectory established by ProGen and ESM-3. The 114M-sequence OpenRNA v1 corpus is itself a valuable community resource for further work in RNA foundation modeling.
Huang, Y., et al. (2026) A Long-Context Generative Foundation Model Deciphers RNA Design Principles. bioRxiv.
DOI: 10.64898/2026.03.17.712398