bio.rodeo
HomeCompetitorsLeaderboardOrganizations
bio.rodeo

The authoritative source for evaluating biological foundation models. No hype, just honest analysis.

© 2026 bio.rodeo. All rights reserved.
Imaging

OmniEM

Peking University

Unified electron microscopy image analysis toolkit built on EM-DINO, a vision foundation model pretrained on 5 million diverse EM images.

Released: 2025

Overview

OmniEM is a comprehensive toolkit for electron microscopy (EM) image analysis built around EM-DINO, the first vision foundation model pretrained specifically on EM data. Developed by the AI4BioMed group at Peking University and released as a preprint in April 2025, the system addresses a long-standing challenge in the field: EM datasets are extraordinarily heterogeneous in terms of species, tissue type, imaging protocol, and resolution, making it difficult to train models that generalize across experiments. By pretraining at scale on a standardized corpus of 5 million EM images spanning resolutions from 0.5 to 70 nm per pixel, EM-DINO learns transferable visual representations that underpin a suite of downstream tasks without requiring task-specific retraining from scratch.

The framework consists of two tightly coupled components. EM-DINO is a vision transformer backbone trained using a self-supervised DINO-style objective, producing multi-scale embeddings that capture rich structural features across magnifications and biological contexts. OmniEM is a U-shaped encoder-decoder architecture built on top of the EM-DINO backbone that handles denoising, super-resolution, and multi-class organelle segmentation within a single unified model, replacing the fragmented array of task-specific pipelines that researchers currently depend on.

OmniEM ships with a Napari plugin that provides an accessible graphical interface for interactive EM analysis, making the model usable by imaging specialists and wet-lab biologists who do not work primarily in Python. The system is available through GitHub and represents one of the first attempts to bring foundation model-scale pretraining to the electron microscopy domain.

Key Features

  • EM-Specific Foundation Model: EM-DINO is the first vision foundation model pretrained exclusively on EM imagery, capturing nanoscale biological structure rather than relying on transfer from natural image datasets.
  • Large-Scale Pretraining Corpus: EM-5M contains 5 million curated EM images spanning multiple species, tissue types, and imaging modalities including TEM, SEM, and FIB-SEM, designed to maximize generalization across diverse experimental settings.
  • Unified Multi-Task Architecture: A single OmniEM model performs denoising, super-resolution, and organelle segmentation simultaneously, eliminating the need to maintain separate specialized models for each task.
  • Hallucination-Resistant Restoration: On image restoration benchmarks, OmniEM matches EM-specific diffusion models in reconstruction quality while producing fewer hallucinated ultrastructural features — a critical advantage when downstream biological interpretation depends on structural accuracy.
  • Integrated Super-Resolution Segmentation: OmniEM can generate high-resolution segmentation maps directly from low-resolution inputs, enabling fine-scale subcellular analysis on legacy datasets collected at lower magnifications.
  • Napari Plugin: An integrated Napari plugin provides a point-and-click interface for the full OmniEM pipeline, lowering the barrier to adoption for non-computational users.

Technical Details

EM-DINO is a vision transformer trained using a self-supervised DINO (self-distillation with no labels) objective on the EM-5M corpus. The DINO framework allows the model to learn meaningful representations without requiring manually annotated training examples, which are expensive to produce at the scale needed to cover EM data diversity. The resulting backbone produces multi-scale patch embeddings that encode both local texture and global contextual structure across a resolution range of 0.5 to 70 nm per pixel.

OmniEM is built as a U-Net-style encoder-decoder on top of the EM-DINO backbone. Skip connections between encoder and decoder stages preserve fine spatial detail while the transformer-encoded global context guides reconstruction and segmentation decisions. The model is trained jointly across all three task types — denoising, super-resolution, and segmentation — which encourages shared representations and reduces per-task overfitting. On segmentation benchmarks, OmniEM outperforms prior methods on both generalized mitochondrial segmentation across diverse datasets and multi-class organelle segmentation distinguishing multiple organelle types simultaneously.

Applications

OmniEM is designed for researchers who acquire or work with EM data and need to denoise images collected under low-dose or otherwise noisy conditions, enhance the apparent resolution of lower-magnification acquisitions, or segment organelles such as mitochondria and nuclei in datasets where training data for task-specific models is scarce. The integrated super-resolution segmentation capability is particularly valuable for mining legacy datasets that were collected at lower resolution but may contain irreplaceable biological information. The Napari plugin extends accessibility to imaging core facility users and wet-lab biologists, making OmniEM a practical tool not only for computational labs but also for groups that generate EM data as part of cell biology or structural studies.

Impact

OmniEM and the accompanying EM-DINO foundation model represent a significant step toward consolidating the fragmented tooling that currently characterizes EM image analysis. By demonstrating that a single pretrained backbone can support denoising, super-resolution, and segmentation across heterogeneous EM data, the work provides a template for foundation model approaches in scientific imaging more broadly. As a bioRxiv preprint from April 2025, the results have not yet undergone formal peer review, and the composition of the EM-5M corpus will require independent scrutiny to assess potential biases toward particular tissue types or imaging modalities. Performance on highly specialized protocols not well-represented in EM-5M — such as cryo-electron tomography subtomogram volumes — has not yet been characterized. The GitHub repository is publicly available, though the project is relatively new and community adoption and independent benchmarking remain ongoing.

Citation

Unifying the Electron Microscopy Multiverse through a Large-scale Foundation Model

Preprint

He, L., et al. (2025) Unifying the Electron Microscopy Multiverse through a Large-scale Foundation Model. bioRxiv.

DOI: 10.1101/2025.04.13.648639

Metrics

Citations

Total Citations2
Influential0
References37

Tags

segmentationsuper resolutionvision transformerfoundation modelelectron microscopy

Resources

GitHub RepositoryResearch Paper