Universal consensus 3D segmentation of cells from 2D segmented stacks

Nat Methods. 2025 Nov;22(11):2386-2399. doi: 10.1038/s41592-025-02887-w. Epub 2025 Nov 11.

Abstract

Cell segmentation is the foundation of a wide range of microscopy-based biological studies. Deep learning has revolutionized two-dimensional (2D) cell segmentation, enabling generalized solutions across cell types and imaging modalities. This has been driven by the ease of scaling up image acquisition, annotation and computation. However, three-dimensional (3D) cell segmentation, requiring dense annotation of 2D slices, still poses substantial challenges. Manual labeling of 3D cells to train broadly applicable segmentation models is prohibitive. Even in high-contrast images annotation is ambiguous and time-consuming. Here we develop a theory and toolbox, u-Segment3D, for 2D-to-3D segmentation, compatible with any 2D method generating pixel-based instance cell masks. u-Segment3D translates and enhances 2D instance segmentations to a 3D consensus instance segmentation without training data, as demonstrated on 11 real-life datasets, comprising >70,000 cells, spanning single cells, cell aggregates and tissue. Moreover, u-Segment3D is competitive with native 3D segmentation, even exceeding when cells are crowded and have complex morphologies.

MeSH terms

  • Algorithms
  • Animals
  • Deep Learning
  • Humans
  • Image Processing, Computer-Assisted* / methods
  • Imaging, Three-Dimensional* / methods
  • Single-Cell Analysis / methods