Affordable Access

deepdyve-link
Publisher Website

CEM500K, a large-scale heterogeneous unlabeled cellular electron microscopy image dataset for deep learning.

Authors
  • Conrad, Ryan1, 2
  • Narayan, Kedar1, 2
  • 1 Center for Molecular Microscopy, Center for Cancer Research, National Cancer Institute, National Institutes of Health, Bethesda, United States. , (United States)
  • 2 Cancer Research Technology Program, Frederick National Laboratory for Cancer Research, Frederick, United States. , (United States)
Type
Published Article
Journal
eLife
Publisher
"eLife Sciences Organisation, Ltd."
Publication Date
Apr 08, 2021
Volume
10
Identifiers
DOI: 10.7554/eLife.65894
PMID: 33830015
Source
Medline
Keywords
Language
English
License
Unknown

Abstract

Automated segmentation of cellular electron microscopy (EM) datasets remains a challenge. Supervised deep learning (DL) methods that rely on region-of-interest (ROI) annotations yield models that fail to generalize to unrelated datasets. Newer unsupervised DL algorithms require relevant pre-training images, however, pre-training on currently available EM datasets is computationally expensive and shows little value for unseen biological contexts, as these datasets are large and homogeneous. To address this issue, we present CEM500K, a nimble 25 GB dataset of 0.5 × 106 unique 2D cellular EM images curated from nearly 600 three-dimensional (3D) and 10,000 two-dimensional (2D) images from >100 unrelated imaging projects. We show that models pre-trained on CEM500K learn features that are biologically relevant and resilient to meaningful image augmentations. Critically, we evaluate transfer learning from these pre-trained models on six publicly available and one newly derived benchmark segmentation task and report state-of-the-art results on each. We release the CEM500K dataset, pre-trained models and curation pipeline for model building and further expansion by the EM community. Data and code are available at https://www.ebi.ac.uk/pdbe/emdb/empiar/entry/10592/ and https://git.io/JLLTz.

Report this publication

Statistics

Seen <100 times