Abstract
Deep learning-based segmentation models can accelerate the analysis of high-throughput microscopy data by automatically identifying and classifying cells in images. However, the datasets needed to train these models are typically assembled via laborious hand-annotation. This limits their scale and diversity, which in turn limits model performance. We present Cell-APP (Cellular Annotation and Perception Pipeline), a tool that automates the annotation of high-quality training data for transmitted-light (TL) cell segmentation. Cell-APP uses two inputs-paired TL and nuclear fluorescence images-and operates in two main steps. First, it extracts each cell's location from the nuclear fluorescence channel and provides these locations to promptable deep learning models to generate cell masks. Then, it classifies each cell as mitotic or non-mitotic based on nuclear features. Together, these masks and classifications form the basis for cell segmentation training data. By training vision-transformer-based models on Cell-APP-generated datasets, we demonstrate how Cell-APP enables the creation of both cell line-specific and multi-cell line segmentation models. Cell-APP thus empowers laboratories to tailor cell segmentation models to their needs, and outlines a scalable path to creating general models for the research community.