Next Article in Journal
Cellular Senescence, Mitochondrial Dysfunction, and Their Link to Cardiovascular Disease
Previous Article in Journal
The Development and Characterization of a Next-Generation Oncolytic Virus Armed with an Anti-PD-1 sdAb for Osteosarcoma Treatment In Vitro
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Software Tools for 2D Cell Segmentation

by
Ping Liu
1,†,
Jun Li
1,2,†,
Jiaxing Chang
1,2,
Pinli Hu
2,
Yue Sun
2,
Yanan Jiang
2,
Fan Zhang
2 and
Haojing Shao
2,*
1
College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Jinzhong 030600, China
2
Shenzhen Branch, Guangdong Laboratory of Lingnan Modern Agriculture, Genome Analysis Laboratory of the Ministry of Agriculture and Rural Affairs, Agricultural Genomics Institute at Shenzhen, Chinese Academy of Agricultural Sciences, No 7, Pengfei Road, Dapeng District, Shenzhen 518120, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Cells 2024, 13(4), 352; https://doi.org/10.3390/cells13040352
Submission received: 9 December 2023 / Revised: 29 January 2024 / Accepted: 4 February 2024 / Published: 17 February 2024

Abstract

:
Cell segmentation is an important task in the field of image processing, widely used in the life sciences and medical fields. Traditional methods are mainly based on pixel intensity and spatial relationships, but have limitations. In recent years, machine learning and deep learning methods have been widely used, providing more-accurate and efficient solutions for cell segmentation. The effort to develop efficient and accurate segmentation software tools has been one of the major focal points in the field of cell segmentation for years. However, each software tool has unique characteristics and adaptations, and no universal cell-segmentation software can achieve perfect results. In this review, we used three publicly available datasets containing multiple 2D cell-imaging modalities. Common segmentation metrics were used to evaluate the performance of eight segmentation tools to compare their generality and, thus, find the best-performing tool.

1. Introduction

Cell segmentation is an important step for imaging studies and is widely used in the life sciences, bioinformatics, and biomedical fields such as oncology, immunology, and histopathology, including the emerging field of spatial transcriptomics. Scientists apply cell segmentation and prior knowledge of cell-type-specific gene expression to analyze the morphology and location of individual cells, obtain single-cell gene counts, and detect fine intracellular variations [1,2,3].
Differences in biological tissue, intercellular heterogeneity, and high cell densities need to be resolved before analyzing cell imaging data. Further, differences in illumination gradients, imaging modalities, and imaging parameters have to be accounted for while considering segmentation solutions [4,5]. In addition, microscopic techniques for cell imaging have improved, resulting in higher resolution, broader visualization, and noninvasive cell images. Common microscopic techniques include bright field microscopy, fluorescent microscopy, confocal microscopy, and phase contrast microscopy [6,7,8,9]. The more-recent imaging methods, such as cell staining combined with immunohistochemistry (IHC), multiplex immunofluorescence (mIF) [10], and CO-Detection by IndEXing (CODEX) [11] can co-detect and co-locate multiple transcriptomes and proteins, resulting in the precise annotation of individual cell types and the resolution of biological functions. In general, cell images obtained using such techniques have different channel types. The main image channels include three-channel RGB, three-channel HSV, single-channel GRAY, and fluorescently labeled specific channels, which are more commonly used by researchers.
Most of the traditional segmentation methods are based on the intensity and spatial relationship of pixels, and the constraint model is found by manual optimization, requiring expertise in basic techniques including code adaptation [12]. Code adaptation is highly subjective, and its development has reached a bottleneck. For example, the threshold only considers the grayscale information of the image and is sensitive to noise, which can easily cause uneven segmentation results; the region algorithm has low adaptability and performs poorly when used alone; the watershed algorithm is susceptible to over-segmentation due to noise; the graph theory algorithm is complex, computationally intensive, and not easy to operate. Therefore, using a sequential combination of these algorithms can effectively avoid these deficiencies and further separate touching or overlapping cells [13]. In addition, threshold processing and watershed algorithm are often used as preprocessing or postprocessing methods for Machine Learning (ML) and Deep Learning (DL).
ML and DL have similar workflows: the selection of the training data, data processing, model training, and model evaluation. These steps in the workflow may be iterative until the model is appropriate and accurate. HK-means [14], Random Forests [15], and EM [16] are trainable ML methods that include part of the knowledge in the segmentation process and improve the Accuracy of the segmentation. DL algorithms are better suited to addressing the challenges of cell segmentation, including multiple object morphologies and imaging techniques. DL network structures provide a generalized framework that can be applied to various tasks in different domains. They learn from data, adapt to different problem settings, and leverage the capabilities of pre-trained models, making them a convenient and effective strategy in many research and application fields. Further, manual tuning is not needed; however, retraining with annotated data is required [12].
The core algorithms in 2D segmentation tools are gradually shifting to more-complex deep learning networks. At the beginning, the earlier 2D segmentation tools CellProfiler [17] and Icy [14] used built-in traditional segmentation algorithms, such as watershed algorithms. Later, the classic U-net [18] structural deep learning model was wildly used and improved. StarDist added a polygon distance output layer [19,20]. Cellpose replaced the standard building blocks with residual blocks [21,22]. Notably, many 2D segmentation tools keep adoptingthe best segmentation models and update their software. For instance, CellProfiler and Icy update their deep learning model plugins (ClassifyPixels-U-net, DoGNet, etc.) for cell segmentation.
For this study, we used three publicly accessible datasets with annotations from several cell-imaging modalities to compare the generality of the tools.

2. Selected Software for Performance Comparison

As shown in Table 1, the selection of the segmentation software for the quantitative comparison was determined by three factors: (1) the ability to analyze cell images from various cell types; (2) those most-commonly used by scientists, especially for biological cell image segmentation; and (3) the feasibility of the installation, training, or use. The summary of the user-friendliness of the software is given in Table 2.

2.1. CellProfiler

CellProfiler (version 4.2.6) is a modular, high-throughput, and open-source software that provides a user-friendly interface, supports automatic cell identification and segmentation, and enables various feature extractions and data analyses, such as morphological measurements and cell counting. CellProfiler supports a variety of cell segmentation algorithms such as threshold algorithms, morphology algorithms, watershed algorithms, etc. It is convenient for users to choose the most-suitable algorithm or combination of algorithms according to different needs and adjust the parameters to improve the segmentation Accuracy and reduce errors. The software also supports segmentation using mainstream DL models via plugins. Scientists run CellProfiler headless or with a pipeline that supports batch processing to analyze multiple image datasets simultaneously and improve the Accuracy and consistency of large-scale data analyses. In general, a segmentation pipeline consists of four steps. First, it applies morphological operations to remove image noise. Second, it utilizes the gradient intensity of the image to obtain edge information. Third, the IdentifyPrimaryObjects module performs binarization to automatically detect and segment cells in the image and label them and further removes signal noise or fills gaps using certain morphological operations. Finally, it converts the cell masks to uint16 TIFF format images for preservation. The software, user manuals, video tutorials, and pipeline examples are available at: https://cellprofiler.org/ (accessed on 8 February 2024).

2.2. Icy

Icy (version 2.4.3.0) is a free, user-friendly, open-source image-analysis software suitable for high-throughput cells. It provides advanced analysis features: cell localization, segmentation, tracking, morphological measurements, etc. Among the plugins or modules for segmentation are Thresholding, Active Contour Line, Parametric Snake, Potts Segmentation, Texture Segmentation, Watershed3D, Spot Detector, and HK-Means [31]. Icy provides a Graphical User Interface (GUI) based on visual programming that allows users to select protocols from existing site libraries or manually configure their workflows for subsequent batch reuse. Java-based and Python-based scripts may be used instead of the plugin modules. For this study, we used a protocol and selected the commonly used HK-Means segmentation, which carries out N-class threshold processing based on the K-Means classification of the image histogram. The threshold processing uses a method based on similarity measurement to assign different brightness values to the cell images and their backgrounds. The object was then extracted from the bottom up using a user-defined minimum and maximum object sizes (pixels) to eliminate groups that are too small or too large. Finally, Gaussian pre-filtering was applied to improve the segmentation of noisy images. Icy is developed in Java. Its source code, standalone versions for Windows and Mac, documentation, user manuals, and video tutorials are distributed at: https://icy.bioimageanalysis.org/ (accessed on 8 February 2024).

2.3. StarDist

StarDist (version 0.8.5) is a DL method for 2D and 3D object detection and segmentation, which can handle complex shapes and structures with high Accuracy and robustness and is applied to the life science and medical image processing. It uses a Convolutional Neural Network (CNN) to predict the distance of each pixel to the object boundary and determines the probability of it belonging to the target object. The program generates a star-shaped convex polygon or convex polyhedron to represent the detected object. StarDist supports the detection of irregularly shaped objects against complex backgrounds. An adaptive model of the dataset is obtained through the application of a training on the original image and corresponding annotated data. The model is then used to test new data for prediction and evaluation. For 2D data, the pre-trained model can be used directly. It has no GUI, but provides a Python-based toolkit that can be used as a plugin for common open-source platforms such as ImageJ/Fiji [32], CellProfiler, Icy, KNIME, QuPath, etc. The source code is available at: https://github.com/stardist/stardist (accessed on 8 February 2024).

2.4. DeepCell

DeepCell [23,24,25,26] (version 0.12.9) is an open-source Python package for automated analysis and segmentation of high-throughput biological images. The software uses Deep Convolutional Neural Networks (DCNNs) to perform cell segmentation on various types of microscopy images and predict cell types. It includes several functions, including cell segmentation, nucleus segmentation, cell tracking, and phenotype measurement. DeepCell supports high customizability and migratory learning, and users can train their Neural Network architectures using pre-trained models or the provided design rules for image normalization, data enhancement, hyperparameter tuning, and post-segmentation processing to improve Accuracy. DeepCell also provides visualization tools to help researchers better understand the output data. It has no GUI, but offers alternative ways to interface with it, including a web portal, ImageJ plugins, and command-line interfaces. Its documentation, user manuals, and tutorials are distributed at: https://deepcell.com/ (accessed on 8 February 2024).

2.5. Cellpose

Cellpose (version 2.2.3) is an open-source general-purpose cell-image-segmentation software that can be applied to a wide range of biological imaging modalities and is suitable for identifying microorganisms and plant and animal cell types. Cellpose uses a modified Neural Network architecture based on U-net. Unlike classical segmentation, which uses image grayscale values to create topological maps, Cellpose uses intermediate image representations to form smoother topological maps to better segment multiple types of non-clumped cells. In addition, Cellpose 2.0 provides a large dataset of trained “model zoos” containing several pre-trained models where users can choose the appropriate pre-trained model and evaluate their data. Pre-training scripts are also provided so that users can efficiently train custom models. Cellpose inputs the approximate average cell diameter in pixels or takes the model’s default values for single-channel images. Image resolution is limited to a maximum of 512 × 512 px, effectively reducing the computation time while bringing fewer segmentation errors. It is provided with a web portal, and a GUI can be found at https://www.cellpose.org/ (accessed on 8 February 2024).

2.6. Omnipose

Omnipose [27] (version 1.0.6) is a user-friendly cell segmentation tool based on DL, using the U-net architecture. Omnipose is based on Cellpose and comes with significant upgrades, such as a new postprocessing step that removes noise and fine-tunes segmentation contours. In addition, it handles images with uneven lighting or blurred backgrounds and is optimized for fast processing while maintaining high Accuracy. Omnipose has been used to segment bacteria in large bacterial image databases. Omnipose can robustly segment bacteria with different morphologies, outperforms Cellpose on bacterial phase contrast images, and provides convincing results even for highly eccentric or elongated cells. In addition, Omnipose can accept images of any size, improving training efficiency and allowing more cell detection. The output of Omnipose is an instance segmentation. By integrating Omnipose with other tools, scientists can determine the average pixel diameter of bacterial cells from real surface masks. Omnipose offers a GUI, command-line interface, and Jupyter notebook at https://omnipose.readthedocs.io (accessed on 8 February 2024).

2.7. Plantseg

Plantseg [28] (version 1.6.0) is a comprehensive software for the 2D or 3D cell segmentation of plant cells. Plantseg utilizes a pipeline that involves a CNN to predict cell contour boundaries and various graph-partitioning algorithms to segment cells based on the previous step. The software was designed to work with confocal microscope and light-sheet microscope images. Users have the option of using a pre-trained network or training a 2D U-net or 3D U-net [33] architecture for the first step. The segmentation algorithms available in Plantseg include GASP [34] (default), Mutex watershed, MultiCut, SimpleTK, and DtWatershed. Plantseg also uses distance-transformation-based watershed and region adjacency graphs to create superpixels for image segmentation. These techniques differ from traditional grayscale segmentation methods, such as extracting DAPI-stained cell nuclei. Plantseg is a free, open-source software that comes equipped with both a command-line and a GUI, making it accessible to a wide range of users. The source code is available at https://github.com/hci-unihd/plant-seg (accessed on 8 February 2024).

2.8. Ilastik

Ilastik [29,30] (version 1.4.0.post1) is a user-friendly, free, open-source software tool for interactive ML-based image analysis in a wide range of applications including neuroscience and cell biology. It enables researchers to perform complex image segmentation, classification, tracking, and counting of cells. Scientists can download the BioImage.IO library (DL models for the bioimaging community) in Ilastik. An interactive ML based on a Random Forest classifier with a GUI that does not require specific ML knowledge adjusts parameters with minimal manual annotation and human intervention, enabling users to implement their image analysis through a supervised ML workflow. Manual annotation is used to predict the class of each unannotated pixel and object, thus classifying pixels and objects and correcting them precisely at the locations where the classifier is wrong. Once the classifier has been trained, the new data can be processed in batch mode. We only successfully tested Ilastik on the Cellpose_cyto dataset without preprocessing operations. First, we used the pixel classification workflow to obtain initial semantic segmentations (two classes: cell and background). Then, we used boundary-based segmentation with the multicut workflow to produce the final cell instance segmentation. Ilastik runs on Windows, macOS, and Linux. User documentation and video tutorials are available at: https://www.ilastik.org/ (accessed on 8 February 2024).

3. Experimental Configuration and Analysis

3.1. Datasets and Preprocessing

As shown in Table 3, three public datasets were selected to evaluate eight cell-segmentation software tools. The first dataset was extrapolated from the 2018 Data Science Bowl [35]; the second was obtained from Cellpose [21]; the third was obtained from the Cell Tracking Challenge of ISBI [36]. The general H5 format image files are easy to process and effectively reduce memory consumption. However, since Icy does not support the input of H5 format files yet, we did not choose the H5 format. Instead, we took another commonly used TIFF image format for conversion. Because CellProfiler and DeepCell require the same-size input dataset for batch processing, we unified the different sizes of the dataset to 512 × 512 px and converted them to a single channel (y,x,c) for grayscale processing. Specifically, the input dimensions of Plantseg were z, y, and x, and those of DeepCell were b, y, x, and c, where “bzxyc” denotes the number of images, depth, height–width, and color channel, respectively. Finally, we normalized the minimum and maximum values of each image in the dataset after removing the upper and lower 1% extremes. We were, thus, able to minimize the impact of these differences in the image brightness and contrast and to ensure the stability and Accuracy of the software cell segmentation. For examples, see Figure 1.

3.2. Hardware Environment

The software was tested using two different hardware environments, as described below:
  • GPU:
    • System: Ubuntu 20.04.2 LTS;
    • GPU: NVIDIA GeForce RTX 3090 24 GB 2 GPUs;
    • RAM: 1 TB;
    • CPU: AMD EPYC 7H12 64-Core Processor;
    • CUDA: version 11.1.
  • CPU:
    • System: Windows 10;
    • CPU: Intel(R) Core(TM) i5-6200U CPU @ 2.30 GHz;
    • RAM: 2 × 4 GB 1867 MHz/s.

3.3. Segmentation Metrics

As in Table 1, we chose eight software tools to compare, selecting pre-trained models and the existing pipelines of the corresponding software, which use default parameters for cell image segmentation in both the models and pipelines. There are several metrics related to image segmentation in the field of computer imaging and ML/DL. In this paper, we chose some generic segmentation metrics [7]: Accuracy, Recall, Precision, and F1. As shown in Table 3, the DSB2018 dataset uses 50 images and the Cellpose_cyto dataset uses 68 images for the average metrics comparison.
A c c u r a c y = T P + T N T P + T N + F N + F P R e c a l l = T P T P + F N P r e c i s i o n = T P T P + F P F 1 = 2 R e c a l l P r e c i s i o n R e c a l l + P r e c i s i o n = 2 T P 2 T P + T N + F N

4. Results

The default pre-trained model was selected for segmentation comparison. As shown in Table 4, Table 5, Table 6 and Table 7, we compared the segmentation results and conducted a quantitative comparison. Among these tables, N_true represents the number of cells in the Ground Truth (GT). N_pred represents the number of cells predicted by each software. Accuracy, Recall, Precision, and F1 are some segmentation metrics. F1-based Rank was used to compare the software performance based on the F1 score.
Without preprocessing, the top-three F1 scores of the software were for StarDist, Cellpose, and Omnipose. After the preprocessing operations, the F1 scores of the remaining five software tools, except for StarDist and Plantseg, improved, verifying the feasibility of the preprocessing operations. CellProfiler leverages the strengths of traditional algorithms, ranking first with an F1 score of 0.6394. StarDist was second with an F1 score of 0.5912, and Cellpose was third with an F1 score of 0.5763. The performance of the other software tools was comparatively low. The cell segmentation results using DeepCell expanded the boundary and, hence, it is more susceptible to noise, which makes the cells with obvious gaps appear adhesive and dense, resulting in over-segmentation. Icy is most affected by noise: the segmentation results contained holes and showed under-segmentation phenomena (Figure 2).
Without preprocessing, the top-three F1 scores of the software were for Cellpose, Omnipose, and StarDist. The same preprocessing operation did not give good results on the Cellpose_cyto dataset, the input of was a three-channel image, which may also be a result of the diversity of its image types. After processing, Cellpose demonstrated the best performance with an F1 score of 0.6929 for the Cellpose_cyto dataset. DeepCell ranked second with an F1 score of 0.3847, followed by StarDist in third place with an F1 score of 0.3072. The performance of the other software tools was notably lower, with Precision, Recall, Accuracy, and F1 scores all falling behind the top-three performers. The poorer results of Omnipose compared to Cellpose, which was highly affected by preprocessing, may be related to its suitability for regular (round or oval) small cells and bacterial cells. Plantseg had the worst results for both datasets, which may be because it is a plant-cell-segmentation software, which is more adapted to the segmentation of tightly arranged cells. For neuron cells, apart from Omnipose, Plantseg and StarDist were among the software tools with poor segmentation results, which are more sensitive to differences in cell shape and more adapted to round or oval conventional cells. However, Icy and CellProfiler are more adaptable for such cells, with finer segmentation (Figure 3).
In addition, we selected two software tools with the best F1 scores (Cellpose and StarDist) to compare the training adaptability of their models with the default parameters and methods using two preprocessing datasets. After model training, the segmentation metrics of the two software tools significantly increased, as shown in Table 6. The F1 score of Cellpose rose from 0.5763 to 0.7026, surpassing the F1 score of StarDist (0.5912) on the DSB2018 dataset. The F1 score of StarDist rose from 0.3072 to 0.6739, lower than the F1 score of Cellpose (0.6929) on the Cellpose_cyto dataset.
We tested an additional 2D dataset (PhC-C2DL-PSC) for cell segmentation with and without preprocessing operations. This dataset was phase contrast images from pancreatic stem cells on a polystyrene substrate. We chose 300 images from its training dataset (folder 01) and the corresponding masks in folder 01_ERR_SEG. Since this dataset has a uniform size, DeepCell and CellProfiler could also perform segmentation without preprocessing. Omnipose and Plantseg among the selected software were ignored as there was no output in this dataset. The results are as shown in Table 7 and Figure 4. Without preprocessing, the top-three F1 scores of the software were StarDist, Cellpose, and Icy. The same preprocessing operation did not give good results on this dataset, except for StarDist.
Finally, we quantitatively compared the computer resources used to obtain the segmentation results for the DSB2018 and the Cellpose_cyto datasets with preprocessing for each of the selected segmentation software, as shown in Table 8. In a GPU hardware environment, Cellpose ran fastest on the DSB2018 dataset and second fastest to Omnipose on the Cellpose_cyto dataset. The maximum memory occupied during StarDist processing was the lowest on both datasets, except for Plantseg. Plantseg took up the least memory, but ran the slowest on both datasets.

5. Discussion

This study examined the performance of different cell-image-segmentation software tools on the DSB2018, the Cellpose_cyto, and the PhC-C2DL-PSC datasets. The DSB2018 dataset is generic, and the overall shape of the cells in the dataset is relatively uniform with differences in the gap densities and cell sizes. The Cellpose_cyto dataset comprises cells with various characteristics. The PhC-C2DL-PSC dataset is consistent on cellular features as it is a 2D time-lapse sequence of cell images. The final results showed performance differences across the software tools, with no one tool performing better than the others across all the measures and datasets evaluated. CellProfiler and StarDist performed well on the DSB2018 dataset, while Cellpose performed well on the Cellpose_cyto dataset, and Cellpose and StarDist performed well on the PhC-C2DL-PSC dataset. After systematic model training and learning, Cellpose and StarDist were trained on their respective datasets and showed similar performance on the two datasets. The segmentation result of Cellpose was slightly better than that of StarDist, indicating that Cellpose has better adaptability to different types of cells. Cellpose and StarDist use distance-transformed gradients to predict the final result, process faster during segmentation, and consume less memory resource, which can satisfy the researchers’ need for the batch processing of cells. Meanwhile, a more-complete process can be constructed using deep learning model extension plug-ins such as Cellpose and StarDist in CellProfiler and the Icy software to achieve further statistical analysis of the segmented cells. Plantseg, Omnipose, and Icy showed limitations in working with specific types of cells. In addition, the adaptability of different software to preprocessing operations varies considerably, and it is not yet possible to choose a uniform preprocessing method to evaluate the performance of software under different datasets. Therefore, the specific requirements of the dataset and application scenarios should be considered when selecting the right software.
This study highlights the need for continuous development and improvement of cell-image-segmentation software. As technology continues to advance, further enhancements in algorithmic methods and optimization techniques may improve the performance of different datasets. In the future, it is expected that general cell-segmentation software with good segmentation results and advanced functions will be developed. These software tools can be used interactively with other software, such as the CellProfiler platform using Cellpose, StarDist, and other DL models for cell segmentation through plug-ins. The segmentation output generated by Icy can be imported into ImageJ for further statistical analysis and processing. It can also be controlled as a process script, which can analyze the cell structure more accurately and effectively and better serve the related biological research.

Author Contributions

Software, J.L. and P.H.; validation, J.C. and Y.S.; investigation, J.L.; resources, J.C.; data curation, Y.J.; writing—original draft preparation, P.L. and J.L.; writing—review and editing, F.Z.; supervision, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Natural Science Foundation of China (Grant number: 32200517 to H.S.).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MLMachine Learning
DLDeep Learning
GUIGraphical User Interface
CNNConvolutional Neural Network
DCNNDeep Convolutional Neural Network

References

  1. Larsson, L.; Frisén, J.; Lundeberg, J. Spatially resolved transcriptomics adds a new dimension to genomics. Nat. Methods 2021, 18, 15–18. [Google Scholar] [CrossRef]
  2. Bortolomeazzi, M.; Montorsi, L.; Temelkovski, D.; Keddar, M.R.; Acha-Sagredo, A.; Pitcher, M.J.; Basso, G.; Laghi, L.; Rodriguez-Justo, M.; Spencer, J.; et al. A SIMPLI (Single-cell Identification from MultiPLexed Images) approach for spatially-resolved tissue phenotyping at single-cell resolution. Nat. Commun. 2022, 13, 781. [Google Scholar] [CrossRef] [PubMed]
  3. Atta, L.; Fan, J. Computational challenges and opportunities in spatially resolved transcriptomic data analysis. Nat. Commun. 2021, 12, 5283. [Google Scholar] [CrossRef] [PubMed]
  4. Ren, H.; Zhao, M.; Liu, B.; Yao, R.; Liu, Q.; Ren, Z.; Wu, Z.; Gao, Z.; Yang, X.; Tang, C. Cellbow: A robust customizable cell segmentation program. Quant. Biol. 2020, 8, 245–255. [Google Scholar] [CrossRef]
  5. Hilsenbeck, O.; Schwarzfischer, M.; Loeffler, D.; Dimopoulos, S.; Hastreiter, S.; Marr, C.; Theis, F.J.; Schroeder, T. fastER: A user-friendly tool for ultrafast and robust cell segmentation in large-scale microscopy. Bioinformatics 2017, 33, 2020–2028. [Google Scholar] [CrossRef]
  6. Piccinini, F.; Balassa, T.; Carbonaro, A.; Diosdi, A.; Toth, T.; Moshkov, N.; Tasnadi, E.A.; Horvath, P. Software tools for 3D nuclei segmentation and quantitative analysis in multicellular aggregates. Comput. Struct. Biotechnol. J. 2020, 18, 1287–1300. [Google Scholar] [CrossRef]
  7. Maddalena, L.; Antonelli, L.; Albu, A.; Hada, A.; Guarracino, M.R. Artificial Intelligence for Cell Segmentation, Event Detection, and Tracking for Label-Free Microscopy Imaging. Algorithms 2022, 15, 313. [Google Scholar] [CrossRef]
  8. Winfree, S. User-Accessible Machine Learning Approaches for Cell Segmentation and Analysis in Tissue. Front. Physiol. 2022, 13, 87. [Google Scholar] [CrossRef]
  9. Hollandi, R.; Moshkov, N.; Paavolainen, L.; Tasnadi, E.; Piccinini, F.; Horvath, P. Nucleus segmentation: Towards automated solutions. Trends Cell Biol. 2022, 32, 295–310. [Google Scholar] [CrossRef]
  10. Taube, J.M.; Akturk, G.; Angelo, M.; Engle, E.L.; Gnjatic, S.; Greenbaum, S.; Greenwald, N.F.; Hedvat, C.V.; Hollmann, T.J.; Juco, J.; et al. The Society for Immunotherapy in Cancer statement on best practices for multiplex immunohistochemistry (IHC) and immunofluorescence (IF) staining and validation. J. Immunother. Cancer 2020, 8, e000155. [Google Scholar] [CrossRef]
  11. Kuswanto, W.; Nolan, G.; Lu, G. Highly multiplexed spatial profiling with CODEX: Bioinformatic analysis and application in human disease. In Proceedings of the Seminars in Immunopathology; Springer: Berlin/Heidelberg, Germany, 2023; Volume 45, pp. 145–157. [Google Scholar]
  12. Scherr, T.; Seiffarth, J.; Wollenhaupt, B.; Neumann, O.; Marcel, P.; Kohlheyer, D.; Scharr, H.; Nöh, K.; Mikut, R. microbeSEG: Accurate Cell Segmentation with OMERO Data Management. bioRxiv 2022. [Google Scholar] [CrossRef]
  13. Meijering, E. Cell segmentation: 50 years down the road [life sciences]. IEEE Signal Process. Mag. 2012, 29, 140–145. [Google Scholar] [CrossRef]
  14. De Chaumont, F.; Dallongeville, S.; Chenouard, N.; Hervé, N.; Pop, S.; Provoost, T.; Meas-Yedid, V.; Pankajakshan, P.; Lecomte, T.; Le Montagner, Y.; et al. Icy: An open bioimage informatics platform for extended reproducible research. Nat. Methods 2012, 9, 690–696. [Google Scholar] [CrossRef] [PubMed]
  15. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  16. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B (Methodol.) 1977, 39, 1–22. [Google Scholar] [CrossRef]
  17. Carpenter, A.E.; Jones, T.R.; Lamprecht, M.R.; Clarke, C.; Kang, I.H.; Friman, O.; Guertin, D.A.; Chang, J.H.; Lindquist, R.A.; Moffat, J.; et al. CellProfiler: Image analysis software for identifying and quantifying cell phenotypes. Genome Biol. 2006, 7, R100. [Google Scholar] [CrossRef]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  19. Schmidt, U.; Weigert, M.; Broaddus, C.; Myers, G. Cell detection with star-convex polygons. In Medical Image Computing and Computer Assisted Intervention—Proceedings of the MICCAI 2018: 21st International Conference, Granada, Spain, 16–20 September 2018, Proceedings, Part II 11; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  20. Weigert, M.; Schmidt, U.; Haase, R.; Sugawara, K.; Myers, G. Star-convex polyhedra for 3D object detection and segmentation in microscopy. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 3666–3673. [Google Scholar]
  21. Stringer, C.; Wang, T.; Michaelos, M.; Pachitariu, M. Cellpose: A generalist algorithm for cellular segmentation. Nat. Methods 2021, 18, 100–106. [Google Scholar] [CrossRef]
  22. Pachitariu, M.; Stringer, C. Cellpose 2.0: How to train your own model. Nat. Methods 2022, 19, 1634–1641. [Google Scholar] [CrossRef]
  23. Van Valen, D.A.; Kudo, T.; Lane, K.M.; Macklin, D.N.; Quach, N.T.; DeFelice, M.M.; Maayan, I.; Tanouchi, Y.; Ashley, E.A.; Covert, M.W. Deep learning automates the quantitative analysis of individual cells in live-cell imaging experiments. PLoS Comput. Biol. 2016, 12, e1005177. [Google Scholar] [CrossRef]
  24. Bannon, D.; Moen, E.; Schwartz, M.; Borba, E.; Kudo, T.; Greenwald, N.; Vijayakumar, V.; Chang, B.; Pao, E.; Osterman, E.; et al. DeepCell Kiosk: Scaling deep learning–enabled cellular image analysis with Kubernetes. Nat. Methods 2021, 18, 43–45. [Google Scholar] [CrossRef]
  25. Greenwald, N.F.; Miller, G.; Moen, E.; Kong, A.; Kagel, A.; Dougherty, T.; Fullaway, C.C.; McIntosh, B.J.; Leow, K.X.; Schwartz, M.S.; et al. Whole-cell segmentation of tissue images with human-level performance using large-scale data annotation and deep learning. Nat. Biotechnol. 2022, 40, 555–565. [Google Scholar] [CrossRef]
  26. Moen, E.; Borba, E.; Miller, G.; Schwartz, M.; Bannon, D.; Koe, N.; Camplisson, I.; Kyme, D.; Pavelchek, C.; Price, T.; et al. Accurate cell tracking and lineage construction in live-cell imaging experiments with deep learning. Biorxiv 2019, 803205. [Google Scholar] [CrossRef]
  27. Cutler, K.J.; Stringer, C.; Lo, T.W.; Rappez, L.; Stroustrup, N.; Brook Peterson, S.; Wiggins, P.A.; Mougous, J.D. Omnipose: A high-Precision morphology-independent solution for bacterial cell segmentation. Nat. Methods 2022, 19, 1438–1448. [Google Scholar] [CrossRef] [PubMed]
  28. Wolny, A.; Cerrone, L.; Vijayan, A.; Tofanelli, R.; Barro, A.V.; Louveaux, M.; Wenzl, C.; Strauss, S.; Wilson-Sánchez, D.; Lymbouridou, R.; et al. Accurate and versatile 3D segmentation of plant tissues at cellular resolution. Elife 2020, 9, e57613. [Google Scholar] [CrossRef]
  29. Sommer, C.; Straehle, C.; Koethe, U.; Hamprecht, F.A. Ilastik: Interactive learning and segmentation toolkit. In Proceedings of the 2011 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, Chicago, IL, USA, 30 March–2 April 2011; pp. 230–233. [Google Scholar]
  30. Berg, S.; Kutra, D.; Kroeger, T.; Straehle, C.N.; Kausler, B.X.; Haubold, C.; Schiegg, M.; Ales, J.; Beier, T.; Rudy, M.; et al. Ilastik: Interactive machine learning for (bio) image analysis. Nat. Methods 2019, 16, 1226–1232. [Google Scholar] [CrossRef] [PubMed]
  31. Dufour, A.; Meas-Yedid, V.; Grassart, A.; Olivo-Marin, J.C. Automated quantification of cell endocytosis using active contours and wavelets. In Proceedings of the 2008 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar]
  32. Schneider, C.A.; Rasband, W.S.; Eliceiri, K.W. NIH Image to ImageJ: 25 years of image analysis. Nat. Methods 2012, 9, 671–675. [Google Scholar] [CrossRef]
  33. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016: 19th International Conference, Athens, Greece, 17–21 October 2016; Proceedings, Part II 19. Springer: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar]
  34. Bailoni, A.; Pape, C.; Wolf, S.; Beier, T.; Kreshuk, A.; Hamprecht, F.A. A generalized framework for agglomerative clustering of signed graphs applied to instance segmentation. arXiv 2019, arXiv:1906.11713. [Google Scholar]
  35. Caicedo, J.C.; Goodman, A.; Karhohs, K.W.; Cimini, B.A.; Ackerman, J.; Haghighi, M.; Heng, C.; Becker, T.; Doan, M.; McQuin, C.; et al. Nucleus segmentation across imaging experiments: The 2018 Data Science Bowl. Nat. Methods 2019, 16, 1247–1253. [Google Scholar] [CrossRef] [PubMed]
  36. Ulman, V.; Maška, M.; Magnusson, K.E.; Ronneberger, O.; Haubold, C.; Harder, N.; Matula, P.; Matula, P.; Svoboda, D.; Radojevic, M.; et al. An objective comparison of cell-tracking algorithms. Nat. Methods 2017, 14, 1141–1152. [Google Scholar] [CrossRef]
Figure 1. An example of raw images and preprocessing result of DSB2018 datasets (left), Cellpose_cyto datasets (center), and PhC-C2DL-PSC datasets (right).
Figure 1. An example of raw images and preprocessing result of DSB2018 datasets (left), Cellpose_cyto datasets (center), and PhC-C2DL-PSC datasets (right).
Cells 13 00352 g001
Figure 2. Some representative sections of the segmentation masks obtained after preprocessing the DSB2018 dataset as shown in (AD). (E) shows the average cell numbers, and the suffix “*” indicates that the software had only successfully made segmentation predictions on preprocessed images. (F) shows FP/TP/FN values, and the prefix “un_” indicates that the software operates on non-preprocess images.
Figure 2. Some representative sections of the segmentation masks obtained after preprocessing the DSB2018 dataset as shown in (AD). (E) shows the average cell numbers, and the suffix “*” indicates that the software had only successfully made segmentation predictions on preprocessed images. (F) shows FP/TP/FN values, and the prefix “un_” indicates that the software operates on non-preprocess images.
Cells 13 00352 g002
Figure 3. Some representative sections of the segmentation masks obtained after preprocessing the Cellpose_cyto dataset as shown in (AF). (G) shows the cell numbers, and the suffix “*” indicates that the software had only successfully made segmentation predictions on preprocessed images. And (H) shows FP/TP/FN values, and the prefix “un_” indicates that the software operates on non-preprocess images.
Figure 3. Some representative sections of the segmentation masks obtained after preprocessing the Cellpose_cyto dataset as shown in (AF). (G) shows the cell numbers, and the suffix “*” indicates that the software had only successfully made segmentation predictions on preprocessed images. And (H) shows FP/TP/FN values, and the prefix “un_” indicates that the software operates on non-preprocess images.
Cells 13 00352 g003
Figure 4. Some representative sections of the segmentation masks obtained after preprocessing the PhC-C2DL-PSC dataset as shown in (A,B). (C) shows the cell numbers, and (D) shows FP/TP/FN values, and the prefix “un_ ” indicates that the software operates on non-preprocess images.
Figure 4. Some representative sections of the segmentation masks obtained after preprocessing the PhC-C2DL-PSC dataset as shown in (A,B). (C) shows the cell numbers, and (D) shows FP/TP/FN values, and the prefix “un_ ” indicates that the software operates on non-preprocess images.
Cells 13 00352 g004
Table 1. Basic information of the evaluated software tools.
Table 1. Basic information of the evaluated software tools.
ToolFirst Release TimeVersionProgramming
Language
Dependent
Library
ArchitecturePre-Trained ModelGUIPlugins for
Other Tools
FunctionPaperAverage Yearly Citations *
CellProfiler20064.2.6PythonScikit-imageNANA1NACell count, size, protein levels cell or organelle shape, and subcellular patterns of DNA or protein staining[17]311.6
Icy20112.4.3.0JavaVtk, TensorFlowNANA1NAVisualize, annotate, and quantify bioimaging data[14]121.2
StarDist20180.8.5PythonTensorFlowU-net2D_paper_DSB20180ImageJ/Fiji, Icy, KNIME, QuPath, NapariObject detection, multi-class prediction[19,20]165.2
DeepCell20180.12.9PythonTensorFlowDCNNNuclearSegmentation0ImageJ/FijiCell division, counting, classification, tracking, natural language processing, speech recognition[23,24,25,26]75.3
Cellpose20202.2.3PythonPytorchU-netcyto1CellProfiler, NapariCell and nucleus segmentation[21,22]432
Omnipose20221.0.6PythonPytorchU-netcyto2_omni1NapariBacterial cell segmentation[27]79.0
Plantseg20201.6.0PythonPytorchU-netconfocal_PNAS_2d1Napari, IlastikCell boundary predictions, graph partitioning[28]51.3
Ilastik20111.4.0.post1PythonScikit-imageNANA1ImageJ/FijiPixel classification, autocontext, object classification, carving, multicut, counting, tracking[29,30]110.1
* The analysis was performed on 5 December 2023 using Google Scholar. NA stands for not available or too diverse to be described individually.
Table 2. The user-friendliness of the software. Group A: documentation. Group B: usability. Group C: segmentation mode. Group D: output.
Table 2. The user-friendliness of the software. Group A: documentation. Group B: usability. Group C: segmentation mode. Group D: output.
Group CellProfilerIcyStarDistDeepCellCellposeOmniposePlantsegIlastik
AUser guide/handbook
website
Video tutorial
Community support
Test dataset/demo
Open-source
BNo programming experience is required
Intuitive visualization settings
Portability on Win/Linux/Mac
CManual
Interactive
Automated
D2D rendering
2D binary mask
Feature statistics
Table 3. The details of the three publicly available datasets.
Table 3. The details of the three publicly available datasets.
DatasetDescriptionTrainTestRAM (MB)Channels (n)Url
DSB2018one challenging dataset of diverse fluorescence microscopy images44750384.481https://www.kaggle.com/competitions/data-science-bowl-2018/data (accessed on 8 February 2024)
Cellpose_cytoone dataset consists of fluorescent cytoplasmic markers, confocal imaging, brightfield microscopy, and non-microscopy images54068161.303https://www.cellpose.org/dataset (accessed on 8 February 2024)
PhC-C2DL-PSCpancreatic stem cells on a polystyrene substrate6006002301http://celltrackingchallenge.net/2d-datasets (accessed on 8 February 2024)
Table 4. Quantitative comparison of the performance of open-source software tools selected for the study on DSB2018 datasets.
Table 4. Quantitative comparison of the performance of open-source software tools selected for the study on DSB2018 datasets.
SoftwareN_trueN_predPrecisionRecallAccuracyF1F1-Based Rank
No PreprocessingCellpose49420.93870.82270.79480.87252
Omnipose49290.78250.44210.40580.53873
StarDist49490.93140.92780.87230.92831
Plantseg4930.00500.00140.00110.00225
Icy49350.52240.43020.33170.46404
PreprocessingCellpose49480.91970.89590.84080.90661
Omnipose49290.76080.51500.49190.59455
StarDist49500.87270.91170.81770.88942
DeepCell49490.76410.80250.66250.78024
Plantseg4930.00070.00030.00020.00057
Icy49380.53000.49370.37700.50426
CellProfiler49470.80870.82150.70580.80893
Table 5. Quantitative comparison of the performance of selected open-source software tools on the Cellpose_cyto datasets.
Table 5. Quantitative comparison of the performance of selected open-source software tools on the Cellpose_cyto datasets.
SoftwareN_trueN_predPrecisionRecallAccuracyF1F1-Based Rank
No PreprocessingCellpose106650.85960.64560.61020.71381
Omnipose106870.76660.65790.55930.69412
StarDist1061160.52990.56090.38440.51423
Plantseg106900.14210.18550.10590.15345
Icy106630.19730.14270.10340.15844
Ilastik106600.04200.04740.02300.03706
PreprocessingCellpose106890.75520.66370.57980.69291
Omnipose10650.18000.04480.03840.06397
StarDist1061080.30370.35750.20350.30723
DeepCell1061290.35500.45960.26740.38472
Plantseg1061100.11750.14760.08300.12516
Icy106870.21030.18310.12600.19035
CellProfiler1061860.17760.24430.12670.19144
Table 6. Quantitative performance comparison of two software tools (Cellpose and StarDist) on 2D datasets.
Table 6. Quantitative performance comparison of two software tools (Cellpose and StarDist) on 2D datasets.
DatasetSoftwareN_trueN_predPrecisionRecallAccuracyF1
DSB2018StarDist49510.57070.61830.45660.5912
Cellpose (untrained)49470.59000.56470.46230.5763
Cellpose (trained)49450.73160.67910.57760.7026
Cellpose_cytoCellpose106890.75520.66370.57980.6929
StarDist (untrained)1061080.30370.35750.20350.3072
StarDist (trained)106930.70500.67080.54260.6739
Table 7. Quantitative comparison of the performance of selected open-source software tools on the PhC-C2DL-PSC datasets.
Table 7. Quantitative comparison of the performance of selected open-source software tools on the PhC-C2DL-PSC datasets.
SoftwareN_trueN_predPrecisionRecallAccuracyF1F1-Based Rank
No PreprocessingCellpose2371180.59190.44100.35040.48572
StarDist2372410.50250.59350.36680.53531
DeepCell2371120.08870.05030.03300.06345
Icy2371420.54270.38940.29920.44813
CellProfiler2373490.21790.35800.15370.26434
PreprocessingCellpose237750.40690.23120.18320.28473
StarDist2371750.57870.53270.37280.53831
DeepCell237610.05960.02060.01540.03015
Icy2371030.42790.26070.20150.31712
CellProfiler2372420.22380.28040.13860.24244
Table 8. Quantitative comparison of computer resource consumption of the selected cell-segmentation software.
Table 8. Quantitative comparison of computer resource consumption of the selected cell-segmentation software.
Softwareon GPU or CPUDSB2018Cellpose_cyto
Time (s)Memory (MB)Time (s)Memory (MB)
CellposeGPU32.895383.8948.855423.26
OmniposeGPU57.876320.0545.176315.86
StarDistGPU85.434533.08178.994628.75
DeepCellGPU110.436922.44112.247254.32
PlantsegGPU293.64234.08375.174274.43
IcyCPU132.85840.15221.01851.24
CellProfilerCPU660.02625.86900.34699.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, P.; Li, J.; Chang, J.; Hu, P.; Sun, Y.; Jiang, Y.; Zhang, F.; Shao, H. Software Tools for 2D Cell Segmentation. Cells 2024, 13, 352. https://doi.org/10.3390/cells13040352

AMA Style

Liu P, Li J, Chang J, Hu P, Sun Y, Jiang Y, Zhang F, Shao H. Software Tools for 2D Cell Segmentation. Cells. 2024; 13(4):352. https://doi.org/10.3390/cells13040352

Chicago/Turabian Style

Liu, Ping, Jun Li, Jiaxing Chang, Pinli Hu, Yue Sun, Yanan Jiang, Fan Zhang, and Haojing Shao. 2024. "Software Tools for 2D Cell Segmentation" Cells 13, no. 4: 352. https://doi.org/10.3390/cells13040352

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop