Next Article in Journal
Sleep Action Recognition Based on Segmentation Strategy
Previous Article in Journal
A Novel Multimedia Player for International Standard—JPEG Snack
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Impact of Training Data, Ground Truth and Shape Variability in the Deep Learning-Based Semantic Segmentation of HeLa Cells Observed with Electron Microscopy

by
Cefa Karabağ
1,
Mauricio Alberto Ortega-Ruíz
1,2 and
Constantino Carlos Reyes-Aldasoro
1,*
1
giCentre, Department of Computer Science, School of Science and Technology, City, University of London, London EC1V 0HB, UK
2
Departamento de Ingeniería, Campus Coyoacán, Universidad del Valle de México, Ciudad de México C.P. 04910, Mexico
*
Author to whom correspondence should be addressed.
J. Imaging 2023, 9(3), 59; https://doi.org/10.3390/jimaging9030059
Submission received: 30 December 2022 / Revised: 16 February 2023 / Accepted: 17 February 2023 / Published: 1 March 2023
(This article belongs to the Topic Medical Image Analysis)

Abstract

:
This paper investigates the impact of the amount of training data and the shape variability on the segmentation provided by the deep learning architecture U-Net. Further, the correctness of ground truth (GT) was also evaluated. The input data consisted of a three-dimensional set of images of HeLa cells observed with an electron microscope with dimensions 8192 × 8192 × 517 . From there, a smaller region of interest (ROI) of 2000 × 2000 × 300 was cropped and manually delineated to obtain the ground truth necessary for a quantitative evaluation. A qualitative evaluation was performed on the 8192 × 8192 slices due to the lack of ground truth. Pairs of patches of data and labels for the classes nucleus, nuclear envelope, cell and background were generated to train U-Net architectures from scratch. Several training strategies were followed, and the results were compared against a traditional image processing algorithm. The correctness of GT, that is, the inclusion of one or more nuclei within the region of interest was also evaluated. The impact of the extent of training data was evaluated by comparing results from 36,000 pairs of data and label patches extracted from the odd slices in the central region, to 135,000 patches obtained from every other slice in the set. Then, 135,000 patches from several cells from the 8192 × 8192 slices were generated automatically using the image processing algorithm. Finally, the two sets of 135,000 pairs were combined to train once more with 270,000 pairs. As would be expected, the accuracy and Jaccard similarity index improved as the number of pairs increased for the ROI. This was also observed qualitatively for the 8192 × 8192 slices. When the 8192 × 8192 slices were segmented with U-Nets trained with 135,000 pairs, the architecture trained with automatically generated pairs provided better results than the architecture trained with the pairs from the manually segmented ground truths. This suggests that the pairs that were extracted automatically from many cells provided a better representation of the four classes of the various cells in the 8192 × 8192 slice than those pairs that were manually segmented from a single cell. Finally, the two sets of 135,000 pairs were combined, and the U-Net trained with these provided the best results.

1. Introduction

The immortal HeLa cell line, which originated from cervical cancer cells of the patient Henrietta Lacks, is the oldest and most commonly used human cell line [1,2]. These cells have been widely used in numerous experiments: from cancer [3,4,5,6,7] to toxoplasmosis [8,9,10,11] to AIDS [12,13,14,15] to radiation [16,17,18,19]. The story of these cells is not just a matter of scientific literature. Recently, it became the subject of books and films based on the life of Henrietta Lacks. The cells were extracted, stored and distributed without the consent or knowledge of the patient or her family, as this was not a requirement in 1951 in the United States. The remarkable and engaging story that the cells followed is narrated by Rebecca Skloot in her book The immortal life of Henrietta Lacks [20] and a film under the same name.
One area of interest in cancer research is the observation of the distribution, shape and morphological characteristics of the cells and cellular structures, such as the nuclear envelope (NE) and the plasma membrane [21,22,23,24]. Whilst it is possible to observe these characteristics of the cells visually, or to manually delineate the structures of interest [25,26,27,28], automatic segmentation of the cell and its structures is crucial for high throughput analysis where large amounts of data, i.e., terabytes, are acquired regularly.
Segmentation is an essential and difficult problem in image processing and machine vision for which a great number of algorithms have been proposed. The particular case of the segmentation and classification of cells and their structures remains as one important and challenging problem of interest for the clinical and programming communities [29,30]. Segmentation can be understood as a process in which an image is divided or partitioned into objects or regions, say o b j e c t / b a c k g r o u n d , or o b j e c t a / o b j e c t b / b a c k g r o u n d . The partition considers that the regions should be “non-intersecting, such that each region is homogeneous and the union of no two homogeneous regions is homogeneous” [31]. When the data are medical or biomedical, the regions usually correspond to anatomical or biological structures [32] e.g., cells, nuclei, cartilages or tumors. Some authors consider that the segmentation and subsequent reconstruction of the shape of medical organs or biological structures is harder than other computer vision problems because of inherent complexity, large shape variability, and characteristic artifacts of the acquisition systems [33]. The large number of segmentation methods in the literature is an indication of the complexity and importance of the problem. At the time of writing this paper (January 2023), PubMed reported more than 50,000 entries for “segmentation” (https://pubmed.ncbi.nlm.nih.gov/?term=(“segmentation”) (URL accessed 23 February 2023)).
Some of the most common segmentation techniques are the following. Grey level thresholding allocates a single or multiple thresholds so that pixels with a grey level lower than the threshold belong to one region and the remaining pixels to another region. The thresholding methods rely on the assumption that the objects to segment are distinctive in their grey levels and use the histogram information, thus ignoring the spatial arrangement of pixels. Whilst these techniques are probably the most basic ones, they are still widely used [34,35,36]. Iterative pixel classifications, such as Markov random fields (MRFs) rely on the spatial interrelation of the pixels with its neighbors. Energy functions, compatibility measures or training of networks determine the classification. In particular, MRFs [37,38,39] have been widely used in the classification of medical images with good results. Superpixel segmentation methods [40] group pixels that share certain characteristics and whilst doing so, reduce significantly the number of elements to be segmented or classified and thus have become a standard tool of many segmentation algorithms [40,41,42]. The watershed transform [43,44] considers not only the single intensity value of a pixel, but its relation with its neighbors in a topographical way. That is, regions are considered homogeneous when they belong to the same catchment basin through which idealized “rain” would flow toward a “lake”. The watershed provides a partition for very different thresholding, and thus is a very useful step of many segmentation algorithms [45,46]. Active contours define a curve or “snake”, which iteratively evolves to detect objects in an image based on constraints within the image itself [47]. Since the curve can evolve over a large number of iterations, the segmentation is very versatile, and even complex shapes, such as the corpus callosum [48], heart ventricles [49] or megakaryocitic cells [50] can be accurately segmented.
Machine learning and deep learning techniques have grown significantly in recent years [51] and a large number of architectures have been proposed with excellent results [52,53]. Whilst deep learning removes the need to hand-craft algorithms, one of the disadvantages is the scarcity of large datasets of training data. The training data of biomedical datasets are generally generated by manual delineation by a single expert, a group of experts or through citizen-science approaches [54]. Automatic labeling for the generation of training data is an attractive alternative to the time-consuming manual interventions. In areas different from biology, the use of synthetic datasets is a useful approach to generate training data [55,56]. Similarly, in some cases, the labels of well-defined objects are simplified, i.e., in the context of navigation, moving objects, such as cars or pedestrians, are labeled as moving, and all other objects are labeled as stationary [57]. In the context of synthetic aperture radar, it is possible to fuse together the image data themselves with a separate source of information, i.e., OpenStreetMap [58] or extra images from the Earth observation mission Sentinel-2 [59], to generate labels. In cell biology, and especially as observed with electron microscopy, none of these approaches are feasible due to the inherent variability of cells and other objects, which are less characteristic than a car or a pedestrian and the lack of alternative sources of information at the required resolution.
The segmentation of cells, their structures and the characteristics obtained from the segmentation, i.e., morphology or numbers, is crucial in the diagnosis of disease [60,61,62,63] and eventually can impact the treatment selected [64]. The segmentation of NE and the plasma membrane depends on the resolution of the acquisition equipment and the contrast it provides, as well complexity of the structures themselves. At higher resolutions and in three dimensions, such as those provided by electron microscopy, the problem is challenging [65,66], and manual delineation is used [67,68], sometimes through citizen-science approaches [54]. Segmentation with traditional image processing algorithms and deep learning approaches [69] are widely used in tasks of segmentation, and have previously been compared for the segmentation of NE and plasma membranes of HeLa cells as observed with electron microscopes [70].
Previous work on Hela cells by the authors focused initially on the development of an automated segmentation of HeLa nuclear envelope, i.e., a hand-crafted traditional image processing algorithm [71], and the modeling of the segmented three-dimensional nuclear envelope against a spheroid to create a two-dimensional (2D) surface [72] of a single cell, which was manually cropped from a larger volume ( 8192 × 8192 × 517 volume elements (voxels)) into a region of interest ( 2000 × 2000 × 300 ). The algorithm was then compared against four deep learning segmentation approaches, namely, VGG16, ResNet18, Inception-ResNet-v2 and U-Net [73]. The traditional image processing algorithm provided better results than the other approaches. However, it was observed that whilst the image processing algorithm segmented only the central nucleus, the U-Net segmented all nuclei that appeared in the images. Thus, the comparison was not like-with-like. A parallel work on HeLa cells [74] exploited the power of a citizen-science approach to enable volunteers to manually segment the nuclear envelope of the HeLa cells. The aggregated segmentations were then used to train a U-Net and segment 20 cells from the data. The image processing algorithm was further developed in [73,75] to segment the plasma membrane in addition to the nuclear envelope. Further, whilst the previous papers had worked with a region of interest, Ref. [75] expanded the scope to a larger dataset and performed an instance segmentation of 25 cells. It is important to highlight that a limitation of the algorithm is that the cells were detected separately and then segmented in an iterative manner. An approach that could segment all the cells in one slice would be an improvement over the algorithm.
In this work, the impact that the training data can have on the outcome of a segmentation was evaluated, and a comparison of the segmentations of HeLa plasma membranes and NE with a U-Net [76] was performed. Furthermore, the possibility of generating training data through the use of the automatic segmentation with a traditional image processing algorithm was explored. In this way, the training data were expanded to include pairs from the hand-segmented regions and the automatically-segmented regions.
Specifically, a region of interest (ROI) of 2000 × 2000 × 300 volume elements (voxels) was cropped from a larger dataset of 8192 × 8192 × 517 voxels, and was manually delineated to obtain a ground truth (GT). The ROI was segmented under the following scenarios: (a) training a U-Net from scratch with 36,000 patches for data and labels from alternate slices of the central region of the cell (101:2:180) considering that there was a single cell in the ROI, (b) training a U-Net from scratch considering the existence of multiple cells in the ROI and 36,000 patches for data and labels from alternate slices of the central region of the cell (101:2:180), and (c) training a U-Net from scratch considering a multiple cells in the ROI and 135,000 patches for data and labels from alternate slices of the whole cell (1:2:300). The U-Net model trained in (c) was then used to segment 8192 × 8192 slices, voxels and then (d) a new U-Net was trained from scratch considering the 135,000 pairs of patches from (c) plus 135,000 new patches that were generated automatically through the segmentation of an image processing algorithm. The results for the 2000 × 2000 × 300 images of the region of interest were compared slice-per-slice against a previously published traditional image processing algorithm using accuracy and the Jaccard similarity index. The results of the 8192 × 8192 slices were visually assessed, as there was no ground truth available. A graphical abstract of the work is presented in Figure 1. All the programming was performed in Matlab® (The MathworksTM, Natick, MA, USA).
Thus, the main contributions of this work are summarized as follows:
  • The impact of the amount and nature of training data on the segmentation of HeLa cells observed with an electron microscope was evaluated in quantitative and qualitative comparisons.
  • A methodology to automatically generate a ground truth using a traditional image processing algorithm is proposed. This ground truth was used to generate training pairs that were later used to train a U-Net. The ground truth was obtained from several cells in several 8192 × 8192 slices.
  • Data, code and ground truth were publicly released through Empiar, GitHub and Zenodo.

2. Materials and Methods

2.1. HeLa Cells Preparation and Acquisition

Details regarding the preparation of the HeLa cells have been previously described [75]. For completeness, these are briefly described here. Cells were embedded in Durcupan and observed with an SBF scanning electron microscope following the National Center for Microscopy and Imaging Research (NCMIR) protocol [77]. The images were acquired with a SBF SEM 3View2XP microscope (Gatan, Pleasanton, CA, USA) attached to a Sigma VP SEM (Zeiss, Cambridge, UK). Voxel dimensions were 10 × 10 × 50 nm with intensity values in the range [0–255]. Five hundred and seventeen images of 8192 × 8192 pixels were acquired. A ROI was cropped by estimating manually the centroid of one cell at the centre of a box of 2000 × 2000 × 300 voxels. Figure 2 illustrates a representative slice of one of the 8192 × 8192 images with the ROI indicated by a dashed blue square and a solid red line a larger region to illustrate the qualitative evaluation. Figure 3 illustrates several slices of the 2000 × 2000 ROI. All the EM dataset are publicly available through EMPIAR [78] at the following URL: http://dx.doi.org/10.6019/EMPIAR-10094 (accessed on 23 February 2023).

2.2. Ground Truth (GT)

Ground truth for the 2000 × 2000 × 300 ROI was manually delineated slice-per-slice in Matlab. Disjoint nuclear regions were assessed by scrolling up and down to decide whether to include them as part of the nucleus. The ground truth for the ROI with (a) only the central nucleus and (b) the central nucleus and other nuclei are publicly available through Zenodo: https://doi.org/10.5281/zenodo.3874949 (accessed on 23 February 2023) for single nucleus, and https://doi.org/10.5281/zenodo.6355622 (accessed on 23 February 2023) for multinuclei.

2.3. Traditional Image Processing Segmentation Algorithm

The semantic segmentations obtained with the different configurations of the U-Net were compared against a traditional image processing algorithm. Details of the algorithm have been described previously [72], but briefly, the algorithm is based on a pipeline formed of traditional image processing steps that are applied to a ROI that has been cropped from the whole dataset. The first step is low pass filtering to reduce the noise intrinsic to the acquisition of the electron microscope. Next, the areas of the image where the intensity changed abruptly were detected with the canny edge detection algorithm [79]. The nuclear envelope is not the only region of abrupt changes and thus, a large number of edges were detected. However, the edges were not used directly; rather, these were used to generate a series of superpixels which corresponded to all the areas of the image where edges did not exist. Again, a large number of superpixels were generated. Next, a series of morphological operations were applied. Specifically, small regions and those in contact with the edge of the image were discarded. Further, the central region of the image was selected, and then closing operations were used to smooth the final central region. This process was repeated on a slice-by-slice basis, starting from the central slice and propagating the results up and down the 3D stack. This process provided the combination of adjacent slices to select disjoint regions that belong to the nucleus. A simplified pipeline is illustrated graphically in Figure 4. This algorithm obtained accurate semantic segmentations, which outperformed other deep learning approaches [73]. However, in the case of U-Net, it was noticed that the inferior accuracy and Jaccard index were due to the discrepancy between the segmentation, which detected several nuclei, and the ground truth, which considered only the central nucleus. The Matlab code is available through GitHub: (https://github.com/reyesaldasoro/HeLa-Cell-Segmentation (accessed 23 February 2023)).

2.4. U-Net Architecture

The U-Net architecture [76] is a convolutional neural network architecture, which has been widely used for the segmentation of medical images. To cite just a few, U-Nets have been used in the segmentation of osteosarcoma in computed tomography scans [80], lung tumors in CT scans [81], lesions of the breast in ultrasound images [82], tumors of the bladder in cystoscopic images [83], breast and fibroglandular tissue in magnetic resonance [84] and nuclei in hematoxylin and eosin-stained slices [85]. The essence of the U-Net architecture is a combination of downsampling steps (also known as the contracting path) obtained by convolutions and downsampling, which are followed by upsampling steps (also known as the expansive path) (Figure 5). The contracting path is formed by a series of operations: 3 × 3 convolutions (i.e., the multiplication of each of the pixels of a small 3 × 3 grid of the image against a filter or kernel and then added together, then the grid is shifted, and the multiplication is repeated until all pixels of the image are covered); by using several kernels, many features can be obtained and the dimensions of the data increase; rectified linear units or ReLUs (i.e., an operation that converts to zero all negative values and maintains the positive values; this is used as an activation function); and a max-pooling operation, (i.e., the maximum value of the output is selected, by selecting one value out of several, the data is being reduced in dimensions). The expansive path follows the reverse process; the features are reduced, and the dimensions are expanded to return to the original size of the input image. There is a final layer in which classes are allocated to each pixel according to the feature vectors that have been created in the process. The number of times that the data are down- or up-sampled, or going up or down, determines the depth of the architecture, and due to the rough shape of a letter “U”, the architecture is called a U-Net.
For this work, a semantic segmentation framework with a U-Net architecture was implemented in Matlab. The U-Net hyperparameters are the following:
  • Adam optimizer [86],
  • four levels of depth, with 46 layers as shown in Figure 5,
  • size of the training patches = 128 × 128 pixels,
  • number of classes = 4 (background, cell, nuclear envelope, nucleus),
  • number of epochs for training = 15,
  • Initial Learn Rate = 1 × 10 3 ,
  • Mini Batch Size = 64.
The whole framework is illustrated in Figure 1. The hyper-parameters of the architecture followed previous work mentioned in the introduction in order to compare the findings of this work with previous results.

2.5. U-Net Training Data, Segmentation and Post-Processing

The U-Net architecture can be trained end-to-end from pairs or patches of images and their corresponding classes. Once trained, the U-Net can be used to segment the input images, which for this work consisted of the 300 images of the ROI (i.e., 2000 × 2000 ) and the 517 original images (i.e., 8192 × 8192 ). To train the network for this work, the following strategies were evaluated. Details of the strategies are illustrated in Figure 6.
1.
A total of 36,000 pairs from manually delineated GT from a single cell, evaluated with a single cell in the GT. Pairs of patches of images and labels of size 128 × 128 with 50% overlap were generated from 40 alternate slices of the central region of the cell (101:2:180). For one 2000 × 2000 image, there were 30 × 30 patches and thus 30 × 30 × 40 corresponded to 36,000 patches. Alternate slices were selected to exploit the similarity between neighboring slices. In this case, the ground truth included only the nucleus of the central cell visible in the ROI.
2.
A total of 36,000 pairs from manually delineated GT from a single cell, evaluated with multiple cells in the GT. The same strategy was followed to generate 36,000 patches and labels from alternate slices of the central region of the cell (101:2:180), however, in this case, the ground truth included the nuclei of all cells visible in the ROI.
3.
A total of 135,000 pairs from manually delineated GT from a single cell, evaluated with multiple cells in the GT. The pairs of patches of labels and data were extended to cover every other slice of the whole region of interest (1:2:300). The size was again 128 × 128 with 50% overlap; therefore, in this case, there were 150 slices and 900 patches per slice, which provided 135,000 pairs of patches for data and labels.
4.
A total of 135,000 pairs from automatically generated GT from multiple cells, evaluated visually. The training was extended beyond the 2000 × 2000 region of interest by performing an automatic segmentation of the 8192 × 8192 slices. This segmentation became a novel ground truth that was used to generate the same amount of pairs and in the previous strategy. The segmentation was performed with a traditional image processing segmentation algorithm [72] previously described. Fifteen non-contiguous slices were selected in the central region of the dataset (230:10:370). In each slice, the background was automatically segmented; distance transform was calculated to locate regions furthest from background, which corresponded to the cells. The 10 most salient cells were selected in each slice. A 2000 × 2000 region was cropped, automatically segmented, and patches of 128 × 128 with 50% overlap were generated. This generated 900 patches per cell, thus 900 × 10 × 15 = 135,000. This training strategy was designed for the segmentation of the 8192 × 8192 slices to compare the impact of segmenting with a U-Net trained on a single cell (even with a significant number of pairs) or with pairs from more than one cell.
5.
A total of 270,000 pairs from manual (135,000) and automatic (135,000) GTs, evaluated visually. Finally, the patches generated in the two previous strategies, that is, the 135,000 from the single cell and the 135,000 from the whole dataset were combined for a total of 270,000.
A simple post-processing step was applied. Each class was processed separately, and traditional image processing steps were applied: filling holes, closing with a disk structural element of 3-pixel radius, and removal of small regions (area <3200 pixels, equivalent to 0.08% of the area of the image). These steps removed small specs of larger uniform areas for the nucleus, cell and background. The nuclear envelope had an extra step: the nucleus and cell were dilated, and the overap between these two classes was added to the nuclear envelope. This removed the small discontinuities of the region. The last step aggregated the individual post-processed classes. Figure 7 illustrates the post-processing per class and the final outcome. The number of pixels modified by the post-processing is less than 1% of the data. All the Matlab code is available through GitHub: (https://github.com/reyesaldasoro/HeLa_Segmentation_UNET2, accessed on 23 February 2023).

2.6. Quantitative Comparisons

The segmentations assigned a class (background, nucleus, nuclear envelope, rest of the cell) to each pixel of the images. Accuracy and Jaccard similarity index (JI) [87] were calculated on a pixel-by-pixel basis for the 2000 × 2000 × 300 region of interest for which ground truth was available. For the 81 , 920 × 8192 slices, the results were visually assessed. Accuracy was calculated as ( T P + T N ) / ( T P + T N + F P + F N ) and JI, or intersection over union of nuclear area corresponded to ( T P ) / ( T P + F P + F N ) , where TP corresponds to true positives, TN to true negatives, FP to false positives and FN to false negatives.

2.7. Hardware Details

All the processing was performed in Matlab® (The Mathworks TM , Natick, MA, USA) and executed in a Dell Alienware m15 R3 Laptop with Intel® Core TM i9-10980HK CPU @ 2.40 GHz, 32 GB RAM, an NVIDIA® GeForce RTX 2070 Super GPU Card with 8 GB RAM.

3. Results and Discussion

3.1. Results on the 2000 × 2000 × 300 Region of Interest

Segmentation of all slices of the 2000 × 2000 × 300 region of interest with the first three training strategies (i.e., excluding the patches obtained with the image processing algorithm) and calculation of accuracy and Jaccard index was performed as described in previous sections. Figure 8, Figure 9 and Figure 10 illustrate the results obtained with the image processing algorithm and the U-Net with different training strategies.
Accuracy and Jaccard index results are shown in Figure 11 and in Table 1. Several points are worth mentioning. When the U-Net was trained and measured with GT with a single nucleus (solid blue line in Figure 11), the results were good between slices 150 and 200 of the ROI, as these slices contained only one nucleus. Toward the top and bottom of the ROI, the accuracy and JI dropped as other nuclei appeared in the images, and these were segmented, but were not part of the GT, and consequently, the numbers dropped. The results of the image processing algorithm (thin dashed black line in Figure 11) were very good, especially toward the central region of the cell, where the nucleus is largest. These results considered only the central cell. Toward the top and bottom of the cell, where the nucleus is smaller and the shape less regular, the accuracy remained high due to the large number of TN, but the JI dropped in a similar way to the U-Net with a single nucleus and was zero in the extremes due to the absence of TP.
For the U-Net trained with multiple nuclei and 36,000 patches (dotted red line in Figure 11), the results were significantly higher than the single nucleus U-Net, close to the image processing algorithm in central slices and better in the top/bottom, as both metrics remained high. However, for the U-Net with 135,000 patches (thick green solid line in Figure 11), the results were higher still, especially in the regions outside the slices from which the 36,000 patches were generated (i.e., 1:2:99 and 181:2:300). The average accuracy for all slices for the four algorithms was as follows: U-Net single nucleus 36,000 patches, 0.9346, U-Net multiple nuclei 36,000 patches, 0.9895, U-Net multiple nuclei, 135,000 patches, 0.9922, and image processing algorithm 0.9926. In general, all these results are very high for all cases where the appropriate GT is considered, but the high values also indicate a large number of TN, especially on the top and bottom slices. For the central slices (150:200), where there is a single cell and results are best, the values are all above 99% (same order 0.9966, 0.9974, 0.9971, 0.9945); thus, the errors are due to small variations on the edge of the nucleus or the fact that the manual delineation can also include small errors. The values for the Jaccard index are more interesting, with the average values for all slices of (0.5138, 0.9158, 0.9378, 0.6436), it is clear that the multinuclei U-Nets have a significant improvement, as they consider better the top and bottom slices. For the central slices (150:200), the values are pretty similar for the three U-Nets and slightly lower for the image processing (0.9712, 0.9778, 0.9760, 0.9564), as a small dip is visible around slice 180. A final comparison in the Jaccard, between slices 60:150, where the curves are very similar, shows the following values (0.8047, 0.9579, 0.9592, 0.9565), that is, nearly identical averages for all methods, except the initial U-Net. These results confirm that the larger the training data, the better the U-Net can learn the characteristics of the cells. It is important to highlight that the results are not completely like-with-like, as the GT is not the same for the single nucleus and the multinuclei approaches.

3.2. Results on the 8192 × 8192 Slices

As a further segmentation test, the U-Nets trained on multiple nuclei were used to segment the 8192 × 8192 slices. The results were evaluated visually due to the lack of ground truth. Figure 12 shows the results of the U-Nets and the image processing algorithm. Figure 12a shows the segmentation obtained with the U-Net trained with 36,000 pairs, Figure 12b shows the segmentation with 135,000 pairs from the single cell from the ROI, Figure 12c shows the segmentation with 135,000 pairs automatically generated with the image processing algorithm, Figure 12d shows the segmentation with 270,000 pairs, and Figure 12e shows the segmentation with the image processing algorithm.
The first observation is that as the training sets grow, the U-Nets provide better results than those trained with smaller sets. This improvement can be observed in several locations: The cells on the top right show a discontinuous nucleus in (a), then a nucleus with holes in (b). These less than satisfactory results on (a,b) may be the outcome of training the U-Nets on a single cell, whose nucleus is considerably smoother than those of the surrounding cells. The second observation is the improvement that is provided with the pairs generated automatically with the image processing algorithm (c). The nuclei on the cells on the top right are much better segmented (dashed arrows). The cell from the ROI on the other hand (bottom left, solid arrows) now presents some small holes inside the nucleus. This is not surprising, as the training data are not from this cell in particular. The results shown in (d) combined the pairs from the ROI and the automatically generated and show improvement in the cell from the ROI and other cells. The sharper delineation of the invaginations should be noticed. These can be compared with the image processing algorithm (e), which provided results equivalent to the segmentation with the network trained with 270,000 pairs.
Training with a larger number of training patches improved the results, but the key element is not necessarily the number of training pairs, but the fact that the training pairs were extracted from different cells. This is important, as just a larger number of samples or an enlargement with augmentation methods will not necessarily improve results unless the samples are representative of the data to be segmented.
The image processing algorithm was run on an instance segmentation, that is, individual cells in 2000 × 2000 regions of interest were automatically cropped and segmented. Only a reduced number of cells per slice were selected, which can be noticed in that some cells (bottom right and left) are not segmented and appear grey. It can also be noticed that some cellular regions are not segmented, especially those that are between cells. However, the shapes of the nuclei appear well segmented, with the exception of the second cell from the left on the top, for which a section of the nuclei was not detected.

4. Conclusions

The quantitative and qualitative comparisons of different semantic segmentation approaches of HeLa cells, their nucleus, nuclear envelope, cell and background, are presented in this work. Whilst, as would be expected, the fact that a larger amount of training data provides more accurate segmentation was sustained; it was further observed that the variability of the cells, even when it is a single line of cells, precludes the generalization that training data can be extended from one cell to another without degradation of the results. The addition of training data from a variety of cells improved considerably the segmentation results, and the combination of two sources further improved the results. Moreover, the consideration of the correct ground truth can be extremely important in the assessment of results of different approaches. An immediate observation is that this research is limited to a single dataset, and even when considerably large, with numerous cells, it is just one dataset. The cells themselves show significant variability as can be seen in Figure 12. It is possible that at lower resolutions, this variability is less significant, and thus, the data used to train a U-Net may have lower impact in the results.
In the context of this work, the main advantage of the U-Net architecture (against an image processing algorithm) is the capability to segment the images at full resolution ( 8192 × 8192 ), without the need to crop into regions of interest select individual cells. Whilst it was not developed in this work, a semantic segmentation of the cells could be integrated as a post-processing step of the U-Net framework. The first disadvantage of the U-Net algorithm is the requirement of training data, which many times requires manual labeling. Further, the amount and diversity of the training data can have a strong impact on the segmentation results. However, as it was demonstrated here, the negative impact can be reduced by increasing the training data with pairs generated automatically by the image processing algorithm.
The main advantage of the image processing algorithm is that it is an automatic process when a region of interest of 2000 × 2000 × 300 voxels with a single cell in the center is provided as input data. There is no need for training data. To process images at full resolution, the algorithm can detect cells and crop automatically, but in some cases, the region of interest may contain more than one cell or contain only part of the cell, especially when close to the edges of the dataset. Whilst a manual interaction is required to determine the number of cells to be processed, the segmentation results are instance based, as the cells are identified as individual cells. The main disadvantage of the algorithm is that it cannot process a dataset with multiple cells, whether this is a crop or a full resolution set. When processing the full dataset, the algorithm requires cells to be cropped into individual volumes, where it is assumed the cell will be centered. The algorithm will only segment one cell per volume.
Future work can be divided in five parallel lines. First, the proposed U-Net framework can be developed to consider the identification of individual cells for an instance-based segmentation. Given that the framework identifies nuclei, it would not be difficult to identify the number of these, which will in turn identify the number of cells present. The difficulty would be in the separation of individual cells, as was performed by traditional image processing algorithms on a cell-by-cell basis [75]. The algorithm first identified cells, cropped regions of interest per cell and segmented one cell per cropped region of interest.
Second, the identification of other sub-cellular structures, such as mitochondria and the Golgi apparatus would be of great interest and could be incorporated into the framework. For the U-Net, a number of training pairs of images and patches, which include the organelles, should be generated. An image processing approach can exploit the visual characteristics, such as the round structure crossed by lines shown by the mitochondria.
Third, the image processing algorithm and U-Net architecture were applied directly, following the use of these in previous publications. Optimization of both approaches could be investigated. For the image processing algorithm, a sensitivity analysis of key parameters, such as structural elements, could be performed. For the U-Net, different optimization techniques could be explored. For example, broad ablation studies could be performed; exploration of the number of layers, epochs, optimizer, loss function, residual connections or decoder attention [88,89,90]. The modification or addition of layers is also a possibility, for instance, interpolation and convolution instead of transposed convolution [91], adding deconvolution and upsampling layers in the splicing process [92], or fuzzy layers in addition to the conventional layers [93].
Fourth, a limitation of the current segmentation is that it was performed on a slice-by-slice basis, without fully exploiting the three-dimensional nature of the data. The image processing algorithm does exploit the 3D by comparing the results of neighboring slices and thus propagating the results. Another option could be to perform a rotation of axes so that segmentations in other planes can be performed and the results combined. This was previously proposed as a “tri-axis prediction” [70].
Fifth, it was previously mentioned that data augmentation methodologies may not necessarily improve the training if the pairs obtained from one cell are not representative of other cells. However, a thorough investigation of different augmentations: crop and resize, rotation, cutout, flips, Gaussian noise and blur, Sobel filter [94], elastic or warping deformations [95], or deep learning methods, e.g., generative adversarial networks [96] could be considered.

Author Contributions

Conceptualization, C.K. and C.C.R.-A.; methodology, C.K. and C.C.R.-A.; software, C.K. and C.C.R.-A.; validation, C.K., M.A.O.-R. and C.C.R.-A.; formal analysis, C.K. and C.C.R.-A.; investigation, C.K.; resources, M.A.O.-R.; writing—original draft preparation, C.K. and C.C.R.-A.; writing—review and editing, C.K., M.A.O.-R. and C.C.R.-A.; supervision, C.C.R.-A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Electron microscopy datasets are available at http://dx.doi.org/10.6019/EMPIAR-10094, ground truths is available at https://doi.org/10.5281/zenodo.3874949, https://doi.org/10.5281/zenodo.6355622. Code is available at https://github.com/reyesaldasoro/Hela-Cell-Segmentation, https://github.com/reyesaldasoro/HeLa_Segmentation_UNET2, all accessed on 24 October 2022. The HeLa cell line was obtained from The Francis Crick Institute and prepared by Christopher J. Peddie.

Acknowledgments

The authors acknowledge the useful discussions with Christopher J. Peddie, Martin L. Jones and Lucy M. Collinson from The Francis Crick Institute, UK.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Masters, J.R. HeLa cells 50 years on: The good, the bad and the ugly. Nat. Rev. Cancer 2002, 2, 315–319. [Google Scholar] [CrossRef] [PubMed]
  2. Rahbari, R.; Sheahan, T.; Modes, V.; Collier, P.; Macfarlane, C.; Badge, R.M. A novel L1 retrotransposon marker for HeLa cell line identification. BioTechniques 2009, 46, 277–284. [Google Scholar] [CrossRef] [PubMed]
  3. Yung, B.Y.; Bor, A.M. Identification of high-density lipoprotein in serum to determine anti-cancer efficacy of doxorubicin in HeLa cells. Int. J. Cancer 1992, 50, 951–957. [Google Scholar] [CrossRef]
  4. Zhang, S.L.; Wang, Y.S.; Zhou, T.; Yu, X.W.; Wei, Z.T.; Li, Y.L. Isolation and characterization of cancer stem cells from cervical cancer HeLa cells. Cytotechnology 2012, 64, 477–484. [Google Scholar] [CrossRef] [Green Version]
  5. Yang, J.; Zhang, B.; Qin, Z.; Li, S.; Xu, J.; Yao, Z.; Zhang, X.; Gonzalez, F.J.; Yao, X. Efflux excretion of bisdemethoxycurcumin-O-glucuronide in UGT1A1-overexpressing HeLa cells: Identification of breast cancer resistance protein (BCRP) and multidrug resistance-associated proteins 1 (MRP1) as the glucuronide transporters. Biofactors 2018, 44, 558–569. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, L.; Qi, Y.; Xiong, Y.; Peng, Z.; Ma, Q.; Zhang, Y.; Song, J.; Zheng, J. Ezrin-Radixin-Moesin Binding Phosphoprotein 50 (EBP50) Suppresses the Metastasis of Breast Cancer and HeLa Cells by Inhibiting Matrix Metalloproteinase-2 Activity. Anticancer Res. 2017, 37, 4353–4360. [Google Scholar] [PubMed] [Green Version]
  7. Zukić, S.; Maran, U. Modelling of antiproliferative activity measured in HeLa cervical cancer cells in a series of xanthene derivatives. SAR QSAR Environ. Res. 2020, 31, 905–921. [Google Scholar] [CrossRef] [PubMed]
  8. Yang, Z.; Ahn, H.J.; Nam, H.W. Gefitinib inhibits the growth of Toxoplasma gondii in HeLa cells. Korean J. Parasitol. 2014, 52, 439–441. [Google Scholar] [CrossRef] [Green Version]
  9. Sanfelice, R.A.; Machado, L.F.; Bosqui, L.R.; Miranda-Sapla, M.M.; Tomiotto-Pellissier, F.; de Alcântara Dalevedo, G.; Ioris, D.; Reis, G.F.; Panagio, L.A.; Navarro, I.T.; et al. Activity of rosuvastatin in tachyzoites of Toxoplasma gondii (RH strain) in HeLa cells. Exp Parasitol 2017, 181, 75–81. [Google Scholar] [CrossRef]
  10. Zhang, Z.; Gu, H.; Li, Q.; Zheng, J.; Cao, S.; Weng, C.; Jia, H. GABARAPL2 Is Critical for Growth Restriction of Toxoplasma gondii in HeLa Cells Treated with Gamma Interferon. Infect. Immun. 2020, 88, e00054-20. [Google Scholar] [CrossRef]
  11. Pan, L.; Yang, Y.; Chen, X.; Zhao, M.; Yao, C.; Sheng, K.; Yang, Y.; Ma, G.; Du, A. Host autophagy limits Toxoplasma gondii proliferation in the absence of IFN-γ by affecting the hijack of Rab11A-positive vesicles. Front. Microbiol. 2022, 13, 1052779. [Google Scholar] [CrossRef] [PubMed]
  12. Tominaga, M.; Kumagai, E.; Harada, S. Effect of electrical stimulation on HIV-1-infected HeLa cells cultured on an electrode surface. Appl. Microbiol. Biotechnol. 2003, 61, 447–450. [Google Scholar] [CrossRef] [PubMed]
  13. Zheng, Y.; Zhou, H.; Zhang, C.; He, Y.; Li, H.; Chen, Z.; Liu, M. The apoptosis-inducing effects of HIV Vpr recombinant eukaryotic expression vectors with different mutation sites on transfected Hela cells. Curr. HIV Res. 2009, 7, 519–525. [Google Scholar] [CrossRef] [PubMed]
  14. Chesebro, B.; Wehrly, K.; Metcalf, J.; Griffin, D.E. Use of a new CD4-positive HeLa cell clone for direct quantitation of infectious human immunodeficiency virus from blood cells of AIDS patients. J. Infect. Dis. 1991, 163, 64–70. [Google Scholar] [CrossRef] [PubMed]
  15. Ekama, S.O.; Ilomuanya, M.O.; Azubuike, C.P.; Ayorinde, J.B.; Ezechi, O.C.; Igwilo, C.I.; Salako, B.L. Enzyme Responsive Vaginal Microbicide Gels Containing Maraviroc and Tenofovir Microspheres Designed for Acid Phosphatase-Triggered Release for Pre-Exposure Prophylaxis of HIV-1: A Comparative Analysis of a Bigel and Thermosensitive Gel. Gels 2021, 8, 15. [Google Scholar] [CrossRef] [PubMed]
  16. Zhang, B.; Liu, J.Y.; Pan, J.S.; Han, S.P.; Yin, X.X.; Wang, B.; Hu, G. Combined treatment of ionizing radiation with genistein on cervical cancer HeLa cells. J. Pharmacol. Sci. 2006, 102, 129–135. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Ziegler, W.; Birkenfeld, P.; Trott, K.R. The effect of combined treatment of HeLa cells with actinomycin D and radiation upon survival and recovery from radiation damage. Radiother. Oncol. 1987, 10, 141–148. [Google Scholar] [CrossRef]
  18. Zinberg, N.; Kohn, A. Dimethyl sulfoxide protection of HeLa cells against ionizing radiation during the growth cycle. Isr. J. Med. Sci. 1971, 7, 719–723. [Google Scholar]
  19. Zhu, C.; Wang, X.; Li, P.; Zhu, Y.; Sun, Y.; Hu, J.; Liu, H.; Sun, X. Developing a Peptide That Inhibits DNA Repair by Blocking the Binding of Artemis and DNA Ligase IV to Enhance Tumor Radiosensitivity. Int. J. Radiat. Oncol. 2021, 111, 515–527. [Google Scholar] [CrossRef]
  20. Skloot, R. The Immortal Life of Henrietta Lacks; Crown: New York, NY, USA, 2010. [Google Scholar]
  21. Rohde, G.K.; Ribeiro, A.J.S.; Dahl, K.N.; Murphy, R.F. Deformation-based nuclear morphometry: Capturing nuclear shape variation in HeLa cells. Cytometry. Part A J. Int. Soc. Anal. Cytol. 2008, 73, 341–350. [Google Scholar] [CrossRef]
  22. Suzuki, R.; Hotta, K.; Oka, K. Spatiotemporal quantification of subcellular ATP levels in a single HeLa cell during changes in morphology. Sci. Rep. 2015, 5, 16874. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Van Peteghem, M.C.; Mareel, M.M. Alterations in shape, surface structure and cytoskeleton of HeLa cells during monolayer culture. Arch. Biol. 1978, 89, 67–87. [Google Scholar]
  24. Welter, D.A.; Black, D.A.; Hodge, L.D. Nuclear reformation following metaphase in HeLa S3 cells: Three-dimensional visualization of chromatid rearrangements. Chromosoma 1985, 93, 57–68. [Google Scholar] [CrossRef] [PubMed]
  25. Bajcsy, P.; Cardone, A.; Chalfoun, J.; Halter, M.; Juba, D.; Kociolek, M.; Majurski, M.; Peskin, A.; Simon, C.; Simon, M.; et al. Survey statistics of automated segmentations applied to optical imaging of mammalian cells. BMC Bioinform. 2015, 16, 330. [Google Scholar] [CrossRef] [Green Version]
  26. Perez, A.; Seyedhosseini, M.; Deerinck, T.; Bushong, E.; Panda, S.; Tasdizen, T.; Ellisman, M. A workflow for the automatic segmentation of organelles in electron microscopy image stacks. Front. Neuroanat. 2014, 8, 1–13. [Google Scholar] [CrossRef]
  27. Wilke, S.; Antonios, J.; Bushong, E.; Badkoobehi, A.; Malek, E.; Hwang, M. Deconstructing complexity: Serial block-face electron microscopic analysis of the hippocampal mossy fiber synapse. J. Neurosci. 2013, 33, 507–522. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Bohorquez, D.; Samsa, L.; Roholt, A.; Medicetty, S.; Chandra, R.; Liddle, R. An enteroendocrine cell-enteric glia connection revealed by 3D electron microscopy. PLoS ONE 2014, 9, e89881. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Gurcan, M.N.; Boucheron, L.; Can, A.; Madabhushi, A.; Rajpoot, N.; Yener, B. Histopathological Image Analysis: A Review. IEEE Rev. Biomed. Eng. 2009, 2, 147–171. [Google Scholar] [CrossRef] [Green Version]
  30. Wang, Z.; Li, H. Generalizing cell segmentation and quantification. BMC Bioinform. 2017, 18, 189. [Google Scholar] [CrossRef] [Green Version]
  31. Pal, N.R.; Pal, S.R. A Review on Image Segmentation Techniques. Pattern Recognit. 1993, 26, 1277–1293. [Google Scholar] [CrossRef]
  32. Kapur, T. Model Based Three Dimensional Medical Image Segmentation. Ph.D. Thesis, AI Lab, Massachusetts Institute of Technology, Cambridge, MA, USA, 1999. [Google Scholar]
  33. Suri, J.S. Two-Dimensional Fast Magnetic Resonance Brain Segmentation. IEEE Eng. Med. Biol. 2001, 20, 84–95. [Google Scholar] [CrossRef] [PubMed]
  34. Zulfira, F.Z.; Suyanto, S.; Septiarini, A. Segmentation technique and dynamic ensemble selection to enhance glaucoma severity detection. Comput. Biol. Med. 2021, 139, 104951. [Google Scholar] [CrossRef] [PubMed]
  35. Zibrandtsen, I.C.; Kjaer, T.W. Fully automatic peak frequency estimation of the posterior dominant rhythm in a large retrospective hospital EEG cohort. Clin. Neurophysiol. Pract. 2021, 6, 1–9. [Google Scholar] [CrossRef] [PubMed]
  36. Xiong, F.; Zhang, Z.; Ling, Y.; Zhang, J. Image thresholding segmentation based on weighted Parzen-window and linear programming techniques. Sci. Rep. 2022, 12, 13635. [Google Scholar] [CrossRef]
  37. Held, K.; Kops, E.R.; Krause, B.J.; Wells, W.M.; Kikinis, R.; Muller-Gartner, H.W. Markov Random Field Segmentation of Brain MR Images. IEEE Trans. Med. Imaging 1997, 16, 6. [Google Scholar] [CrossRef] [Green Version]
  38. Zhang, Y.; Brady, M.; Smith, S. Segmentation of brain MR Images through a hidden Markov random field model and expectation maximization algorithm. IEEE Trans. Med. Imaging 2001, 20, 45–57. [Google Scholar] [CrossRef]
  39. Zhang, Z.; Zhao, T.; Gay, H.; Zhang, W.; Sun, B. ARPM-net: A novel CNN-based adversarial method with Markov random field enhancement for prostate and organs at risk segmentation in pelvic CT images. Med. Phys. 2020, 48, 227–237. [Google Scholar] [CrossRef]
  40. Ren, X.; Malik, J. Learning a classification model for segmentation. In Proceedings of the Ninth IEEE International Conference on Computer Vision, Nice, France, 13–16 October 2003. [Google Scholar] [CrossRef] [Green Version]
  41. Stutz, D.; Hermans, A.; Leibe, B. Superpixels: An evaluation of the state-of-the-art. Comput. Vis. Image Underst. 2018, 166, 1–27. [Google Scholar] [CrossRef] [Green Version]
  42. Albayrak, A.; Bilgin, G. Automatic cell segmentation in histopathological images via two-staged superpixel-based algorithms. Med. Biol. Eng. Comput. 2018, 57, 653–665. [Google Scholar] [CrossRef]
  43. Vincent, L.; Soille, P. Watersheds in digital spaces: An efficient algorithm based on immersion simulations. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 583–598. [Google Scholar] [CrossRef] [Green Version]
  44. Cousty, J.; Bertrand, G.; Najman, L.; Couprie, M. Watershed Cuts: Thinnings, Shortest Path Forests, and Topological Watersheds. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 925–939. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Gamarra, M.; Zurek, E.; Escalante, H.J.; Hurtado, L.; San-Juan-Vergara, H. Split and merge watershed: A two-step method for cell segmentation in fluorescence microscopy images. Biomed. Signal Process. Control 2019, 53, 101575. [Google Scholar] [CrossRef] [PubMed]
  46. Zumbado-Corrales, M.; Esquivel-Rodríguez, J. EvoSeg: Automated Electron Microscopy Segmentation through Random Forests and Evolutionary Optimization. Biomimetics 2021, 6, 37. [Google Scholar] [CrossRef] [PubMed]
  47. Chan, T.; Vese, L. Active contours without edges. IEEE Trans. Image Process. 2001, 10, 266–277. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Ciecholewski, M.; Spodnik, J. Semi–Automatic Corpus Callosum Segmentation and 3D Visualization Using Active Contour Methods. Symmetry 2018, 10, 589. [Google Scholar] [CrossRef] [Green Version]
  49. Zhu, X.; Wei, Y.; Lu, Y.; Zhao, M.; Yang, K.; Wu, S.; Zhang, H.; Wong, K.K. Comparative analysis of active contour and convolutional neural network in rapid left-ventricle volume quantification using echocardiographic imaging. Comput. Methods Programs Biomed. 2021, 199, 105914. [Google Scholar] [CrossRef]
  50. Song, T.H.; Sanchez, V.; EIDaly, H.; Rajpoot, N.M. Dual-Channel Active Contour Model for Megakaryocytic Cell Segmentation in Bone Marrow Trephine Histology Images. IEEE Trans. Biomed. Eng. 2017, 64, 2913–2923. [Google Scholar] [CrossRef] [Green Version]
  51. Arafat, Y.; Reyes-Aldasoro, C.C. Computational Image Analysis Techniques, Programming Languages and Software Platforms Used in Cancer Research: A Scoping Review. In Proceedings of the Medical Image Understanding and Analysis; Yang, G., Aviles-Rivero, A., Roberts, M., Schönlieb, C.B., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2022; Volume 13413, pp. 833–847. [Google Scholar]
  52. Michael, E.; Ma, H.; Li, H.; Kulwa, F.; Li, J. Breast Cancer Segmentation Methods: Current Status and Future Potentials. BioMed Res. Int. 2021, 2021, 9962109. [Google Scholar] [CrossRef]
  53. Vicar, T.; Balvan, J.; Jaros, J.; Jug, F.; Kolar, R.; Masarik, M.; Gumulec, J. Cell segmentation methods for label-free contrast microscopy: Review and comprehensive comparison. BMC Bioinform. 2019, 20, 360. [Google Scholar] [CrossRef]
  54. Jones, M.L.; Spiers, H. The crowd storms the ivory tower. Nat. Methods 2018, 15, 579–580. [Google Scholar] [CrossRef]
  55. Tripathi, S.; Chandra, S.; Agrawal, A.; Tyagi, A.; Rehg, J.M.; Chari, V. Learning to Generate Synthetic Data via Compositing. In Proceedings of the CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 461–470. [Google Scholar]
  56. Such, F.P.; Rawal, A.; Lehman, J.; Stanley, K.; Clune, J. Generative Teaching Networks: Accelerating Neural Architecture Search by Learning to Generate Synthetic Training Data. In Proceedings of the 37th International Conference on Machine Learning, Virtual Event, 13–18 July 2020; PMLR: Brookline, MA, USA, 2020; pp. 9206–9216. [Google Scholar]
  57. Chen, X.; Mersch, B.; Nunes, L.; Marcuzzi, R.; Vizzo, I.; Behley, J.; Stachniss, C. Automatic Labeling to Generate Training Data for Online LiDAR-Based Moving Object Segmentation. IEEE Robot. Autom. Lett. 2022, 7, 6107–6114. [Google Scholar] [CrossRef]
  58. Schmitz, S.; Weinmann, M.; Weidner, U.; Hammer, H.; Thiele, A. Automatic generation of training data for land use and land cover classification by fusing heterogeneous data sets. Publ. Der Dtsch. Ges. Für Photogramm. Fernerkund. Und Geoinf. 2020, 29, 73–86. [Google Scholar]
  59. Voelsen, M.; Torres, D.L.; Feitosa, R.Q.; Rottensteiner, F.; Heipke, C. Investigations on Feature Similarity and the Impact of Training Data for Land Cover Classification. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2021, 3, 181–189. [Google Scholar] [CrossRef]
  60. Wen, T.; Tong, B.; Liu, Y.; Pan, T.; Du, Y.; Chen, Y.; Zhang, S. Review of research on the instance segmentation of cell images. Comput. Methods Programs Biomed. 2022, 227, 107211. [Google Scholar] [CrossRef] [PubMed]
  61. Zhang, L.; Jia, Z.; Leng, X.; Ma, F. Artificial Intelligence Algorithm-Based Ultrasound Image Segmentation Technology in the Diagnosis of Breast Cancer Axillary Lymph Node Metastasis. J. Healthc. Eng. 2021, 2021, 8830260. [Google Scholar] [CrossRef]
  62. Yang, M.; Zhang, Y.; Chen, H.; Wang, W.; Ni, H.; Chen, X.; Li, Z.; Mao, C. AX-Unet: A Deep Learning Framework for Image Segmentation to Assist Pancreatic Tumor Diagnosis. Front. Oncol. 2022, 12, 894970. [Google Scholar] [CrossRef] [PubMed]
  63. Tsochatzidis, L.; Koutla, P.; Costaridou, L.; Pratikakis, I. Integrating segmentation information into CNN for breast cancer diagnosis of mammographic masses. Comput. Methods Programs Biomed. 2021, 200, 105913. [Google Scholar] [CrossRef]
  64. Tahir, H.B.; Washington, S.; Yasmin, S.; King, M.; Haque, M.M. Influence of segmentation approaches on the before-after evaluation of engineering treatments: A hypothetical treatment approach. Accid. Anal. Prev. 2022, 176, 106795. [Google Scholar] [CrossRef] [PubMed]
  65. Ostroff, L.; Zeng, H. Electron Microscopy at Scale. Cell 2015, 162, 474–475. [Google Scholar] [CrossRef] [Green Version]
  66. Peddie, C.; Collinson, L. Exploring the third dimension: Volume electron microscopy comes of age. Micron 2014, 61, 9–19. [Google Scholar] [CrossRef]
  67. Tsai, W.T.; Hassan, A.; Sarkar, P.; Correa, J.; Metlagel, Z.; Jorgens, D.M.; Auer, M. From voxels to knowledge: A practical guide to the segmentation of complex electron microscopy 3D-data. J. Vis. Exp. 2014, 90, e51673. [Google Scholar] [CrossRef] [Green Version]
  68. Russell, M.R.; Lerner, T.R.; Burden, J.J.; Nkwe, D.O.; Pelchen-Matthews, A.; Domart, M.C.; Durgan, J.; Weston, A.; Jones, M.L.; Peddie, C.J.; et al. 3D correlative light and electron microscopy of cultured cells using serial blockface scanning electron microscopy. J. Cell Sci. 2017, 130, 278–291. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  69. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  70. Spiers, H.; Songhurst, H.; Nightingale, L.; de Folter, J.; Zooniverse Volunteer Community, Z.; Hutchings, R.; Peddie, C.J.; Weston, A.; Strange, A.; Hindmarsh, S.; et al. Deep learning for automatic segmentation of the nuclear envelope in electron microscopy data, trained with volunteer segmentations. Traffic 2021, 22, 240–253. [Google Scholar] [CrossRef]
  71. Karabag, C.; Jones, M.L.; Peddie, C.J.; Weston, A.E.; Collinson, L.M.; Reyes-Aldasoro, C.C. Automated Segmentation of HeLa Nuclear Envelope from Electron Microscopy Images. In Proceedings of the Medical Image Understanding and Analysis, Southampton, UK, 9–11 July 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 241–250. [Google Scholar] [CrossRef] [Green Version]
  72. Karabağ, C.; Jones, M.L.; Peddie, C.J.; Weston, A.E.; Collinson, L.M.; Reyes-Aldasoro, C.C. Segmentation and Modelling of the Nuclear Envelope of HeLa Cells Imaged with Serial Block Face Scanning Electron Microscopy. J. Imaging 2019, 5, 75. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  73. Karabağ, C.; Jones, M.L.; Peddie, C.J.; Weston, A.E.; Collinson, L.M.; Reyes-Aldasoro, C.C. Semantic segmentation of HeLa cells: An objective comparison between one traditional algorithm and four deep-learning architectures. PLoS ONE 2020, 15, e0230605. [Google Scholar] [CrossRef] [PubMed]
  74. Jones, M.; Songhurst, H.; Peddie, C.; Weston, A.; Spiers, H.; Lintott, C.; Collinson, L.M. Harnessing the Power of the Crowd for Bioimage Analysis. Microsc. Microanal. 2019, 25, 1372–1373. [Google Scholar] [CrossRef] [Green Version]
  75. Karabağ, C.; Jones, M.L.; Reyes-Aldasoro, C.C. Volumetric Semantic Instance Segmentation of the Plasma Membrane of HeLa Cells. J. Imaging 2021, 7, 93. [Google Scholar] [CrossRef]
  76. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  77. Deerinck, T.J.; Bushong, E.A.; Thor, A.; Ellisman, M.H. NCMIR: A New Protocol for Preparation of Biological Specimens for Serial Block-Face SEM Microscopy. 2010. Available online: https://ncmir.ucsd.edu/sbem-protocol (accessed on 23 February 2023).
  78. Iudin, A.; Korir, P.K.; Salavert-Torres, J.; Kleywegt, G.J.; Patwardhan, A. EMPIAR: A public archive for raw electron microscopy image data. Nat. Methods 2016, 13, 387–388. [Google Scholar] [CrossRef]
  79. Canny, J. A Computational Approach to Edge Detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, PAMI-8, 679–698. [Google Scholar] [CrossRef]
  80. Zhang, R.; Huang, L.; Xia, W.; Zhang, B.; Qiu, B.; Gao, X. Multiple supervised residual network for osteosarcoma segmentation in CT images. Comput. Med. Imaging Graph. 2018, 63, 1–8. [Google Scholar] [CrossRef] [PubMed]
  81. Zhou, T.; Dong, Y.; Lu, H.; Zheng, X.; Qiu, S.; Hou, S. APU-Net: An Attention Mechanism Parallel U-Net for Lung Tumor Segmentation. Biomed Res. Int. 2022, 2022, 5303651. [Google Scholar] [CrossRef] [PubMed]
  82. Zhao, H.; Niu, J.; Meng, H.; Wang, Y.; Li, Q.; Yu, Z. Focal U-Net: A Focal Self-attention based U-Net for Breast Lesion Segmentation in Ultrasound Images. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, UK, 11–15 July 2022; pp. 1506–1511. [Google Scholar]
  83. Mutaguchi, J.; Morooka, K.; Kobayashi, S.; Umehara, A.; Miyauchi, S.; Kinoshita, F.; Inokuchi, J.; Oda, Y.; Kurazume, R.; Eto, M. Artificial Intelligence for Segmentation of Bladder Tumor Cystoscopic Images Performed by U-Net with Dilated Convolution. J. Endourol. 2022, 36, 827–834. [Google Scholar] [CrossRef] [PubMed]
  84. Zhang, Y.; Chen, J.H.; Chang, K.T.; Park, V.Y.; Kim, M.J.; Chan, S.; Chang, P.; Chow, D.; Luk, A.; Kwong, T.; et al. Automatic Breast and Fibroglandular Tissue Segmentation in Breast MRI Using Deep Learning by a Fully-Convolutional Residual Neural Network U-Net. Acad. Radiol. 2019, 26, 1526–1535. [Google Scholar] [CrossRef] [PubMed]
  85. Zhao, B.; Chen, X.; Li, Z.; Yu, Z.; Yao, S.; Yan, L.; Wang, Y.; Liu, Z.; Liang, C.; Han, C. Triple U-net: Hematoxylin-aware nuclei segmentation with progressive dense feature aggregation. Med. Image Anal. 2020, 65, 101786. [Google Scholar] [CrossRef]
  86. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  87. Jaccard, P. Étude comparative de la distribution florale dans une portion des Alpes et des Jura. Bull. Del La Société Vaudoise Des Sci. Nat. 1901, 37, 547–579. [Google Scholar]
  88. Futrega, M.; Milesi, A.; Marcinkiewicz, M.; Ribalta, P. Optimized U-Net for Brain Tumor Segmentation. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Springer International Publishing: Berlin/Heidelberg, Germany, 2022; pp. 15–29. [Google Scholar] [CrossRef]
  89. Karabağ, C.; Verhoeven, J.; Miller, N.R.; Reyes-Aldasoro, C.C. Texture Segmentation: An Objective Comparison between Five Traditional Algorithms and a Deep-Learning U-Net Architecture. Appl. Sci. 2019, 9, 3900. [Google Scholar] [CrossRef] [Green Version]
  90. Jaffari, R.; Hashmani, M.A.; Reyes-Aldasoro, C.C. A Novel Focal Phi Loss for Power Line Segmentation with Auxiliary Classifier U-Net. Sensors 2021, 21, 2803. [Google Scholar] [CrossRef]
  91. Astono, I.P.; Welsh, J.S.; Chalup, S.; Greer, P. Optimisation of 2D U-Net Model Components for Automatic Prostate Segmentation on MRI. Appl. Sci. 2020, 10, 2601. [Google Scholar] [CrossRef] [Green Version]
  92. Li, S.; Xu, J.; Chen, R. U-Net Neural Network Optimization Method Based on Deconvolution Algorithm. In Neural Information Processing; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 592–602. [Google Scholar] [CrossRef]
  93. Kirichev, M.M.; Slavov, T.S.; Momcheva, G.D. Fuzzy U-Net Neural Network Architecture Optimization for Image Segmentation. IOP Conf. Ser. Mater. Sci. Eng. 2021, 1031, 012077. [Google Scholar] [CrossRef]
  94. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A Simple Framework for Contrastive Learning of Visual Representations. arXiv 2020, arXiv:2002.05709. [Google Scholar]
  95. Castro, E.; Cardoso, J.S.; Pereira, J.C. Elastic deformations for data augmentation in breast cancer mass detection. In Proceedings of the 2018 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Las Vegas, NV, USA, 4–7 March 2018. [Google Scholar] [CrossRef]
  96. Chlap, P.; Min, H.; Vandenberg, N.; Dowling, J.; Holloway, L.; Haworth, A. A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 2021, 65, 545–563. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Graphical abstract of the framework analysed in this paper. The input data consist of 517 electron microscope images with dimensions 8192 × 8192 pixels. A region of interest (ROI) (blue box) of 300 images each 2000 × 2000 was cropped and manually labeled into 4 classes (nucleus, nuclear envelope, cell, and background) to generate a ground truth. Pairs of patches of data and labels were generated and used to train a U-Net architecture with different strategies: 36,000 pairs, 135,000 pairs from a single cell, 135,000 from all the dataset extracted automatically with the image processing algorithm, and finally 270,000 with 135,000 from the ROI and 135,000 from other cells. The segmentations for the ROI were evaluated quantitatively, and the accuracy and Jaccard index were calculated. For a larger subsection of the data (red box), the evaluation was qualitative due to lack of ground truth.
Figure 1. Graphical abstract of the framework analysed in this paper. The input data consist of 517 electron microscope images with dimensions 8192 × 8192 pixels. A region of interest (ROI) (blue box) of 300 images each 2000 × 2000 was cropped and manually labeled into 4 classes (nucleus, nuclear envelope, cell, and background) to generate a ground truth. Pairs of patches of data and labels were generated and used to train a U-Net architecture with different strategies: 36,000 pairs, 135,000 pairs from a single cell, 135,000 from all the dataset extracted automatically with the image processing algorithm, and finally 270,000 with 135,000 from the ROI and 135,000 from other cells. The segmentations for the ROI were evaluated quantitatively, and the accuracy and Jaccard index were calculated. For a larger subsection of the data (red box), the evaluation was qualitative due to lack of ground truth.
Jimaging 09 00059 g001
Figure 2. One representative slice of one of the 8192 × 8192 × 517 images. The image shows several HeLa cells which are darker than the background. It should be noticed that the background is not completely uniform in intensity but it is smooth, and changes of intensity are not as abrupt as those for the cells. The main visible characteristic of the cells is a large central region that corresponds to the nucleus, surrounded by a dark thin region corresponding to the nuclear envelope. Cells with different characteristics are visible: for some, the nucleus is relatively smooth and regular (dashed blue box; this will be referred to in the manuscript as the region of interest (ROI)) as opposed to other nuclei that present invaginations and more complex geometries. Other structures, such as mitochondria are visible, but these are outside the scope of this work. A solid red box shows a larger region of interest, which will be used for qualitative comparisons. Scale bar at the bottom right indicates 5 μ m.
Figure 2. One representative slice of one of the 8192 × 8192 × 517 images. The image shows several HeLa cells which are darker than the background. It should be noticed that the background is not completely uniform in intensity but it is smooth, and changes of intensity are not as abrupt as those for the cells. The main visible characteristic of the cells is a large central region that corresponds to the nucleus, surrounded by a dark thin region corresponding to the nuclear envelope. Cells with different characteristics are visible: for some, the nucleus is relatively smooth and regular (dashed blue box; this will be referred to in the manuscript as the region of interest (ROI)) as opposed to other nuclei that present invaginations and more complex geometries. Other structures, such as mitochondria are visible, but these are outside the scope of this work. A solid red box shows a larger region of interest, which will be used for qualitative comparisons. Scale bar at the bottom right indicates 5 μ m.
Jimaging 09 00059 g002
Figure 3. (ae): Five representative slices of the 2000 × 2000 × 300 region of interest (ROI) with the central slice (c) corresponding to the blue box in Figure 2. It should be noticed how the nucleus decreases in size towards top (d,e) and bottom (a,b) of the cell. Also, the nuclei of the cells that surround the central cell are particularly visible towards the bottom of the cell (a) at the left side and top of the cell (e) at the right and bottom of the image. Scale bar at the bottom right indicates 5 μ m.
Figure 3. (ae): Five representative slices of the 2000 × 2000 × 300 region of interest (ROI) with the central slice (c) corresponding to the blue box in Figure 2. It should be noticed how the nucleus decreases in size towards top (d,e) and bottom (a,b) of the cell. Also, the nuclei of the cells that surround the central cell are particularly visible towards the bottom of the cell (a) at the left side and top of the cell (e) at the right and bottom of the image. Scale bar at the bottom right indicates 5 μ m.
Jimaging 09 00059 g003
Figure 4. Graphical illustration of the image processing algorithm. The image is low-passed filtered to reduce noise, edges are generated and then the regions that are not covered by edges become superpixels. Morphological operations are used to discard all regions except the one that corresponds to the nucleus, which in this case is the large central one. Finally, closing operations are used to fill the small gaps and provide a smooth nuclear segmentation.
Figure 4. Graphical illustration of the image processing algorithm. The image is low-passed filtered to reduce noise, edges are generated and then the regions that are not covered by edges become superpixels. Morphological operations are used to discard all regions except the one that corresponds to the nucleus, which in this case is the large central one. Finally, closing operations are used to fill the small gaps and provide a smooth nuclear segmentation.
Jimaging 09 00059 g004
Figure 5. U-Net architecture used for semantic segmentation in this work. Each level consists of a series of operations: convolutions (abbreviated as Conv), non-linear operations (rectified linear unit or ReLU) and subsampling (MaxPool). The data are first reduced in resolution (going down in the diagram) and then expanded (going up) to return to the same level of the input image.
Figure 5. U-Net architecture used for semantic segmentation in this work. Each level consists of a series of operations: convolutions (abbreviated as Conv), non-linear operations (rectified linear unit or ReLU) and subsampling (MaxPool). The data are first reduced in resolution (going down in the diagram) and then expanded (going up) to return to the same level of the input image.
Jimaging 09 00059 g005
Figure 6. Illustration of different approaches to generate training data. The first approach shown in the top row required the manual segmentation of a 2000 × 2000 × 300 region of interest. The region was manually cropped, and then each slice was manually segmented. The image and the ground truth were used to generate pairs of patches of data and labels used in strategies 1, 2 and 3. The second approach illustrated in the bottom row applied an image processing algorithm without manual intervention. A series of cells were detected in an image. Then, in a sequential manner, each of these was automatically cropped and segmented, and the image and ground truth were used to generate pairs of patches of data and labels used in strategy 4. The number of pairs is described in Section 2.5.
Figure 6. Illustration of different approaches to generate training data. The first approach shown in the top row required the manual segmentation of a 2000 × 2000 × 300 region of interest. The region was manually cropped, and then each slice was manually segmented. The image and the ground truth were used to generate pairs of patches of data and labels used in strategies 1, 2 and 3. The second approach illustrated in the bottom row applied an image processing algorithm without manual intervention. A series of cells were detected in an image. Then, in a sequential manner, each of these was automatically cropped and segmented, and the image and ground truth were used to generate pairs of patches of data and labels used in strategy 4. The number of pairs is described in Section 2.5.
Jimaging 09 00059 g006
Figure 7. Illustration of the post-processing. The classes corresponding to a small section of one slice of the dataset are shown on the left. On the two central columns, the pre-processing and post-processing are shown per class, from top to bottom, nucleus, cell and nuclear envelope. On the right is the final post-processed region. It can be seen how small misclassifications over the nucleus and cell have been smoothed, and the nuclear envelope improved its continuity.
Figure 7. Illustration of the post-processing. The classes corresponding to a small section of one slice of the dataset are shown on the left. On the two central columns, the pre-processing and post-processing are shown per class, from top to bottom, nucleus, cell and nuclear envelope. On the right is the final post-processed region. It can be seen how small misclassifications over the nucleus and cell have been smoothed, and the nuclear envelope improved its continuity.
Jimaging 09 00059 g007
Figure 8. Illustration of the semantic segmentation of one slice (71/300) of the ROI with the image processing algorithm and U-Net (36,000 patches). Top row (ad) show slice with results overlaid, bottom row (eh) shows ground truth (GT) and results as classes. (a,e) GT, central nucleus in green/white whilst other nuclei in purple/grey and background in greyscale/black. (b,f) IP segmentation. (c,g) U-Net segmentation. (d,h) U-Net segmentation after post-processing. In (fh) true positives = light grey, true negatives = darker grey, false positives = black, false negatives = white.
Figure 8. Illustration of the semantic segmentation of one slice (71/300) of the ROI with the image processing algorithm and U-Net (36,000 patches). Top row (ad) show slice with results overlaid, bottom row (eh) shows ground truth (GT) and results as classes. (a,e) GT, central nucleus in green/white whilst other nuclei in purple/grey and background in greyscale/black. (b,f) IP segmentation. (c,g) U-Net segmentation. (d,h) U-Net segmentation after post-processing. In (fh) true positives = light grey, true negatives = darker grey, false positives = black, false negatives = white.
Jimaging 09 00059 g008
Figure 9. Illustration of the semantic segmentation of one slice (71/300) of the region of interest (ROI) with the image processing (IP) algorithm and U-Net trained with 135,000 patches. For a description of (ah), see Figure 8.
Figure 9. Illustration of the semantic segmentation of one slice (71/300) of the region of interest (ROI) with the image processing (IP) algorithm and U-Net trained with 135,000 patches. For a description of (ah), see Figure 8.
Jimaging 09 00059 g009
Figure 10. Illustration of the semantic segmentation of one slice (251/300) of the region of interest (ROI) with the image processing (IP) algorithm and U-Net trained with 135,000 patches. For a description, see Figure 8.
Figure 10. Illustration of the semantic segmentation of one slice (251/300) of the region of interest (ROI) with the image processing (IP) algorithm and U-Net trained with 135,000 patches. For a description, see Figure 8.
Jimaging 09 00059 g010
Figure 11. Numerical comparison of the four algorithms (U-Net single nucleus 36,000 patches, U-Net multiple nuclei 36,000 patches, U-Net multiple nuclei, 135,000 patches, image processing (IP) algorithm). (a) Accuracy. Average values for all slices (0.9346, 0.9895, 0.9922, and 0.9926), for slices 150:200 (0.9966, 0.9974, 0.9971, and 0.9945). (b) Jaccard similarity index (JI). Average values for all slices (0.5138, 0.9158, 0.9378, and 0.6436), for slices 150:200 (0.9712, 0.9778, 0.9760, and 0.9564).
Figure 11. Numerical comparison of the four algorithms (U-Net single nucleus 36,000 patches, U-Net multiple nuclei 36,000 patches, U-Net multiple nuclei, 135,000 patches, image processing (IP) algorithm). (a) Accuracy. Average values for all slices (0.9346, 0.9895, 0.9922, and 0.9926), for slices 150:200 (0.9966, 0.9974, 0.9971, and 0.9945). (b) Jaccard similarity index (JI). Average values for all slices (0.5138, 0.9158, 0.9378, and 0.6436), for slices 150:200 (0.9712, 0.9778, 0.9760, and 0.9564).
Jimaging 09 00059 g011
Figure 12. Segmentation of the section of interest shown in Figure 2 (red box). (a) U-Net trained with 36,000 patches from one cell (bottom left). (b) U-Net trained with 135,000 patches from one cell. (c) U-Net trained with 135,000 patches automatically generated with the image processing algorithm. (d) U-Net trained with 270,000 patches; the combination of the 135,000 of previous approaches. (e) Image processing (IP) algorithm. Cells are shown with red shading, nuclei are shown with a green shade, and for the U-Net results, the nuclear envelope is shown with a blue shade. It can be noticed that the segmentation of the single cell from which patches were extracted (bottom left, solid arrows) is good in all cases, with slight errors in (c), which is natural, as the training for this strategy originated from other cells. On the other hand, the segmentation of the top right cells (dashed arrows) is much better in (c) than (a) or (b), which shows the improvement provided by training in more than one cell. Whilst the image processing algorithm provides very good results, it proceeds iteratively, and unlike the U-Net, it will not process all the cells as those in the edges (dashed-dot arrow in the bottom right), and in some cases, it may miss a section of the nucleus (dashed arrow in the top left).
Figure 12. Segmentation of the section of interest shown in Figure 2 (red box). (a) U-Net trained with 36,000 patches from one cell (bottom left). (b) U-Net trained with 135,000 patches from one cell. (c) U-Net trained with 135,000 patches automatically generated with the image processing algorithm. (d) U-Net trained with 270,000 patches; the combination of the 135,000 of previous approaches. (e) Image processing (IP) algorithm. Cells are shown with red shading, nuclei are shown with a green shade, and for the U-Net results, the nuclear envelope is shown with a blue shade. It can be noticed that the segmentation of the single cell from which patches were extracted (bottom left, solid arrows) is good in all cases, with slight errors in (c), which is natural, as the training for this strategy originated from other cells. On the other hand, the segmentation of the top right cells (dashed arrows) is much better in (c) than (a) or (b), which shows the improvement provided by training in more than one cell. Whilst the image processing algorithm provides very good results, it proceeds iteratively, and unlike the U-Net, it will not process all the cells as those in the edges (dashed-dot arrow in the bottom right), and in some cases, it may miss a section of the nucleus (dashed arrow in the top left).
Jimaging 09 00059 g012
Table 1. Numerical results of the accuracy and Jaccard similarity index obtained in one cell for which ground truth was available. It is important to highlight that “single nucleus” and “multiple nuclei” refer to the evaluation, not the training.
Table 1. Numerical results of the accuracy and Jaccard similarity index obtained in one cell for which ground truth was available. It is important to highlight that “single nucleus” and “multiple nuclei” refer to the evaluation, not the training.
U-Net
Single Nucleus
36,000
Strategy 1
U-Net
Multiple Nuclei
36,000
Strategy 2
U-Net
Multiple Nuclei
135,000
Strategy 3
Image
Processing
Algorithm
Accuracy 1:3000.93460.98950.99220.9926
Accuracy 150:2000.99660.99740.99710.9945
Jaccard 1:3000.51380.91580.93780.6436
Jaccard 150:2000.97120.97780.97600.9564
Jaccard 60:1500.80470.95790.95920.9565
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Karabağ, C.; Ortega-Ruíz, M.A.; Reyes-Aldasoro, C.C. Impact of Training Data, Ground Truth and Shape Variability in the Deep Learning-Based Semantic Segmentation of HeLa Cells Observed with Electron Microscopy. J. Imaging 2023, 9, 59. https://doi.org/10.3390/jimaging9030059

AMA Style

Karabağ C, Ortega-Ruíz MA, Reyes-Aldasoro CC. Impact of Training Data, Ground Truth and Shape Variability in the Deep Learning-Based Semantic Segmentation of HeLa Cells Observed with Electron Microscopy. Journal of Imaging. 2023; 9(3):59. https://doi.org/10.3390/jimaging9030059

Chicago/Turabian Style

Karabağ, Cefa, Mauricio Alberto Ortega-Ruíz, and Constantino Carlos Reyes-Aldasoro. 2023. "Impact of Training Data, Ground Truth and Shape Variability in the Deep Learning-Based Semantic Segmentation of HeLa Cells Observed with Electron Microscopy" Journal of Imaging 9, no. 3: 59. https://doi.org/10.3390/jimaging9030059

APA Style

Karabağ, C., Ortega-Ruíz, M. A., & Reyes-Aldasoro, C. C. (2023). Impact of Training Data, Ground Truth and Shape Variability in the Deep Learning-Based Semantic Segmentation of HeLa Cells Observed with Electron Microscopy. Journal of Imaging, 9(3), 59. https://doi.org/10.3390/jimaging9030059

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop