Next Article in Journal
Digital Incoherent Compressive Holography Using a Geometric Phase Metalens
Previous Article in Journal
Extracting Key Traffic Parameters from UAV Video with On-Board Vehicle Data Validation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Caenorhabditis elegans Multi-Tracker Based on a Modified Skeleton Algorithm

by
Pablo E. Layana Castro
,
Joan Carles Puchalt
,
Antonio García Garví
and
Antonio-José Sánchez-Salmerón
*
Instituto de Automática e Informática Industrial, Universitat Politècnica de València, 46022 Valencia, Spain
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5622; https://doi.org/10.3390/s21165622
Submission received: 13 July 2021 / Revised: 13 August 2021 / Accepted: 18 August 2021 / Published: 20 August 2021
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Automatic tracking of Caenorhabditis elegans (C. egans) in standard Petri dishes is challenging due to high-resolution image requirements when fully monitoring a Petri dish, but mainly due to potential losses of individual worm identity caused by aggregation of worms, overlaps and body contact. To date, trackers only automate tests for individual worm behaviors, canceling data when body contact occurs. However, essays automating contact behaviors still require solutions to this problem. In this work, we propose a solution to this difficulty using computer vision techniques. On the one hand, a skeletonization method is applied to extract skeletons in overlap and contact situations. On the other hand, new optimization methods are proposed to solve the identity problem during these situations. Experiments were performed with 70 tracks and 3779 poses (skeletons) of C. elegans. Several cost functions with different criteria have been evaluated, and the best results gave an accuracy of 99.42% in overlapping with other worms and noise on the plate using the modified skeleton algorithm and 98.73% precision using the classical skeleton algorithm.

1. Introduction

The nematode Caenorhabditis elegans (C. elegans) is a widely studied animal model, as its diverse age-related behavioral patterns provide valuable information on the function of its nervous system and is, therefore, an attractive model to evaluate the effects of mutations [1]. This facilitates the study and treatment of aging, as well as age-related pathologies and neurodegenerative disorders in humans at advanced ages [2,3]. Many of these studies have shown that automatic tracking applications based on computer vision systems help to reduce the manual cost of data acquisition and research hours, improving potential observation of the effects of drug trials [4] and improvements in lifespan or “shelf-life” [5,6,7]. These systems provide quantitative information on alterations in the individual motility and behavior of worms produced by chemical substances in their environment (chemotaxis), providing statistical data that allow further research in the field of health and wellness.
C. elegans demonstrate group behavior [8,9,10], among the best known are courtship, mating, aggression, rearing and foraging. Group behavior assays are currently being performed, for example, research into the effect of 0 2 in food search analysis, and aggregation [11,12]. These assays, like others, are visualized manually, due to the complexity of solving the identification problem during an overlapping or body contact of these worms. Currently, the automatic [13,14,15,16,17,18,19,20,21] or semi-automatic applications discard the data from tracks where there are these particular cases (overlapping and bodies contacts) [22,23,24]. Overlapping can take place among worms or may also be due to plate noise. Plate noise was defined as segmentation errors due to edges, or opaque waste in the plate.
Certain applications use diverse techniques and methods such as length [25], smoothness [26,27], previous segmentation [28], or other complex methods [29,30,31,32,33,34] to solve tracking problems during aggregation. The aforementioned methods were tested by our team and results indicated that previous segmentation is one of the most significant criteria to help identify worms within an aggregation. This was taken as a starting point to design different trackers, adding and combining numerous criteria, and it was found that the criteria of completeness and color used with the improvement of the form of skeletonizing [35] can help to identity the worms in the next image. Furthermore, we introduced three new criteria: length, smoothness, and noise, revealing that they can be useful in re-identification.
We present different multi-trackers which, unlike any other trackers, are fully automatic (Table 1). Segmentation and identification of the edge, the interior rim of the Petri dish (area of interest), and segmentation tracks made by worms are performed without the intervention of human operators. These new methods show worm-tracking accuracy of above 98% in nematode aggregations and problems related to plate noise. The best result reaches 99.42% accuracy.

2. Materials and Methods

2.1. Nematode Strains and Culture

The strains (N2) and CB1370 daf-2 (e1370) used in the study were provided by the University of Minnesota Caenorhabditis Genetics Center. The C. elegans were kept at 20 °C and cultured on 55 mm diameter NGM plates with 1 µg/mL of fungizone to prevent fungal contamination [38]. Escherichia coli (E. coli) strain OP50 was used as standard prey for C. elegans in the laboratory. FUdR (0.2 mM) was used to prevent replication. The cultured plates were 10, 15, 30, 60 and 100 adult-young worms synchronized from worm eggs incubated at 20 °C. This variety in the number of worms per plate allowed to obtain aggregations of two C. elegans or more.

2.2. Proposed Tracking Method

At present, automatic or semi-automatic trackers discard the data of tracks where there are overlap or body contacts, due to the difficulty of solving the identity of each individual in these situations. Our method does solve these situations using a fully automated pipeline based on an intelligent active backlight, a modified skeletonization method and new criteria to solve the final optimization problem.
Steps of the proposed tracking method are described in Figure 1a–f, and more detailed in Section 2.3, Section 2.4, Section 2.5, Section 2.6 and Section 2.7, respectively. This method starts with the acquisition of sequences of C. elegans images, Figure 1a. These image sequences are then processed (Figure 1b) to obtain mathematical models of C. elegans Figure 1c, and possible solutions to these in an aggregation. Finally, the result (Figure 1f) is the optimal minimization cost between models and possible solutions.

2.3. Image Acquisition

The image acquisition process was performed using the capture system [37]. To use this system, first, a laboratory operator removed the plates from the incubator, then the plates were analyzed to find condensation on the covers, if so, they were removed, otherwise, the image sequence was captured. The system [37] is automatic and captures an image every second. Escherichia coli (E. coli) strain OP50 was placed in the center of the plate to capture the worms inside the Petri dish and not scaling the edges or near them. The intelligent active backlighting method was used as the illumination technique [36]. This method is more robust than standard backlighting methods, as it allows us to obtain constant intensity values for the bottom of the Petri dish and the worms (greater than 48 and less than 35, respectively). This facilitated automatic segmentation with fixed thresholds on all images.
The image sequence was acquired at a resolution of 1944 × 1944 pixels and a frequency of 1 Hz (1 image per second) using the system [37] (Table 2).
This image acquisition system is open hardware and its assembly procedure, parts and description are described in detail in another work [37]. Worm tracking, using these image acquisition conditions, is a very challenging problem. An image resolution of 1944 × 1944 pixels is the lowest able to detect worms, when a complete Petri dish (55 mm. of diameter) is monitored with a fixed camera. In addition, this problem was solved by using a low frame rate of 1 Hz. The dataset collected was composed by sequences of 30 images where contact events between worms occurred, to perform all the experiments.

2.4. Image Processing

Image processing began with the segmentation of the region of interest and C. elegans tracks in the image sequence (white circle and red track in Figure 1b). To obtain the region of interest, first, a segmentation was carried out on all the images of the sequence using a threshold with a fixed intensity value (35 in the gray scale). Then, an AND operation was executed between all the segmented images, the result obtained went through a “Fillhole” operation to fill small holes. The region of interest was selected as the largest connected component of the resulting image. It is important to mention that the use of a fixed threshold for all images is due to the intelligent active backlighting method as a lighting technique [36], as this system allows to conserve background intensity values constantly.
In parallel to the previous step, the C. elegans tracks were segmented with a process using different threshold levels Figure 1b. Threshold levels were below 35 on the gray scale. The resulting segmentations went through two filters in order to eliminate those tracks that did not correspond to worms. In the first stage, those tracks with an area smaller than the minimum area of a worm were filtered. The second filter analyzed the skeleton of each image, if the skeleton did not correspond to a minimum expected length, it was classified as noise. The results of the number of skeletons found in each image were stored in a 30 × N matrix, where N is total number of tracks and 30 is the number of images in the sequence.
Due to low resolution of the worms, a scale factor of 3 was applied to increase resolution. This process was applied by obtaining the model to the end of tracking. The model of each worm was obtained by analyzing 30 × N matrix, finding image “k” of 30 images, where worms were further apart from each other, and their ends, head–tail, were also separate. Tracking of each worm started from image k + 1 to image 30 and from image k − 1 to the first. The skeletonization method proposed in previous work [35] was used in each image. This method used distance transformation [39] to obtain possible worm skeletons, and through of an optimization method using different criteria found the best skeleton prediction. Once the tracking process had finished, the results were reconverted to the original scale to be saved.

2.5. Worm Model

At present, there are different skeletonization methods [39,40]. Matlab’s bwmorph function was used as a classic skeletonization method to obtain worm skeleton model (color pixels in Figure 2b). The proposed worm model consists of width and color values along the skeleton of each individual, Figure 2c. The width values are obtained using classical skeleton in resulting image after using distance transformation function on segmentation of the image “k”. Grayscale image is shown in Figure 2a, while color values of pixels are obtained using the classical skeleton in gray image in Figure 2b. The length value is the total number of pixels in the skeleton. The length model is averaged while the C. elegans are separate and tracking progresses.

2.6. Extraction of Possible Solutions

The skeletonization method proposed in the previous work [35], unlike classical methods, enables the separation of aggregated worms (Figure 3a), creating new paths in the skeleton, and some possible solutions for each worm (Figure 3b–e). Maximum and minimum values of the width vector are used in the distance to transform images of each segmentation to find this new skeleton as mentioned in [35]. The possible solution skeletons are obtained from a recursive function, which runs through ends and branch points of the new skeleton that overlap the previous segmentation of the worm’s body (red circle in Figure 3b–e).

2.7. Optimization Method

The prediction “S” of the following postures will be the possible solution with the minimum value within all the possible “P” combinations of skeletons in one segmentation (1). The value of each possible combination “ C P ” of skeletons is obtained from the sum of criteria “m”, for the number of worms “n” in an aggregation (2). The criterion “ C p ” is evaluated for each possible worm “i” and for each criterion “j”. The criteria analyzed were the length of skeleton, overlap with the previous body, completeness, smoothness of the skeleton, noise in segmentation and the colors of each worm.
S =   arg min p ( C p )
C p = j = 1 m i = 1 n C j i
The length and color criteria prevent the prediction differing from the model in length and color. The smoothness criterion prevents sudden changes in the direction of the skeletons. The completeness criterion prevents the current segmentation from being incomplete. The overlap criterion with the previous body prevents the identity change during aggregation. Furthermore, the noise criterion prevents the skeleton prediction falling on the plate noise.
The reconstruction of the body of each worm was used for the evaluation of the different criteria. This was performed by using the skeleton pixels in each possible prediction with the width and color values obtained in the model (prediction start), respectively, for each individual.

2.7.1. Length Criterion

The length criterion “ C L ” is obtained from the sum of the multiplication of average squared width [ W i 2 ¯ ] with the difference in length ( Δ i ), (3). This difference is obtained from subtraction between the model length of each worm ( L i ) and length of the skeleton obtained ( W L i ). The average squared of the width was used so that the resulting length criterion is as significant as the rest of the criteria where the error in pixels was evaluated. Figure 4a shows the aggregation of two worms in a gray scale, and Figure 4b,c shows possible posture predictions where length errors occur.
C L = i = 1 n Δ i W i 2 ¯

2.7.2. Overlap Criterion

The overlap criterion “ C O ” is obtained from the sum of the absolute difference of the reconstruction of the worm’s body in the previous state, B P p x , (white dashed line in Figure 5b,c,e,f) and the current state, B C p x , (green and blue segmentation in Figure 5b,c,e,f) for each possible worm “i” in the aggregation (4) and (5). This is done for all “m” pixels in each reconstruction. Figure 5a shows in gray scale the aggregation of two worms, Figure 5d,g shows the next postures predictions (blue and green segmentation), parallel to these in dashed lines showing the previous state ( B P p x ) and the rest show the current state ( B C p x ). In Figure 5d it can be seen that the overlap criterion is low because it belongs to the same worms, while in Figure 5g it is higher due to the identity change.
C O i = p x = 1 m 1 i f B P p x B C p x = = 1 0
C O = i = 1 n C O i

2.7.3. Completeness Criterion

The completeness criterion “ C C p ” is obtained from the sum of the absolute difference of the current segmentation of the image ( A S ) and the reconstruction of each possible worm body “i” in the current state ( B C p x ) (6) and (7). This is done for all “m” pixels in each reconstruction or current segmentation. Figure 6a shows the aggregation of two worms in gray scale, Figure 6b–d shows completeness values for each image. Figure 6b is the correct prediction, while Figure 6c shows incorrect prediction, due to the identity change and final extremes. Figure 6d shows low completeness criterion and an incorrect prediction too, due to the change of identity in both worms.
C C p i = p x = 1 m 1 i f A S p x B C p x = = 1 0
C C p = i = 1 n C C p i

2.7.4. Smoothness Criterion

The smoothness criterion “ C S ” is obtained from the average of the absolute values of the angles obtained for the “nk” pixels of the skeleton and for each worm “i” in the segmentation (9). For each pixel of the skeleton, there is an angle ( θ p x ), which is obtained by an average of the sum of “nA” angles before and “nA” angles after divided by the total pixels found “c” (8). The value of “c” will be 2*nA if there are “nA” pixels before and after the pixel to be evaluated (8). Figure 7a shows the aggregation of two worms in gray scale, Figure 7b shows a possible prediction of skeletons with low softness criterion, while Figure 7c,d show possible predictions of skeletons with high softness criterion.
θ p x = a = x n A x + n A θ a c
C S = i = 1 n p x = 1 n k | θ p x | n k

2.7.5. Noise Criterion

The noise criterion “ C N ” is obtained from the intersection of the noise segmentation ( N S ) and the reconstruction of each body of a possible worm “i” in the current state ( B C p x ) (10) and (11). This is done for all “m” pixels in each reconstruction or noise segmentation. Figure 8a shows the aggregation of one worm with noise in a gray scale, while Figure 8b shows the correct prediction (segmentation in blue) with a low noise criterion (segmentation in magenta). Figure 8c shows an incorrect prediction (blue-magenta segmentation) with a high noise index (magenta segmentation).
C N i = p x = 1 m 1 i f N S p x & B C p x = = 1 0
C N = i = 1 n C N i

2.7.6. Color Criterion

The color criterion “ C C l ” is obtained by adding all the pixels with an absolute difference greater than the threshold value (U = 1) of the color model ( C M ) with respect to the reconstruction of the current body ( C A ) (12) for each worm in the aggregation (13). This means that the error only increases the value by 1 if the absolute difference between a pixel of the reconstruction of the body of the possible worm “i” using the color model ( C M p x ) and the reconstruction of the same worm “i” in the current segmentation ( C A p x ) is greater than the threshold value. This is done for all “m” pixels in the reconstruction. For both worm body reconstructions (model and current prediction), the same width and length values are used, in order to compare pixels 1 to 1. Figure 9a shows the aggregation of two worms in gray scale, while Figure 9b,c shows the comparison of each prediction with its respective model.
C C l i = p x = 1 m 1 i f C M p x C A p x > U 0
C C l = i = 1 n C C l i

2.8. Evaluation Method

Manually labeled skeletons were used as a reference to compare all the results. These skeletons were obtained using an application designed to select each pixel belonging to the skeleton of each worm one by one in the image sequence. This operation was performed for all 3779 postures of the 70 plates used. The shape of these nematodes was recovered using a disk-shaped dilation operation of radius equal to half the width (approx. 2 pixels) on the skeletons obtained.
The Jaccard coefficient, or intersection over the union (IoU), was used to measure the degree of precision in locating worms (14). As its name indicates, it is obtained by dividing the total area of the intersection by the union of the elements [41]. For the evaluation, the area of the reconstructed bodies of the manually labeled skeletons was used, skeletons using the skeletonization method proposed in [35] and the classical skeletonization method using the Matlab bwmorph command.
I o U = P w 1 P w 2 P w 1 P w 2
The IoU index was expected to be higher because a predicted pose (Figure 10b) is compared to an annotated ground-true pose (Figure 10a), which must overlap (Figure 10c–e). The results for the example below are IoU = 0.9784, 0.5667, and 0.2649, respectively.
Matlab 2018b Machine Learning Toolbox was used to obtain the comparison statistics between prediction models (Appendix A Figure A1a,b, Figure A2a,b, Figure A3a,b and Figure A4a,b) and the two skeletonization methods (Appendix A Figure A5a,b and Figure A6a,b). The Kolmogorov–Smirnov normality test was used for large samples (n > 50) and the Wilcoxon Signed Ranks test.

3. Results

Experiments were performed with 70 plates. Of these, 54 corresponded to plates with 10 and 15 worms, one plate with 30 worms, four plates with 60 worms, and 11 plates with 100 worms, totaling 65,400 worm poses. All these data were analyzed to obtain contact between worms and noise. As demonstrated in [5,14], a higher number of worms per plate will increase the likelihood of contact between them. Nematodes studied were young-adult wild-type (N2) and CB1370, daf-2 (e1370), as mentioned above.
53 tracks with 3240 poses were used to evaluate aggregation between worms, and 17 tracks with 509 poses for aggregation between worms and noise. The IoU index was used to evaluate the percentage of success in tracking the worms and also to compare both skeletonization methods. The area of worm bodies reconstructed from skeletons obtained manually and skeletons obtained with the two skeletonization methods (new and classical) was used to evaluate the IoU index.
Different prediction models were implemented in order to find the most significant criteria. The name of each model has been coded using letters from the criteria names. “O” for overlap, “L” for length, “Cp” for completeness, “N” for noise, “S” for smoothness, and “Cl” for color. The model with the best results was model 7 (OCpCl) with a 99.42% percentage accuracy in aggregated worm tracks and an IoU value of 0.70 in average. Figure 11a–e, shows an example using the model 7, in this image you can see the evaluation of the three criteria of this model and optimization result. Some examples of aggregation cases are presented at the end of the Appendix B using model 7 (Figure A7, Figure A8, Figure A9, Figure A10 and Figure A11).
To measure the percentage accuracy in tracking the worms, non-zero IoU values were used, from the beginning to the end of the tracks. The accuracy of the results of the different prediction models using the 2 skeletonization methods are shown in Table 3. In addition, the average IoU value was obtained for all the prediction models (see Table 4), from the beginning of the aggregation to the end. Then, 790 poses were used in aggregations between worms and 509 poses for worms aggregated with noise, 1299 poses in total.

Comparison with Other Trackers

At present, automatic or semi-automatic trackers discard the data of tracks where there are overlap or body contacts, due to the difficulty of solving the identity of each individual in these situations. Our method does solve these situations, so a direct comparison cannot be made with the results of other trackers.
To compare our method graphically with other trackers (tierpsy-tracker [22], WF-NTP.v3 [23]), labeled data was first shaded in different grays and overlapped predicted data in colors (Figure 12a–c). Errors for each comparison are shown in grayscale. The comparison was made with bodies reconstructed from skeletons (skeleton dilation with disk equal to 2), except for WF-NTP.v3 [23], whose results are skeleton centroid points. The tierpsy-tracker [22] multi-tracker, Figure 12a, did not resolve path 2 where aggregation occurs, while the other tracks (1, 3, 4) were partially resolved, due to the occlusions that the worms make on themselves. The multi-tacker WF-NTP.v3 [23], Figure 12b, did not solve the identity problem in track 2. Furthermore, like the previous multi-tracker it presented problems in the tracks with occlusions (1, 3, 4). Model 7, on the other hand, had almost zero errors, tracked all worms, resolved the identity in the aggregation of track 2 and the occlusions in the remaining tracks, Figure 12c.

4. Discussion

Caenorhabditis elegans are aggregated in different ways, such as aggregation of end parts (head or tail), partial aggregation of bodies, and aggregation of parallel bodies, among others. The experiments were conducted with the above mentioned methods with a single criterion and found that the overlapping criterion with the previous prediction is the most significant. This criterion allows part of the previous state to be preserved, helping to solve the next state. A result of 98.49% was obtained using this criterion individually. However, when aggregation gives rise to an overlap in most worms, it is difficult to identify them, even for human observers. To solve this problem, different tests were designed, combining the criteria mentioned in the optimization methods section.
The completeness in cases of aggregation, where there are overlaps between aggregated bodies, partial and total identity changes could be obtained, as observed in the optimization method Figure 6c,d. When using the overlap criterion with completeness criterion, a percentage accuracy of 99.13% was achieved. As shown in Appendix A (Figure A1a,b and Figure A2a,b) this criterion is statistically significant, helping to solve some cases where the overlap criterion presented problems, Figure 5b.
The color criterion also helped to improve worm prediction, although not so significant (Appendix A Figure A3a,b and Figure A4a,b). The end portion of C. elegan (tail) has a greater number of light pixels (higher gray levels) than the head, where there are fewer. These features are important because before, during and after an aggregation one of the final parts remains visible, and with this color singularity a more accurate prediction can be obtained, helping to solve the identity problem after an aggregation between worms or noise. The use of these three criteria (overlap, completeness, and color) allowed us to obtain a percentage accuracy of 99.42%.
The length, smoothness, and noise criteria were discarded because when they were used individually or together with the overlap criterion, their percentage accuracy decreased. On the one hand, the problem with the smoothness criterion was due to the low resolution of the worms in our images, which provided a poor estimate of this criterion. On the other hand, the problem with the length criterion was due to errors in the length model and the increase or decrease in the length of worms owing to overlaps. The problem with the noise criterion was due to, when the worm is visualized against background noise, many possible solutions were generated.
Finally, it is worth mentioning that the best combination of criteria depends on image quality. In this work, we demonstrated that the combination of the three criteria mentioned above (overlap, completeness, and color) was the best option for automatic tracking of interaction behaviors among C. elegans (contacts or overlapping) with our low-resolution dataset.

5. Conclusions

This paper presents a method for tracking multiple C. elegans in standard Petri dishes where some worms can come into contact or overlap. This method was evaluated in a difficult scenario using a low-image resolution and a low frame rate. Using an optimizer with the appropriate criteria (overlap, completeness, and color) was shown to solve many worm overlap and contact situations. The accuracy obtained under these conditions was 99.42% and 98.73% using the modified skeleton algorithm and the classical skeleton algorithm, respectively. In addition, the proposed method employs an improved active backlight system and an improved skeletonization algorithm. The active backlight system provides fixed gray levels in all captured images, which allows automatic segmentation using a fixed threshold. The improved skeletonization algorithm uses width information from each worm model to extract skeletons, enabling the tracking of worms moving in parallel (side by side). Our proposal, unlike other trackers that discard worm overlaps and contacts, solves many of these situations, increasing the number of worms tracked in a Petri dish and therefore paving the way to automate new assays where interaction between worms occurs.

Author Contributions

Conceptualization, P.E.L.C. and A.-J.S.-S.; methodology, P.E.L.C. and A.-J.S.-S.; software, P.E.L.C.; validation, P.E.L.C., J.C.P., A.-J.S.-S. and A.G.G.; formal analysis, P.E.L.C. and A.-J.S.-S.; investigation, P.E.L.C. and A.-J.S.-S.; resources, A.-J.S.-S.; data curation, P.E.L.C., J.C.P. and A.G.G.; writing—original draft preparation, P.E.L.C. and A.-J.S.-S.; writing—review and editing, P.E.L.C., J.C.P., A.G.G. and A.-J.S.-S.; visualization, P.E.L.C. and A.-J.S.-S.; supervision, A.-J.S.-S.; project administration, A.-J.S.-S.; funding acquisition, A.-J.S.-S. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the Plan Nacional de I+D with Project RTI2018-094312-B-I00, FPI Predoctoral contract PRE2019-088214 and by European FEDER funds.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The program was developed in Matlab2018b with Windows 10 and works correctly in later versions with Image Processing, Communications and Bioinformatics toolbox. Its source code is in GitHub. It is open-source MIT (Massachusetts Institute of Technology) and can be downloaded from the repository at https://github.com/playanaC/WT_ISA, accessed on 17 August 2021. The dataset with all aggregation experiments can be downloaded from https://active-vision.ai2.upv.es/wp-content/uploads/2021/02/dataset_skeletons.zip, accessed on 17 August 2021.

Acknowledgments

ADM Nutrition, Biopolis S.L. and Archer Daniels Midland supplied the C. elegans plates. Some strains were provided by the CGC, which is funded by NIH Office of Research Infrastructure Programs (P40 OD010440). Maria-Gabriela Salazar-Secada developed the skeleton annotation application. Jordi Tortosa-Grau annotated worm skeletons.

Conflicts of Interest

No conflict of interest exists.

Appendix A

Appendix A.1. IoU Comparison in Models 3 and 1

Figure A1. Box plot and normality test of the difference of both models. (a) Box plot, green line indicates the mean in both graphs, and gray line indicates the median. Model1, N = 1299, mean = 0.6358, median = 0.7027, std. deviation = 0.2289, variance = 0.0540. Model3, N = 1299, mean = 0.6933, median = 0.7517, std. deviation = 0.1940, variance = 0.0376. (b) Normality test on the difference of methods (Model 3–Model 1). The p-value obtained was 8.24E-96 less than the significance value of 0.05, so the null hypothesis was rejected and the alternative hypothesis H1 was accepted (data did not come from normal distribution). Once the alternative hypothesis was accepted, Wilcoxon signed ranks test was used to evaluate both methods.
Figure A1. Box plot and normality test of the difference of both models. (a) Box plot, green line indicates the mean in both graphs, and gray line indicates the median. Model1, N = 1299, mean = 0.6358, median = 0.7027, std. deviation = 0.2289, variance = 0.0540. Model3, N = 1299, mean = 0.6933, median = 0.7517, std. deviation = 0.1940, variance = 0.0376. (b) Normality test on the difference of methods (Model 3–Model 1). The p-value obtained was 8.24E-96 less than the significance value of 0.05, so the null hypothesis was rejected and the alternative hypothesis H1 was accepted (data did not come from normal distribution). Once the alternative hypothesis was accepted, Wilcoxon signed ranks test was used to evaluate both methods.
Sensors 21 05622 g0a1
Figure A2. Wilcoxon signed rank test. (a) The Wilcoxon signed rank test table shows the difference that exists in 2 related samples through positive, negative and tie ranges. (b) p-value obtained with Wilcoxon rank test was 741E-34 less than the significance value of 0.05, so it was concluded that there was a statistically significant difference between both models.
Figure A2. Wilcoxon signed rank test. (a) The Wilcoxon signed rank test table shows the difference that exists in 2 related samples through positive, negative and tie ranges. (b) p-value obtained with Wilcoxon rank test was 741E-34 less than the significance value of 0.05, so it was concluded that there was a statistically significant difference between both models.
Sensors 21 05622 g0a2

Appendix A.2. IoU Comparison in Models 6 and 1

Figure A3. Box plot and normality test of the difference of both models. (a) Box plot, green line indicates the mean in both graphs, and gray line indicates the median.Model1, N = 1299, mean = 0.6358, median = 0.7027, std. deviation = 0.2289, variance = 0.0540. Model6, N = 1299, mean = 0.6464, median = 0.7197, std. deviation = 0.2300, variance = 0.0529. (b) Normality test on the difference of methods (Model 6–Model 1). The p-value obtained was 1.48E-151 less than the significance value of 0.05, so the null hypothesis was rejected and the alternative hypothesis H1 was accepted (data did not come from normal distribution). Once the alternative hypothesis was accepted, Wilcoxon signed ranks test was used to evaluate both methods.
Figure A3. Box plot and normality test of the difference of both models. (a) Box plot, green line indicates the mean in both graphs, and gray line indicates the median.Model1, N = 1299, mean = 0.6358, median = 0.7027, std. deviation = 0.2289, variance = 0.0540. Model6, N = 1299, mean = 0.6464, median = 0.7197, std. deviation = 0.2300, variance = 0.0529. (b) Normality test on the difference of methods (Model 6–Model 1). The p-value obtained was 1.48E-151 less than the significance value of 0.05, so the null hypothesis was rejected and the alternative hypothesis H1 was accepted (data did not come from normal distribution). Once the alternative hypothesis was accepted, Wilcoxon signed ranks test was used to evaluate both methods.
Sensors 21 05622 g0a3
Figure A4. Wilcoxon signed rank test. (a) The Wilcoxon signed rank test table shows the difference that exists in two related samples through positive, negative and tie ranges. (b) p-value obtained with Wilcoxon rank test was 0.1054 less than the significance value of 0.05, so it was concluded that there was not a statistically significant difference between both models.
Figure A4. Wilcoxon signed rank test. (a) The Wilcoxon signed rank test table shows the difference that exists in two related samples through positive, negative and tie ranges. (b) p-value obtained with Wilcoxon rank test was 0.1054 less than the significance value of 0.05, so it was concluded that there was not a statistically significant difference between both models.
Sensors 21 05622 g0a4

Appendix A.3. IoU Comparison New and Classical Method (Model 7)

Figure A5. Box plot and normality test of the difference of both methods. (a) Box plot, green line indicates the mean in both graphs, and gray line indicates the median. Classical method, N = 1299, mean = 0.6719, median = 0.7218, std. deviation = 0.1842, variance = 0.0339. New method, N = 1299, mean = 0.6975, median = 0.7562, std. deviation = 0.1929, variance = 0.0372. (b) Normality test on the difference of methods (New-classical). The p-value obtained was 1.26E-152 less than the significance value of 0.05, so the null hypothesis was rejected and the alternative hypothesis H1 was accepted (data did not come from normal distribution). Once the alternative hypothesis was accepted, Wilcoxon signed ranks test was used to evaluate both methods.
Figure A5. Box plot and normality test of the difference of both methods. (a) Box plot, green line indicates the mean in both graphs, and gray line indicates the median. Classical method, N = 1299, mean = 0.6719, median = 0.7218, std. deviation = 0.1842, variance = 0.0339. New method, N = 1299, mean = 0.6975, median = 0.7562, std. deviation = 0.1929, variance = 0.0372. (b) Normality test on the difference of methods (New-classical). The p-value obtained was 1.26E-152 less than the significance value of 0.05, so the null hypothesis was rejected and the alternative hypothesis H1 was accepted (data did not come from normal distribution). Once the alternative hypothesis was accepted, Wilcoxon signed ranks test was used to evaluate both methods.
Sensors 21 05622 g0a5
Figure A6. Wilcoxon signed rank test. (a) The Wilcoxon signed rank test table shows the difference that exists in two related samples through positive, negative and tie ranges. (b) p-value obtained with Wilcoxon rank test was 3.46E-22 less than the significance value of 0.05, so it was concluded that there was a statistically significant difference between both methods.
Figure A6. Wilcoxon signed rank test. (a) The Wilcoxon signed rank test table shows the difference that exists in two related samples through positive, negative and tie ranges. (b) p-value obtained with Wilcoxon rank test was 3.46E-22 less than the significance value of 0.05, so it was concluded that there was a statistically significant difference between both methods.
Sensors 21 05622 g0a6

Appendix B

Appendix B.1. Evaluation Examples with Model 7

Figure A7. Model 7 evaluation, example 1. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Overlap criterion evaluation. (c) Completeness criterion evaluation. (d) Color criterion evaluation. (e) Optimization result.
Figure A7. Model 7 evaluation, example 1. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Overlap criterion evaluation. (c) Completeness criterion evaluation. (d) Color criterion evaluation. (e) Optimization result.
Sensors 21 05622 g0a7
Figure A8. Model 7 evaluation, example 2. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Overlap criterion evaluation. (c) Completeness criterion evaluation. (d) Color criterion evaluation. (e) Optimization result.
Figure A8. Model 7 evaluation, example 2. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Overlap criterion evaluation. (c) Completeness criterion evaluation. (d) Color criterion evaluation. (e) Optimization result.
Sensors 21 05622 g0a8
Figure A9. Model 7 evaluation, example 3. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Overlap criterion evaluation. (c) Completeness criterion evaluation. (d) Color criterion evaluation. (e) Optimization result.
Figure A9. Model 7 evaluation, example 3. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Overlap criterion evaluation. (c) Completeness criterion evaluation. (d) Color criterion evaluation. (e) Optimization result.
Sensors 21 05622 g0a9

Appendix B.2. Failure Cases

Figure A10. Model 7 evaluation, example 4. Errors occurred by the presence of noise. (a) Grayscale image. (b) Resulting skeleton using the improved form of skeletonizing. (c) Optimization result.
Figure A10. Model 7 evaluation, example 4. Errors occurred by the presence of noise. (a) Grayscale image. (b) Resulting skeleton using the improved form of skeletonizing. (c) Optimization result.
Sensors 21 05622 g0a10
Figure A11. Model 7 evaluation, example 5. Errors occurred by the presence of noise. (a) Grayscale image. (b) Resulting skeleton using the improved form of skeletonizing (c) Optimization result.
Figure A11. Model 7 evaluation, example 5. Errors occurred by the presence of noise. (a) Grayscale image. (b) Resulting skeleton using the improved form of skeletonizing (c) Optimization result.
Sensors 21 05622 g0a11

References

  1. Olsen, A.; Gill, M.S. Ageing: Lessons from C. elegans; Springer International Publishing: Cham, Switzerland, 2017; pp. 2–5. [Google Scholar]
  2. Teo, E.; Lim, S.J.; Fong, S.; Larbi, A.; Wright, G.D.; Tolwinski, N.; Gruber, J. A high throughput drug screening paradigm using transgenic Caenorhabditis elegans model of Alzheimer’s disease. Transl. Med. Aging 2020, 4, 11–21. [Google Scholar] [CrossRef]
  3. Kim, M.; Knoefler, D.; Quarles, E.; Jakob, U.; Bazopoulou, D. Automated phenotyping and lifespan assessment of a C. elegans model of Parkinson’s disease. Transl. Med. Aging 2020, 4, 38–44. [Google Scholar] [CrossRef]
  4. Spensley, M.; Del Borrello, S.; Pajkic, D.; Fraser, A.G. Acute Effects of Drugs on Caenorhabditis elegans Movement Reveal Complex Responses and Plasticity. G3 Genes Genomes Genet. 2018, 8, 2941–2952. [Google Scholar] [CrossRef] [Green Version]
  5. Puchalt, J.C.; Sánchez-Salmerón, A.J.; Ivorra, E.; Genovés Martínez, S.; Martínez, R.; Martorell Guerola, P. Improving lifespan automation for Caenorhabditis elegans by using image processing and a post-processing adaptive data filter. Sci. Rep. 2020, 10, 8729. [Google Scholar] [CrossRef]
  6. Jung, S.K.; Aleman Meza, B.; Riepe, C.; Zhong, W. QuantWorm: A Comprehensive Software Package for Caenorhabditis elegans Phenotypic Assays. PLoS ONE 2014, 9, e84830. [Google Scholar] [CrossRef] [Green Version]
  7. Puchalt, J.C.; Layana Castro, P.E.; Sánchez-Salmerón, A.J. Reducing Results Variance in Lifespan Machines: An Analysis of the Influence of Vibrotaxis on Wild-Type Caenorhabditis elegans for the Death Criterion. Sensors 2020, 20, 5981. [Google Scholar] [CrossRef]
  8. Sokolowski, M.B. Social Interactions in “Simple” Model Systems. Neurone 2010, 65, 780–794. [Google Scholar] [CrossRef] [Green Version]
  9. De Bono, M.; Bargmann, C.I. Natural Variation in a Neuropeptide Y Receptor Homolog Modifies Social Behavior and Food Response in C. elegans. Cell 1998, 94, 679–689. [Google Scholar] [CrossRef] [Green Version]
  10. Srinivasan, J.; Von Reuss, S.H.; Bose, N.; Zaslaver, A.; Mahanti, P.; Margaret, M.C.; O’Doherty, O.G.; Edison, A.S.; Sternberg, P.W.; Schroeder, F.C. A Modular Library of Small Molecule Signals Regulates Social Behaviors in Caenorhabditis elegans. PLoS Biol. 2012, 10, e1001237. [Google Scholar] [CrossRef] [Green Version]
  11. Rogers, C.; Persson, A.; Cheung, B.; De Bono, M. Behavioral Motifs and Neural Pathways Coordinating O2 Responses and Aggregation in C. elegans. Curr. Biol. 2006, 16, 649–659. [Google Scholar] [CrossRef] [Green Version]
  12. Dunsenbery, D.B. Video camera-computer tracking of nematode Caenorhabditis elegans to record behavioral responses. J. Chem. Ecol. 1985, 11, 1239–1247. [Google Scholar] [CrossRef]
  13. Restif, C.; Ibáñez Ventoso, C.; Vora, M.M.; Guo, S.; Metaxas, D.; Driscoll, M. CeleST: Computer Vision Software for Quantitative Analysis of C. elegans Swim Behavior Reveals Novel Features of Locomotion. PLoS ONE 2014, 10, e1003702. [Google Scholar] [CrossRef] [Green Version]
  14. Swierczek, N.A.; Giles, A.C.; Rankin, C.H.; Kerr, R.A. High-throughput behavioral analysis in C. elegans. Nat. Methods 2011, 8, 592–598. [Google Scholar] [CrossRef] [Green Version]
  15. Ramot, D.; Johnson, B.E.; Berry, B., Jr.; Carnell, L.; Goodman, M.B. The Parallel Worm Tracker: A Platform for Measuring Average Speed and Drug-Induced Paralysis in Nematodes. PLoS ONE 2008, 3, e2208. [Google Scholar] [CrossRef] [Green Version]
  16. Dusenbery, D.B. Using a microcomputer and video camera to simultaneously track 25 animals. Comput. Biol. Med. 1985, 15, 169–175. [Google Scholar] [CrossRef]
  17. Simonetta, S.H.; Golombek, D.A. An automated tracking system for Caenorhabditis elegans locomotor behavior and circadian studies application. J. Neurosci. Methods 2007, 161, 273–280. [Google Scholar] [CrossRef] [PubMed]
  18. Jaensch, S.; Decker, M.; Hyman, A.A.; Myers, E.W. Automated tracking and analysis of centrosomes in early Caenorhabditis elegans embryos. Bioinformatics 2010, 26, i13–i20. [Google Scholar] [CrossRef] [Green Version]
  19. Boyd, W.; Anderson, G.; Dusenbery, D.; Williams, P. Computer Tracking Method for Assessing Behavioral Changes in the Nematode Caenorhabditis elegans. ASTM Int. 2000, 9, 225–238. [Google Scholar] [CrossRef]
  20. Dzyubachyk, O.; Jelier, R.; Lehner, B.; Niessen, W.; Meijering, E. Model-based approach for tracking embryogenesis in Caenorhabditis elegans fluorescence microscopy data. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009; pp. 5356–5359. [Google Scholar] [CrossRef]
  21. Lorimer, T.; Goodridge, R.; Bock, A.K.; Agarwal, V.; Saberski, E.; Sugihara, G.; Rikin, S.A. Tracking changes in behavioural dynamics using prediction error. PLoS ONE 2021, 16, e0251053. [Google Scholar] [CrossRef]
  22. Javer, A.; Currie, M.; Lee, C.; Hokanson, J.; Li, K.; Martineau, C.N.; Yemini, E.; Grundy, L.; Li, C.; Ch’ng, Q.; et al. An open-source platform for analyzing and sharing worm-behavior data. Nat. Methods 2018, 15, 645–646. [Google Scholar] [CrossRef]
  23. Koopman, M.; Peter, Q.; Seinstra, R.I.; Perni, M.; Vendruscolo, M.; Dobson, C.; Knowles, T.; Nollen, E. Assessing motor-related phenotypes of Caenorhabditis elegans with the wide field-of-view nematode tracking platform. Nat. Protoc. 2018, 15, 2071–2106. [Google Scholar] [CrossRef]
  24. Leonard, N.; Vidal-Gadea, A.G. Affordable Caenorhabditis elegans tracking system for classroom use. MicroPubl. Biol. 2021. [Google Scholar] [CrossRef]
  25. Wählby, C.; Kamentsky, L.; Liu, Z.; Riklin Raviv, T.; Conery, A.; O’Rourke, E.; Sokolnicki, K.; Visvikis, O.; Ljosa, V.; Irazoqui, J.; et al. An image analysis toolbox for high-throughput C. elegans assays. Nat. Methods 2012, 9, 714–716. [Google Scholar] [CrossRef] [Green Version]
  26. Rizvandi, N.B.; Pižurica, A.; Rooms, F.; Philips, W. Skeleton analysis of population images for detection of isolated and overlapped nematode C. elegans. In Proceedings of the 2008 16th European Signal Processing Conference, Lausanne, Switzerland, 25–29 August 2008; pp. 1–5. [Google Scholar]
  27. Rizvandi, N.B.; Pižurica, A.; Philips, W. Machine vision detection of isolated and overlapped nematode worms using skeleton analysis. In Proceedings of the 2008 15th IEEE International Conference on Image Processing, San Diego, CA, USA, 12–15 October 2008; pp. 2972–2975. [Google Scholar] [CrossRef]
  28. Winter, P.B.; Brielmann, R.M.; Timkovich, N.P.; Navarro, H.T.; Teixeira Castro, A.; Morimoto, R.I.; Amaral, L.A.N. A network approach to discerning the identities of C. elegans in a free moving population. Sci. Rep. 2016, 10, 34859. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Huang, K.M.; Cosman, P.; Schafer, W.R. Machine vision based detection of omega bends and reversals in C. elegans. J. Neurosci. Methods 2006, 158, 323–336. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Fontaine, E.; Burdick, J.; Barr, A. Automated Tracking of Multiple C. elegans. In Proceedings of the 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, New York, NY, USA, 30 August–3 September 2006; pp. 3716–3719. [Google Scholar] [CrossRef]
  31. Roussel, N.; Morton, C.A.; Finger, F.P.; Roysam, B. A Computational Model for C. elegans Locomotory Behavior: Application to Multiworm Tracking. IEEE Trans. Biomed. Eng. 2007, 54, 1786–1797. [Google Scholar] [CrossRef] [PubMed]
  32. Nagy, S.; Goessling, M.; Amit, Y.; Biron, D. A Generative Statistical Algorithm for Automatic Detection of Complex Postures. PLoS Comput. Biol. 2015, 11, e1004517. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Uhlmann, V.; Unser, M. Tip-seeking active contours for bioimage segmentation. In Proceedings of the 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), Brooklyn, NY, USA, 16–19 April 2015; pp. 544–547. [Google Scholar] [CrossRef]
  34. Kiel, M.; Berh, D.; Daniel, J.; Otto, N.; Steege, A.T.; Jiang, X.; Liebau, E.; Risse, B. A Multi-Purpose Worm Tracker Based on FIM. bioRxiv 2018. [Google Scholar] [CrossRef]
  35. Layana Castro, P.E.; Puchalt, J.C.; Sánchez Salmerón, A.J. Improving skeleton algorithm for helping Caenorhabditis elegans trackers. Sci. Rep. 2020, 10, 22247. [Google Scholar] [CrossRef] [PubMed]
  36. Puchalt, J.C.; Sánchez-Salmerón, A.J.; Martorell Guerola, P.; Ivorra, E.; Genovés Martínez, S. Active backlight for automating visual monitoring: An analysis of a lighting control technique for Caenorhabditis elegans cultured on standard Petri plates. PLoS ONE 2019, 14, e0215548. [Google Scholar] [CrossRef]
  37. Puchalt, J.C.; Sánchez-Salmerón, A.J.; Ivorra, E. Small flexible automated system for monitoring Caenorhabditis elegans lifespan based on active vision and image processing techniques. Sci. Rep. 2021, 11, 12289. [Google Scholar] [CrossRef] [PubMed]
  38. Stiernagle, T. Maintenance of C. elegans. In C. elegans a Practical Approach; Oxford University Press: New York, NY, USA, 2006; pp. 2005–2018. [Google Scholar] [CrossRef] [Green Version]
  39. Russ, J.C.; Neal, F.B. The Image Processing Handbook. In The Image Processing Handbook, 7th ed.; CRC Press, Inc.: Boca Raton, FL, USA, 2015; pp. 479–492. [Google Scholar]
  40. Sossa, J.H. An improved parallel algorithm for thinning digital patterns. Pattern Recognit. Lett. 1989, 10, 77–80. [Google Scholar] [CrossRef]
  41. Koul, A.; Ganju, S.; Kasam, M. Building The Purrfect Cat Locator App with TensorFlow Object Detection API. In Practical Deep Learning for Cloud, Mobile, and Edge: Real-World AI & Computer-Vision Projects Using Python, Keras & TensorFlow; O’Reilly Media: Sebastopol, CA, USA, 2019; Chapter 14; pp. 679–680. [Google Scholar]
Figure 1. General scheme of image processing. The image shows the different stages that the tracker goes through to obtain the results of the skeletons and to follow all the worms within the plate. (a) Image acquisition [37]. (b) Pre-processing image. (c) Worm models. (d) Improved skeleton using proposed method [35]. (e) Optimization method. (f) Optimization results (predicted poses).
Figure 1. General scheme of image processing. The image shows the different stages that the tracker goes through to obtain the results of the skeletons and to follow all the worms within the plate. (a) Image acquisition [37]. (b) Pre-processing image. (c) Worm models. (d) Improved skeleton using proposed method [35]. (e) Optimization method. (f) Optimization results (predicted poses).
Sensors 21 05622 g001
Figure 2. Prediction model. Skeleton gray values were changed with a HOT color map for better visualization. (a) Grayscale image. (b) Color values obtained from grayscale image. (c) Values of widths marked with the colors of the model; the length is the total pixels in the skeleton.
Figure 2. Prediction model. Skeleton gray values were changed with a HOT color map for better visualization. (a) Grayscale image. (b) Color values obtained from grayscale image. (c) Values of widths marked with the colors of the model; the length is the total pixels in the skeleton.
Sensors 21 05622 g002
Figure 3. Possible solutions. Circles in red mark the starting point to find the possible skeleton. Cyan and green lines are the possible solutions for each worm. (a) Grayscale image. (be) Possible solutions.
Figure 3. Possible solutions. Circles in red mark the starting point to find the possible skeleton. Cyan and green lines are the possible solutions for each worm. (a) Grayscale image. (be) Possible solutions.
Sensors 21 05622 g003
Figure 4. Length criterion evaluation. The pixels in red are the length error. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Length criterion evaluation with length error in worm1 (blue). (c) Length criterion evaluation with length error in worm2 (green).
Figure 4. Length criterion evaluation. The pixels in red are the length error. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Length criterion evaluation with length error in worm1 (blue). (c) Length criterion evaluation with length error in worm2 (green).
Sensors 21 05622 g004
Figure 5. Overlap criterion evaluation. The pixels in red and magenta are the error of overlap with the previous state. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). The dashed line segmentation is the previous state segmentation. (a) Grayscale image. (b,e) Previous state ( B p s p x ) in white dashed line and current state ( B c s p x ) in green. (c,f) Previous state ( B p s p x in white dashed line and current state ( B c s p x ) in blue. (d) Evaluation with low overlap criterion. (g) Evaluation with high overlap criterion.
Figure 5. Overlap criterion evaluation. The pixels in red and magenta are the error of overlap with the previous state. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). The dashed line segmentation is the previous state segmentation. (a) Grayscale image. (b,e) Previous state ( B p s p x ) in white dashed line and current state ( B c s p x ) in green. (c,f) Previous state ( B p s p x in white dashed line and current state ( B c s p x ) in blue. (d) Evaluation with low overlap criterion. (g) Evaluation with high overlap criterion.
Sensors 21 05622 g005
Figure 6. Completeness criterion evaluation. The pixels in red are the completeness error. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Correct prediction with low completeness criterion. (c) Incorrect prediction with high completeness criterion. (d) Identities changed and with the same completeness criterion as image (b).
Figure 6. Completeness criterion evaluation. The pixels in red are the completeness error. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Correct prediction with low completeness criterion. (c) Incorrect prediction with high completeness criterion. (d) Identities changed and with the same completeness criterion as image (b).
Sensors 21 05622 g006
Figure 7. Smoothness criterion evaluation. The yellow and white pixels are the resulting skeleton using the improved skeletonizing method. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Evaluation of smoothness, angles in red are those that have more weight and increase the index of smoothness. (c) Incorrect prediction of worm1 with high softness criterion. (d) Incorrect prediction of worm2 with high softness criterion.
Figure 7. Smoothness criterion evaluation. The yellow and white pixels are the resulting skeleton using the improved skeletonizing method. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Evaluation of smoothness, angles in red are those that have more weight and increase the index of smoothness. (c) Incorrect prediction of worm1 with high softness criterion. (d) Incorrect prediction of worm2 with high softness criterion.
Sensors 21 05622 g007
Figure 8. Noise criterion evaluation. The pixels in magenta (intersection of blue and red) are the noise error. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the worm’s body (segmentation in blue). (a) Grayscale image. (b) Correct prediction with low noise criterion. (c) Incorrect prediction with high noise criterion.
Figure 8. Noise criterion evaluation. The pixels in magenta (intersection of blue and red) are the noise error. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the worm’s body (segmentation in blue). (a) Grayscale image. (b) Correct prediction with low noise criterion. (c) Incorrect prediction with high noise criterion.
Sensors 21 05622 g008
Figure 9. Color criterion evaluation. For each possible skeleton, its color values are obtained using the current grayscale image, and with those color values and the width values of the model, each worm is reconstructed and compared with the reconstruction of the model (model worm). The gray values were changed for a HOT color map in order to better visualize them. The blue lines indicate those pixels different from the model. (a) Grayscale image. (b) Comparison of worm1 model with current prediction of worm1. (c) Comparison of worm2 model with current prediction of worm2.
Figure 9. Color criterion evaluation. For each possible skeleton, its color values are obtained using the current grayscale image, and with those color values and the width values of the model, each worm is reconstructed and compared with the reconstruction of the model (model worm). The gray values were changed for a HOT color map in order to better visualize them. The blue lines indicate those pixels different from the model. (a) Grayscale image. (b) Comparison of worm1 model with current prediction of worm1. (c) Comparison of worm2 model with current prediction of worm2.
Sensors 21 05622 g009
Figure 10. IoU index. This index evaluates how close the response of the automatic method comes to the reference, and compares both results to determine the improvement of one method over the other. The higher the IoU value, the closer the response comes to the reference. The evaluation is performed by reconstructing the skeletons obtained with a radio 2 disk. (a) Reconstructed body of the manually labeled skeleton. (b) Reconstructed body of the skeleton obtained using the new skeletonization method [35] or the classical method. (ce) Evaluation of the reconstructed skeletons.
Figure 10. IoU index. This index evaluates how close the response of the automatic method comes to the reference, and compares both results to determine the improvement of one method over the other. The higher the IoU value, the closer the response comes to the reference. The evaluation is performed by reconstructing the skeletons obtained with a radio 2 disk. (a) Reconstructed body of the manually labeled skeleton. (b) Reconstructed body of the skeleton obtained using the new skeletonization method [35] or the classical method. (ce) Evaluation of the reconstructed skeletons.
Sensors 21 05622 g010
Figure 11. Model 7 evaluation. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Overlap criterion evaluation. (c) Completeness criterion evaluation. (d) Color criterion evaluation. (e) Optimization result.
Figure 11. Model 7 evaluation. The yellow and white pixels are the resulting skeleton using the improved form of skeletonizing. The white pixels are the pixels of the skeleton prediction, which are used to reconstruct the body of each worm (segmentation in blue and green). (a) Grayscale image. (b) Overlap criterion evaluation. (c) Completeness criterion evaluation. (d) Color criterion evaluation. (e) Optimization result.
Sensors 21 05622 g011
Figure 12. Comparison of trackers. Comparison of reconstruction of C. elegans bodies between labeled data (shaded in grays) and predictions obtained with different trackers (shaded in colors). (a) Results obtained using tierpsy-tracker [22]. (b) Labeled data (grays) compared with colored lines that connect centroids obtained using WF-NTP.v3 [23]. (c) Results obtained using model 7.
Figure 12. Comparison of trackers. Comparison of reconstruction of C. elegans bodies between labeled data (shaded in grays) and predictions obtained with different trackers (shaded in colors). (a) Results obtained using tierpsy-tracker [22]. (b) Labeled data (grays) compared with colored lines that connect centroids obtained using WF-NTP.v3 [23]. (c) Results obtained using model 7.
Sensors 21 05622 g012
Table 1. Comparison with other multi-trackers. This table shows the comparison of our multi-tracker with respect to others (tierpsy-tracker [22], WF-NTP.v3 [23]).
Table 1. Comparison with other multi-trackers. This table shows the comparison of our multi-tracker with respect to others (tierpsy-tracker [22], WF-NTP.v3 [23]).
Comparative Table
NameIllumination TechniqueFeaturesMethodSituations solved
Tierpsy Tracker [22]Standard backlightManual parameter settingSkeletons, outlines and segmentationsIndividual tracking
WF-NTP.v3 [23]Flat-field illuminationManual parameter settingSkeletons and centroidsIndividual tracking
OursActive backlight system [36,37]Fully automaticImproved skeleton and segmentationsIndividual tracking, overlaps, body contacts, rolled worms and occlusions.
Table 2. Components of the data acquisition system.
Table 2. Components of the data acquisition system.
System Components
N °NameDescription
1Raspberry Pi V3 b+Procesador: 64-bit ARM Cortex-A53, 1.4 GHz
RAM Size: 1 GB LPDDR2 SDRAM
2Raspberry Pi Camera V1.3Sensor: OmniVision OV5647
Pixel resolution: 2592 × 1944
Pixel size: 1.4 × 1.4 µm
Field of view: 53.50° × 41.41°
Optical size: 1/4′′
Focal length: 2.9
3Raspberry Pi displayScreen display size: 7′′
Resolution: 800 × 480 and 60 fps
Color: 24-bit RGB colour
Table 3. Comparative table of percentage accuracy of postures for models and methods. The table shows the percentage accuracy of poses (skeletons) for each model and method used during the tracking of C. elegans. Overall, 3240 poses were used to evaluate tracks where there is aggregation of two or more worms, and 509 poses to evaluate the aggregation between worms and noise on the plate.
Table 3. Comparative table of percentage accuracy of postures for models and methods. The table shows the percentage accuracy of poses (skeletons) for each model and method used during the tracking of C. elegans. Overall, 3240 poses were used to evaluate tracks where there is aggregation of two or more worms, and 509 poses to evaluate the aggregation between worms and noise on the plate.
Worms Aggregation
N ModelNewClassical
1O98.4997.80
2OL98.1597.99
3OCp99.1398.09
4ON98.0997.88
5OS98.0797.72
6OCl98.5798.20
7OCpCl99.4298.73
Table 4. Summary of model and method comparison. This table shows the results for each prediction model and two skeletonization methods, new [35] and classical (using the Matlab bwmorph function). The results column indicates the percentage value of the improvement using the new skeleton with respect to the classic skeletonization.
Table 4. Summary of model and method comparison. This table shows the results for each prediction model and two skeletonization methods, new [35] and classical (using the Matlab bwmorph function). The results column indicates the percentage value of the improvement using the new skeleton with respect to the classic skeletonization.
Average IoUStandard DeviationResults %
N ModelTotal PoseNewClassicalNewClassicalImprovement
1O12990.640.630.230.220.39
2OL12990.630.630.230.220.10
3OCp12990.690.670.190.192.61
4ON12990.630.630.240.220.50
5OS12990.630.630.230.220.04
6OCl12990.650.650.230.20−0.15
7OCpCl12990.700.670.190.182.66
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Layana Castro, P.E.; Puchalt, J.C.; García Garví, A.; Sánchez-Salmerón, A.-J. Caenorhabditis elegans Multi-Tracker Based on a Modified Skeleton Algorithm. Sensors 2021, 21, 5622. https://doi.org/10.3390/s21165622

AMA Style

Layana Castro PE, Puchalt JC, García Garví A, Sánchez-Salmerón A-J. Caenorhabditis elegans Multi-Tracker Based on a Modified Skeleton Algorithm. Sensors. 2021; 21(16):5622. https://doi.org/10.3390/s21165622

Chicago/Turabian Style

Layana Castro, Pablo E., Joan Carles Puchalt, Antonio García Garví, and Antonio-José Sánchez-Salmerón. 2021. "Caenorhabditis elegans Multi-Tracker Based on a Modified Skeleton Algorithm" Sensors 21, no. 16: 5622. https://doi.org/10.3390/s21165622

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop