Next Article in Journal
A Framework for Detecting and Analyzing Behavior Changes of Elderly People over Time Using Learning Techniques
Next Article in Special Issue
Siamese Architecture-Based 3D DenseNet with Person-Specific Normalization Using Neutral Expression for Spontaneous and Posed Smile Classification
Previous Article in Journal
Mode Analysis of Pt/LGS Surface Acoustic Wave Devices
Previous Article in Special Issue
Feature Analysis of Smart Shoe Sensors for Classification of Gait Patterns
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated White Blood Cell Counting in Nailfold Capillary Using Deep Learning Segmentation and Video Stabilization

1
Department of Computer Engineering, Kwangwoon University, Seoul 01897, Korea
2
School of Applied Science, Telkom University, Bandung 40257, Indonesia
3
Department of Electrical Communication, Daelim University, Anyang-Si 13916, Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2020, 20(24), 7101; https://doi.org/10.3390/s20247101
Submission received: 14 October 2020 / Revised: 30 November 2020 / Accepted: 9 December 2020 / Published: 11 December 2020

Abstract

:
White blood cells (WBCs) are essential components of the immune system in the human body. Various invasive and noninvasive methods to monitor the condition of the WBCs have been developed. Among them, a noninvasive method exploits an optical characteristic of WBCs in a nailfold capillary image, as they appear as visual gaps. This method is inexpensive and could possibly be implemented on a portable device. However, recent studies on this method use a manual or semimanual image segmentation, which depends on recognizable features and the intervention of experts, hindering its scalability and applicability. We address and solve this problem with proposing an automated method for detecting and counting WBCs that appear as visual gaps on nailfold capillary images. The proposed method consists of an automatic capillary segmentation method using deep learning, video stabilization, and WBC event detection algorithms. Performances of the three segmentation algorithms (manual, conventional, and deep learning) with/without video stabilization were benchmarks. Experimental results demonstrate that the proposed method improves the performance of the WBC event counting and outperforms conventional approaches.

1. Introduction

White blood cells (WBCs), also called leukocytes, are important components of the immune system in the human body, since their deficiency could cause various health conditions, such as sepsis [1], infectious diseases [2,3], and cancer [4]. The WBC monitoring usually requires the extraction of a blood sample by an experienced medical staff using specialized equipment. Patients who need continuous monitoring of the WBCs daily, for conditions such as neutropenia, have drawbacks due the equipment’s inefficiency and inconvenience. To improve the WBC monitoring method, various noninvasive techniques based on optical methods have been proposed. One method assesses the vessel in the oral mucosa [5], and another one is based on optical characteristics in nailfold capillaries, where WBCs are recognized as visual gaps or particles [6]. This study focuses on the nailfold capillary images because it could implement the WBC monitoring system while improving both patient comfort and measurement accessibility.
When the diameter of a WBC reaches that of the capillary through which it flows, the blood flow along that capillary is interrupted by the WBC, causing a gap in a capillary image. A sequence of images with gaps allows for seeing the WBCs flowing through the capillary, as illustrated in Figure 1, where each gap is considered as a WBC event. Using the visual gap characteristic, Bourquard et al. [7] proposed a semi-automated pipeline for gap numeration to achieve the noninvasive WBC analysis using a portable and low-cost capillaroscope and an image processing method, namely, spatiotemporal representation and the Radon transform. However, they conducted the capillary segmentation process in a semi-manual way. As manual processing is substantially slower than a computerized processing, it represents a bottleneck for automated analysis. To prevent this, an automated image segmentation approach based on a deep learning algorithm could be a solution, since the convolutional neural networks have fostered efficiency in computer vision tasks, including object recognition [8,9], detection [10,11], and segmentation [12,13,14,15]. For instance, the U-Net has been widely utilized for image segmentation [16,17] given its high performance and efficient use of GPU memory [18]. Likewise, we also adopt the U-Net [13] for our semantic segmentation problem aiming to automate capillary identification.
Additionally, inaccurate capillary detection and miss-detection of the WBC event might occur as a small finger motion is magnified when seen through a microscope. To mitigate this type of artifact, we apply motion compensation while also adopting an efficient version of image registration for translations [19] that reduces the capillary motion and accumulates data in case of a large unwanted movement in the recorded capillary videos [20]. The main contribution of this study is the automation of WBC event counting using a deep learning approach supported by video stabilization for robustness against motion artifacts. The proposed method consists of an automatic capillary segmentation method using deep learning, video stabilization, and WBC event detection algorithms.
The remainder of this paper is organized as follows: In Section 2, related works about semantic segmentation and video stabilization are presented; details of the proposed method are explained in Section 3; Section 4 provides the results of the experiment settings and results; Section 5 provides the discussion; lastly, Section 6 gives the summary and conclusions of this work.

2. Related Work

2.1. Semantic Segmentation

Image segmentation aims to extract useful regions from an image as a set of contours or sub-images for subsequent analysis and interpretation. To extract regions of interest, thresholding approaches [21,22] can be used to determine pixel intensities that discard the background. In addition, clustering approaches, such as the K-means [23], group similar pixel intensities over a region. However, these approaches are sensitive to the image characteristics and scenes, such as shaded images and the presence of multicolored objects, thus undermining their performance.
Since the introduction of convolutional neural networks [24], the consistency of visual recognition has substantially improved in many visual tasks, including classification [8,9,25,26], detection [10,11,27], and segmentation [12,28,29,30,31]. Such deep learning approaches leverage high-level features of input images, providing robustness against noise compared to conventional methods. Deep learning using convolutional neural networks has been exploited in various applications of medical image processing including segmentation [13,14,15].
We adopt u-net architecture for capillary segmentation because it uses global location and context information simultaneously, and works well with few training samples [13]. The u-net can be trained using the few data images in end-to-end manner, where the whole image in the forward pass can directly produce segmentation maps in order to retain the full context of the input images [13].

2.2. Video Stabilization

Camera motion while recording video causes the captured objects to move accordingly. When unintended, such motion can cause noise such as image blurring, which undermines the quality and consistency of the recorded image. To prevent this problem, stabilization improves the quality of video by eliminating unintended movements, including translations and rotations. Various approaches are available to remove motion artifacts from video. As the proposed method relies on video recorded by a human operator, the video is likely to include unwanted movements. Moreover, noise is amplified in the video because the capillaroscopy device magnifies the capillaries during recording. Therefore, video stabilization becomes essential for processing and analyzing capillary videos.
We only assume device translations during video recording [32]. Therefore, video stabilization in the proposed method should determine translation vectors. Given that these vectors contain the directions and magnitudes of camera movements, it is possible to fix the capillary positions by shifting the frames based on the vectors along the opposite direction. To determine the translation vector between two consecutive images, we adopt a computationally efficient version of an image registration method [19]. Specifically, given two images, Guizar-Sicairos et al. [19] determine the translation vector that maximizes the cross-correlation between them. To this end, each image is expressed in the frequency domain by applying the fast Fourier transform. Then, elementwise multiplication is performed between the converted images, and then the result is reverted back into the spatial domain. For image registration, capillary labels corresponding to frames are exploited rather than raw frames to mitigate the impact of noise.
If large camera movements occur, image registration may fail in the worst case. The magnification in capillary videos of small hand motions can produce such large camera movements, representing a potentially severe problem in the proposed method. A previous study [20] has addressed this problem by accumulating movements. Likewise, we adopt the same approach to reduce the extent of large movements.

3. Proposed WBC Counting Method

The proposed WBC counting method is illustrated in Figure 2, where the capillary video and predicted events are its input and output, respectively. During frame extraction, the input video is preprocessed to extract regions of interest containing capillaries for the subsequent analysis. Capillary segmentation extracts the capillaries and removes the background from the video. Then, coordinate determination selects the coordinates from the labels of each capillary to extract the corresponding pixel intensities. Video stabilization is applied using image registration based on the translation vectors acquired from the capillary labels. A spatiotemporal representation transforms the intensities acquired from the selected coordinates of each capillary in a frame into a 1D array, and the set of arrays obtained from all the frames are represented as a 2D matrix, whose x- and y-axes represent time and the corresponding intensity array, respectively. Finally, event detection predicts the events in the spatiotemporal map using the Radon transform and local maxima detection. Below, we detail each step of the proposed method.

3.1. Frame Extraction

In this step, all the frames are extracted from an input video. The capillaries in raw images are hardly distinguishable from the background (see Figure 3a), which results in poor capillary segmentation and event counting due to their low-quality information. To enhance the contrast for better visibility of capillaries in a video, we apply histogram equalization to the red, green, and blue channels in the frames (see Figure 3b).

3.2. Capillary Segmentation

To extract the representative information from the capillaries captured in a video, the capillary labels should be determined to characterize their locations and appearances. To this end, we adopt deep learning segmentation and compare its performance with a conventional segmentation method.

3.2.1. Deep Learning-Based Segmentation

For a deep learning-based segmentation of the capillaries captured in a video, we adopt the semantic segmentation model introduced in [13]. Given a frame image (see Figure 3b), the model outputs pixelwise capillary labels (see Figure 3c), which determine the locations and shapes of the capillaries. Specifically, an RGB image is fed to the model encoder and compressed into a dense representation as a multidimensional vector through consecutive convolutional and pooling layers. Then, the model decoder up-samples the compressed representation through consecutive convolutional and up-sampling layers.
Figure 4 describes the model architecture of the proposed deep learning segmentation. Each box represents a feature map with dimension [c, w, h], where c is the channel size (a value on top of the box) and w and h are the width and height (values on the side of the box) of the feature map, respectively. Each colored arrow denotes the corresponding operation (see the figure legends) between the connected feature maps. A gray arrow indicates concatenation, which is followed by a 2 × 2 up-convolution represented as a green arrow. An orange arrow represents a 2 × 2 max-pooling layer to reduce the spatial size followed by a 3 × 3 convolutional layer represented as a blue arrow. At the end of the model, the dark yellow arrow indicates a 1 × 1 convolution to map each multidimensional channel onto a scalar, such that the feature map is converted into a single-channel map, which contains the pixelwise labels of the capillaries in the input image.
We use a dataset containing 1358 capillary images and the corresponding ground-truth labels for training and validating the model. We divide this dataset into 950 and 408 images for the training and validation sets, respectively. After training the model for 100 epochs with Adam optimization on binary cross-entropy loss, learning rate of 0.001, and batch size of 3, 91.11% validation accuracy was obtained regarding the mean intersection over union.

3.2.2. Conventional Segmentation

As a benchmark test, a conventional segmentation algorithm is implemented based on the capillary optical characteristics. Since a capillary appears as a red region such as Figure 3b, the color information of the region could be utilized for the segmentation process. Therefore, pixels in a capillary area contain less green and blue components than those from the background. For the segmentation, capillaries are labelled by subtracting the sum of the green and blue components from the weighted red component in an image to achieve a relatively large contrast with respect to the background. In addition, the subtraction results are squared to highlight the intensities of the capillary and discard those of the background, whose intensity is low in general. The conventional segmentation algorithm for an images is implemented as follows:
L a b e l = R e L U ( λ I R ( I G + I B ) ) 2
R e L U ( x ) = m a x ( 0 , x )
where λ is the weight of the red component and I R , I G , and I B denote the intensities of the red, green, and blue components of an image. Weight λ (1.5 in this study) compensates for the intensity of the red component by the summation of the green and blue components.

3.3. Video Stabilization

To minimize unwanted motion artifacts in a capillary video stream, the translation vectors of video frames corresponding to a reference frame are determined [20]. A motion artifact is measured between two frames; current frame and reference frame. The video stabilization process estimates how much a frame has moved from the reference frame, where the amount and direction of the movement are represented as a translation vector. By adding the translation vectors to the frames, the capillary positions could be corrected to keep the initial position as the reference, which is usually the first frame [32]. Figure 5a–c illustrate an example of the video stabilization process. Considering the first frame in Figure 5a as the reference, the frame in Figure 5b is aligned to the position of the reference frame. The stabilized frame in Figure 5c shows the aligned capillaries to the reference locations.
However, it is empirically observed that the video stabilization process shows poor performance when the amount of the translation becomes large. As the amount of the translation has more chances to become large as time goes given a fixed reference frame, there would be more chances for the video stabilization process not to show the best performance if the first frame is consistently used as the reference frame. In the sense, a simple yet effective way to keep the amount of the translation small would be to periodically update the reference frame every p frames (p is set as 50), not to fix the reference frame to the first frame. It could alleviate the performance degradation issue because the periodic update shortens the time interval between frames to be stabilized and a reference frame and therefore the amount of the translation could be kept small.
When a reference frame is updated, subsequent frames are stabilized with respect to the next reference frame. However, the video stabilization process translates the positions of capillaries in the subsequent frames to the position of those in not the first frame but the next reference frame. To stabilize the subsequent frames with respect to the first frame, the video stabilization process also keeps a history of the translations among all reference frames. When the translation between the first and the last frames is equivalent to the sum of all translation vectors, a single variable, denoted by “reference” in Algorithm 1, would be enough to represent the history.
It is empirically observed that a capillary video is better stabilized with the translation vectors acquired from not raw frames but the labels obtained from the segmentation process detailed in Section 3.2. For the reason, the video stabilization process stabilizes all capillary videos based on the translation vectors from the labels.
In addition, the video stabilization algorithm is also applied to the labels. Even though the capillary segmentation could generate such a great quality of capillary masks, it would be much better to utilize multiple capillary masks as an ensemble for more robust masks. However, the multiple capillary masks include capillaries in different positions across time. To keep all capillaries of the masks in the same position, the same video stabilization process is applied to the masks.
Algorithm 1 describes the video stabilization process, where “ g e t _ t r a n s l a t i o n _ v e c ” returns a translation vector between a reference and an image, and “ a p p l y _ t r a n s l a t i o n ” translates the image based on the translation vector.
Algorithm 1: Pseudo code for video stabilization.
Sensors 20 07101 i001

3.4. Capillary Coordinate Determination

A label obtained from capillary segmentation includes multiple pixels, but they are not ordered or all of them are not needed for the analysis. Therefore, the important pixels and their orders should be precisely determined. Unlike the method in [7], which applies an interpolation [33] with manual selected information, the automatically selected pixels along the boundary of a capillary with a label are utilized to orderly extract the internal, external, and intermediate coordinates corresponding to the capillary.
To determine the coordinates, a virtual rectangle around the labelled capillary is adjusted to establish a region of interest, as illustrated in Figure 6a. The rectangle fully covers the short horizontal side between the two branches of a capillary, and all the coordinates in the internal boundary of each capillary are selected, as illustrated in Figure 6b. Then, an equal number of external coordinates are also chosen, and the external boundary is divided into as many coordinates as the internal coordinates like Figure 6c. This procedure enables the internal coordinates to be paired with the external ones.
Using the internal and external coordinates, the intermediate coordinates are created with the connecting lines across the paired internal and external coordinates, similar to the method in [7], as illustrated in Figure 6d. Using all the determined coordinates, a 1D array is constructed, whose length is the number of the internal coordinates and each element is the average intensity of the corresponding internal, external, and intermediate coordinates. Figure 6 illustrates all the processes of the capillary coordinate determination described in this section, and Figure 3c,d display the results example.

3.5. Spatiotemporal Representation

Given a frame, the coordinates of each capillary (Section 3.4) are used to obtain a capillary vector that represents the intensities of the capillary. By concatenating the vectors across all frames over time, we obtain a matrix that represents the intensities of a capillary throughout the video, establishing what we call an ST map. The x- and y-axes of an ST map indicate the time and capillary vector corresponding to time, respectively. An element of a capillary vector has a high value if it corresponds to the area of the event containing the WBC. As the element of an event moves along the capillary vector over time, the event appears as a line in the ST map. The problem of counting visual gaps is thus a problem of counting lines in an ST map. An example of events appearing in an ST map is shown in Figure 3e.
The raw ST map is sensitive to the quality of the capillary image. Therefore, the ST map should be processed to highlight the event information and eliminate ambient noise. First, we apply a median filter to the ST map to remove weak event information. Then, the unnecessary background is eliminated to highlight the event lines in the ST map. To this end, each row of the ST map is zero-meaned, and the resulting negative values are zeroed. To enhance the contrast between the background after the removal and lines in the ST map, the derivative of the intensities of the lines is calculated while applying the Sobel operator to the processed ST map. As a result, the lines in the ST map become more distinguishable, as illustrated in Figure 3f.

3.6. WBC Event Detection

To detect events occurring as lines in an ST map, we adopt a local maxima detection method using the Radon transform, which was introduced in [7]. The Radon transform maps an image given in Cartesian coordinates into an image given in polar coordinates which is called a polar map. The Radon transform is defined as
R ( θ , z ) = f ( s sin θ + z cos θ , s cos θ + z sin θ ) d s
where f ( i , j ) and R ( θ , z ) are a pixel of a zero-centered image at ( i , j ) in Cartesian coordinates and a polar map at ( θ , z ) in polar coordinates, respectively. The map converts lines in an image (ST map) into peaks, as illustrated in Figure 3g. The lines of a processed ST map are thus represented as peaks in the polar map. The peaks in a polar map are illustrated in Figure 3h. By detecting the local maxima in a polar map, the WBC events in a video sequence could be identified.
Given its thickness, a line in an ST map might lead to duplicate detections in the corresponding polar map, as the line is not an ideal one with zero area. To prevent the duplicate detection, some events are eliminated. First, the local maxima whose angle in the polar map contains most events are selected as base events and each local maximum except the base events is removed if its reconstructed line crosses any reconstructed lines of the base events in the ST map. Otherwise, the local maximum is added to the set of base events. The resulting base events are illustrated in Figure 3i, and the reconstructed lines from the base events are shown in Figure 3j.

4. Experiments and Results

4.1. Experiment Settings

To validate the proposed method, three capillary videos, denoted as videos 1–3, were recorded from three healthy subjects by experts using a portable capillaroscopy device. Each RGB video was recorded for 30 s at 30 frames per second and 640 × 480 resolution. Therefore, the input for the proposed model is a 900 × 480 × 640 × 3 array. Among the capillaries in each video, the best two were selected to ensure a clear shape in the video for segmentation. The WBC events were counted by four experts to establish the ground truth. We obtained the Korean IRB approval (No. P01-201903-11-02) to conduct the experiments involving human participants.
As hyper-parameters for the segmentation methods in Section 3.3, the labels were binarized with empirical thresholds of 150 and 80 for the deep learning and conventional segmentation, respectively. Figure 3b,c respectively show a frame of a video and the averaged label of the outputs using deep learning segmentation across all frames.
Hereinafter, the following abbreviations are used for the methods, that is, the method names with and without ‘S’ indicate a video with and without stabilization. ‘Manual’, ‘Conventional’, and ‘DNN’ denote that the capillary labels in segmentation process are obtained manually, the conventional segmentation, and the semantic segmentation based on a deep learning algorithm, respectively. For the evaluation of the proposed methods, the six combinations of the stabilization and segmentation methods are explored to evaluate the performance of WBC event counting. The configuration of each combination method is showed in Table 1.

4.2. Results

Box-plot in Figure 7 shows the variability of the WBC event counting among the experts and the orange line inside each box indicates its median value, while the different colors of ‘x’ show the number of WBC events predicted by each methods and ‘cap.’ means capillary. In cap. 2 and cap. 5, only four methods were presented because the other two, the Traditional and S-Traditional methods, failed to detect the capillary. It can be seen that our proposed method, denoted by the orange cross-mark, consistently predicts the WBC event number around the median value as the box for each capillary. It shows that the stabilization process and automatic segmentation using DNN improve the performance of predicted events.
In addition to the number of predicted events, Figure 8 shows that the proposed method correctly predicts the locations of events. In other words, a method may predict events in incorrect positions from an ST map, as shown in the results obtained from other methods. If a method predicts some events in incorrect positions, and, despite the number of predictions corresponding to the ground truth, the method could not ensure success for other input images. The proposed method captures all the lines in an ST map with correct position and number, further verifying its detection accuracy. The reconstructed lines from the ST maps of all the evaluated videos are given in Figure A1, Figure A2, Figure A3, Figure A4, Figure A5 and Figure A6 in Appendix B.

5. Discussion

5.1. Main Contributions

In this paper, we propose an automated method for detecting and counting WBCs that appear as visual gaps in nailfold capillary images. The proposed method consists of an automatic capillary segmentation method using deep learning, video stabilization, and WBC event detection algorithms. There has been no specific previous research on the WBC event detection using the nailfold capillary images. The most similar work has been conducted by Trinidad et al. [34]. However, our work has two major differences from theirs. While our work implements the automatic segmentation method and exploits only capillary-relevant intensity information, they used bounding box and brightness variations which could affect WBC count performance with noises in the background of capillary videos [34]. On the other hand, as the proposed method exploits intensities only in capillaries, it has less chance to be affected by noises in the background and therefore would be robust to the background noise. Since there is no specific previous work, the performances of the three segmentation algorithms (manual, conventional, and deep learning) with and without video stabilization were compared in this study. Experimental results demonstrate that the proposed method improves the performance of the WBC event counting and outperforms the conventional approach.

5.2. Segmentation Method

The semantic segmentation method uses a deep learning approach that could contribute to the automation of WBC event counting while replacing manual capillary segmentation. From left to the right in Figure 9, video frames and their corresponding capillary labels estimated using the manual, conventional, and deep learning segmentation are displayed.
The labels obtained by the experts and using the deep learning are similar. Likewise, methods ‘S-DNN’ and ‘S-Manual’ in Figure 10 and the Figure A7, Figure A8, Figure A9, Figure A10, Figure A11 and Figure A12 in Appendix A show that labeling using the deep learning approach is close to that obtained manually. Therefore, event counting based on the deep learning has comparable performance to that based on the manual segmentation.
Furthermore, the deep learning approach is more reliable compared with the conventional approach for the automatic capillary segmentation. Figure 9 shows that the conventional segmentation is relatively sensitive to the image quality, whereas deep learning-based segmentation is more robust. As a matter of fact, some capillaries are not reflected in the conventional labels, as shown in Figure A1, Figure A2, Figure A3, Figure A4, Figure A5 and Figure A6 in Appendix B. Therefore, deep learning-based segmentation is robust to an image in low quality and outperforms the conventional segmentation approach.

5.3. Video Stabilization

To measure how much the video stabilization process affects the WBC event counting process, the results of the WBC event counting with and without the video stabilization were compared. Figure 10 shows that the methods with video stabilization correctly predict the seven ground-truth events, whereas those without video stabilization miss some of them and identify incorrect events. This is because the ST map from a video without stabilization captures the intensities of the background rather than those of a capillary, which appear as white areas in the ST map, and they are subsequently represented as peaks in the corresponding polar map. As the intensities of the background and events are almost the same without stabilization and both represent peaks in the polar map, event detection (Section 3.6) is degraded, resulting in the miscounts.
Moreover, the norms of all translation vectors from the video stabilization process were acquired for each capillary video to measure the amount of the stabilization. If a capillary video contains more unwanted motion artifacts, then the overall norm of the translation vectors would become large and therefore the event counting results with and without the video stabilization process would differ a lot. Conversely, if a capillary video contains less motion artifacts, then the results would not differ that much. Figure 11 shows the distribution of the norms of translation vectors for each video, which could infer that videos 1 and 2 are more intensely stabilized than video 3 due to their bigger means of the L2 norm distribution. The Figure A7, Figure A8, Figure A9, Figure A10, Figure A11 and Figure A12 in Appendix A demonstrate that the predictions for capillaries 1–4 in videos 1 and 2 with all different stabilization/segmentation methods considerably differ from those for capillaries 5 and 6 in video 3.
Figure 8 and Figure 10 show the consistency between the proposed method and experts’ evaluation regarding the WBC event detection. In Figure 10, the gray vertical dashed lines and colored crosses indicate the events counted by an experts and those counted using the various methods, respectively. The cyan lines in Figure 8 visualize the counted events. The method with stabilization consistently predicts the events when compared with other methods. The event detection for all videos are reported in Figure A7, Figure A8, Figure A9, Figure A10, Figure A11 and Figure A12 in Appendix A, whose figures demonstrate the effectiveness of the video stabilization process on the prediction consistency of the event counting compared with the ground truths.

6. Conclusions

In this paper, a fully automated WBC event counting method is proposed to determine the number of visual gaps representing WBCs on capillary images using deep learning and video stabilization approaches. The proposed method segments capillary labels using a deep learning model and stabilizes the video frames with respect to the capillary labels to improve the performance of the event counting. The labels determined by the deep-learning-based segmentation are more reliable than those determined by the conventional segmentation, as the deep-learning-based segmentation gives labels close to ones obtained manually and the conventional segmentation fails to label capillaries in the case of low quality of images. Moreover, the video stabilization aligns capillaries to the same position to remove unwanted motion artifacts that might cause miscounts of events and therefore could guarantee accurate WBC event counting, resulting in comparable performance with the human experts.

Author Contributions

Conceptualization, Y.-S.H.; methodology, B.K. and Y.-S.H.; software, B.K.; validation, B.K.; formal analysis, B.K.; investigation, B.K.; resources, B.K.; data curation, B.K.; writing—original draft preparation, B.K.; writing—review and editing, Y.-S.H., C.P. and Y.-H.C.; visualization, B.K.; supervision, C.P.; project administration, Y.-H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government(MSIT) (NRF-2017R1A5A1015596) and Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Education(NRF-2020R1F1A1071712)). Additionally, the work reported in this paper was conducted during the sabbatical year of Kwangwoon University in 2020.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A. Consistency of WBC Event Counting for Various Methods

The figures in this appendix show the ST maps and corresponding reconstructed events. The cyan lines are the reconstructed events after event counting using various methods. The abbreviations denoting method combinations are detailed in Section 4.
Figure A1. Events reconstructed from event counting using various methods for capillary 1.
Figure A1. Events reconstructed from event counting using various methods for capillary 1.
Sensors 20 07101 g0a1
Figure A2. Events reconstructed from event counting using various methods for capillary 2.
Figure A2. Events reconstructed from event counting using various methods for capillary 2.
Sensors 20 07101 g0a2
Figure A3. Events reconstructed from event counting using various methods for capillary 3.
Figure A3. Events reconstructed from event counting using various methods for capillary 3.
Sensors 20 07101 g0a3
Figure A4. Events reconstructed from event counting using various methods for capillary 4 (same as Figure 8).
Figure A4. Events reconstructed from event counting using various methods for capillary 4 (same as Figure 8).
Sensors 20 07101 g0a4
Figure A5. Events reconstructed from event counting using various methods for capillary 5.
Figure A5. Events reconstructed from event counting using various methods for capillary 5.
Sensors 20 07101 g0a5
Figure A6. Events reconstructed from event counting using various methods for capillary 6.
Figure A6. Events reconstructed from event counting using various methods for capillary 6.
Sensors 20 07101 g0a6

Appendix B. Spatiotemporal Maps of Capillaries for Various Methods

The figures in this appendix show the prediction consistency of WBC event counting using various methods with respect to the ground truths provided by experts. The gray vertical dashed lines and colored crosses indicate events counted by experts and were obtained from the various methods, respectively. The abbreviations denoting method combinations are detailed in Section 4.
Figure A7. Prediction consistency of event counting compared with ground truths for capillary 1.
Figure A7. Prediction consistency of event counting compared with ground truths for capillary 1.
Sensors 20 07101 g0a7
Figure A8. Prediction consistency of event counting compared with ground truths for capillary 2.
Figure A8. Prediction consistency of event counting compared with ground truths for capillary 2.
Sensors 20 07101 g0a8
Figure A9. Prediction consistency of event counting compared with ground truths for capillary 3.
Figure A9. Prediction consistency of event counting compared with ground truths for capillary 3.
Sensors 20 07101 g0a9
Figure A10. Prediction consistency of event counting compared with ground truths for capillary 4 (same as Figure 10).
Figure A10. Prediction consistency of event counting compared with ground truths for capillary 4 (same as Figure 10).
Sensors 20 07101 g0a10
Figure A11. Prediction consistency of event counting compared with ground truths for capillary 5.
Figure A11. Prediction consistency of event counting compared with ground truths for capillary 5.
Sensors 20 07101 g0a11
Figure A12. Prediction consistency of event counting compared with ground truths for capillary 6.
Figure A12. Prediction consistency of event counting compared with ground truths for capillary 6.
Sensors 20 07101 g0a12

References

  1. Newman, T.B.; Draper, D.; Puopolo, K.M.; Wi, S.; Escobar, G.J. Combining immature and total neutrophil counts to predict early onset sepsis in term and late preterm newborns: Use of the I/T2. Pediatr. Infect. Dis. J. 2014, 33, 798. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Honda, T.; Uehara, T.; Matsumoto, G.; Arai, S.; Sugano, M. Neutrophil left shift and white blood cell count as markers of bacterial infection. Clin. Chim. Acta 2016, 457, 46–53. [Google Scholar] [CrossRef]
  3. van Wolfswinkel, M.E.; Vliegenthart-Jongbloed, K.; de Mendonça Melo, M.; Wever, P.C.; McCall, M.B.; Koelewijn, R.; van Hellemond, J.J.; van Genderen, P.J. Predictive value of lymphocytopenia and the neutrophil-lymphocyte count ratio for severe imported malaria. Malar. J. 2013, 12, 101. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Crawford, J.; Dale, D.C.; Lyman, G.H. Chemotherapy-induced neutropenia: Risks, consequences, and new directions for its management. Cancer 2004, 100, 228–237. [Google Scholar] [CrossRef] [PubMed]
  5. Golan, L.; Yeheskely-Hayon, D.; Minai, L.; Dann, E.J.; Yelin, D. Noninvasive imaging of flowing blood cells using label-free spectrally encoded flow cytometry. Biomed. Opt. Express 2012, 3, 1455–1464. [Google Scholar] [CrossRef]
  6. Uji, A.; Hangai, M.; Ooto, S.; Takayama, K.; Arakawa, N.; Imamura, H.; Nozato, K.; Yoshimura, N. The source of moving particles in parafoveal capillaries detected by adaptive optics scanning laser ophthalmoscopy. Investig. Ophthalmol. Vis. Sci. 2012, 53, 171–178. [Google Scholar] [CrossRef] [Green Version]
  7. Bourquard, A.; Butterworth, I.; Sánchez-Ferro, A.; Giancardo, L.; Soenksen, L.; Cerrato, C.; Flores, R.; Castro-González, C. Analysis of white blood cell dynamics in nailfold capillaries. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milano, Italy, 25–29 August 2015; pp. 7470–7473. [Google Scholar]
  8. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  10. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  12. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2015; pp. 3431–3440. [Google Scholar]
  13. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  14. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heisenberg, Germany, 2018; pp. 3–11. [Google Scholar]
  15. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.W.; Heng, P.A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans. Med Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Hariyani, Y.S.; Eom, H.; Park, C. DA-Capnet: Dual Attention Deep Learning Based on U-Net for Nailfold Capillary Segmentation. IEEE Access 2020, 8, 10543–10553. [Google Scholar] [CrossRef]
  17. Hegazy, M.A.; Cho, M.H.; Cho, M.H.; Lee, S.Y. U-net based metal segmentation on projection domain for metal artifact reduction in dental ct. Biomed. Eng. Lett. 2019, 9, 375–385. [Google Scholar] [CrossRef]
  18. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  19. Guizar-Sicairos, M.; Thurman, S.T.; Fienup, J.R. Efficient subpixel image registration algorithms. Opt. Lett. 2008, 33, 156–158. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Huynh, L.; Choi, J.; Medioni, G. Aerial implicit 3D video stabilization using epipolar geometry constraint. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; pp. 3487–3492. [Google Scholar]
  21. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  22. dos Anjos, A.; Shahbazkia, H.R. Bi-level image thresholding. Biosignals 2008, 2, 70–76. [Google Scholar]
  23. MacQueen, J. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Oakland, CA, USA, 21 June–18 July 1965 and 27 December 1965–7 January 1966; Volume 1, pp. 281–297. [Google Scholar]
  24. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  25. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–8 December 2012; pp. 1097–1105. [Google Scholar]
  26. Ibrokhimov, B.; Hur, C.; Kim, H.; Kang, S. An Optimized Deep Belief Network Model for Accurate Breast Cancer Classification. IEIE Trans. Smart Process. Comput. 2020, 9, 266–273. [Google Scholar] [CrossRef]
  27. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  28. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  29. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  30. Munirathinam, R.; M, S. Hybrid LSM-based Image Segmentation and Analysis of Morphological Variations of the Brainstem in Alzheimer MR Images. Trans. Smart Process. Comput. 2018, 7, 124–131. [Google Scholar] [CrossRef]
  31. Comelli, A.; Dahiya, N.; Stefano, A.; Benfante, V.; Gentile, G.; Agnese, V.; Raffa, G.M.; Pilato, M.; Yezzi, A.; Petrucci, G.; et al. Deep learning approach for the segmentation of aneurysmal ascending aorta. Biomed. Eng. Lett. 2020, 1–10. [Google Scholar] [CrossRef]
  32. Bourquard, A.; Pablo-Trinidad, A.; Butterworth, I.; Sánchez-Ferro, Á.; Cerrato, C.; Humala, K.; Urdiola, M.F.; Del Rio, C.; Valles, B.; Tucker-Schwartz, J.M.; et al. Non-invasive detection of severe neutropenia in chemotherapy patients by optical imaging of nailfold microcirculation. Sci. Rep. 2018, 8, 5301. [Google Scholar] [CrossRef] [PubMed]
  33. De Boor, C. Applied Mathematical Sciences. In A Practical Guide to Splines; Springer: Berlin/Heisenberg, Germany, 1978; Volume 27. [Google Scholar]
  34. Pablo-Trinidad, A.; Butterworth, I.; Ledesma-Carbayo, M.J.; Vettenburg, T.; Sánchez-Ferro, A.; Soenksen, L.; Durr, N.J.; Muñoz-Barrutia, A.; Cerrato, C.; Humala, K.; et al. Automated detection of neutropenia using noninvasive video microscopy of superficial capillaries. Am. J. Hematol. 2019, 94, E219–E222. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Example of a WBC event (black ellipse) over a sequence of nailfold capillary images.
Figure 1. Example of a WBC event (black ellipse) over a sequence of nailfold capillary images.
Sensors 20 07101 g001
Figure 2. Overview of the proposed WBC counting process.
Figure 2. Overview of the proposed WBC counting process.
Sensors 20 07101 g002
Figure 3. Application of the proposed WBC event counting method. (a) raw frame; (b) processed frame; (c) labels of capillaries; (d) coordinates of labels; (e) ST maps from capillaries; (f) processed ST maps; (g) polar maps obtained by applying Radon transform; (h) detected local maxima from (g); (i) local maxima selected as base events; (j) reconstructed lines according to events in (i).
Figure 3. Application of the proposed WBC event counting method. (a) raw frame; (b) processed frame; (c) labels of capillaries; (d) coordinates of labels; (e) ST maps from capillaries; (f) processed ST maps; (g) polar maps obtained by applying Radon transform; (h) detected local maxima from (g); (i) local maxima selected as base events; (j) reconstructed lines according to events in (i).
Sensors 20 07101 g003aSensors 20 07101 g003b
Figure 4. Proposed deep learning model for capillary segmentation.
Figure 4. Proposed deep learning model for capillary segmentation.
Sensors 20 07101 g004
Figure 5. Example of the video stabilization process: (a) the first frame set as a reference, (b) a frame before the stabilization, and (c) a frame stabilized based on the reference. The yellow rectangles indicate the initial position of capillaries in the reference frame. Note the shift of the frame after the stabilization process.
Figure 5. Example of the video stabilization process: (a) the first frame set as a reference, (b) a frame before the stabilization, and (c) a frame stabilized based on the reference. The yellow rectangles indicate the initial position of capillaries in the reference frame. Note the shift of the frame after the stabilization process.
Sensors 20 07101 g005
Figure 6. Example of capillary coordinate determination. (a) virtual rectangle determining capillary region of interest; (b) internal coordinates (red dots); (c) external coordinates (blue dots); (d) intermediate coordinates (green dots) between pairs of internal and external coordinates.
Figure 6. Example of capillary coordinate determination. (a) virtual rectangle determining capillary region of interest; (b) internal coordinates (red dots); (c) external coordinates (blue dots); (d) intermediate coordinates (green dots) between pairs of internal and external coordinates.
Sensors 20 07101 g006
Figure 7. Prediction variability for the capillary videos using all the different segmentation methods. The x-axis denotes the index of each capillary from videos 1 to 3, and the y-axis the number of predicted events. Each ‘×’ mark indicates the number of predicted events using the corresponding segmentation and video stabilization. The boxplot for each capillary is obtained from the events counted by four experts.
Figure 7. Prediction variability for the capillary videos using all the different segmentation methods. The x-axis denotes the index of each capillary from videos 1 to 3, and the y-axis the number of predicted events. Each ‘×’ mark indicates the number of predicted events using the corresponding segmentation and video stabilization. The boxplot for each capillary is obtained from the events counted by four experts.
Sensors 20 07101 g007
Figure 8. ST maps and reconstructed events (cyan lines) for ST maps detected by event counting for the combinations of video stabilization and capillary segmentation for capillary 4.
Figure 8. ST maps and reconstructed events (cyan lines) for ST maps detected by event counting for the combinations of video stabilization and capillary segmentation for capillary 4.
Sensors 20 07101 g008
Figure 9. Capillary segmentation for videos (a) 1, (b) 2, and (c) 3. From left to right: the raw frame, manual segmentation, conventional segmentation, and (deep learning) semantic segmentation are displayed. The selected capillaries are indicated by the yellow arrows.
Figure 9. Capillary segmentation for videos (a) 1, (b) 2, and (c) 3. From left to right: the raw frame, manual segmentation, conventional segmentation, and (deep learning) semantic segmentation are displayed. The selected capillaries are indicated by the yellow arrows.
Sensors 20 07101 g009
Figure 10. Predicted WBC event counts and ground truths for capillary 4. The x-axis denotes time stamp in the video, and the y-axis represents the different combinations of the video stabilization process and capillary segmentation methods. The gray vertical dashed lines indicate the events identified by experts. Note that all the segmentation methods with the stabilization preprocessing correctly recognize the counts of WBCs.
Figure 10. Predicted WBC event counts and ground truths for capillary 4. The x-axis denotes time stamp in the video, and the y-axis represents the different combinations of the video stabilization process and capillary segmentation methods. The gray vertical dashed lines indicate the events identified by experts. Note that all the segmentation methods with the stabilization preprocessing correctly recognize the counts of WBCs.
Sensors 20 07101 g010
Figure 11. The distribution of the norms of translation vectors for three capillary videos. Each x- and y-axis indicates the L2 norm of a translation vector and the occurrence frequency of each norm.
Figure 11. The distribution of the norms of translation vectors for three capillary videos. Each x- and y-axis indicates the L2 norm of a translation vector and the occurrence frequency of each norm.
Sensors 20 07101 g011
Table 1. Method configuration.
Table 1. Method configuration.
No.MethodSegmentation MethodVideo Stabilization
1ManualManualNo
2ConventionalConventionalNo
3DNNDNNNo
4S-ManualManualYes
5S-ConventionalConventionalYes
6S-DNNDNNYes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, B.; Hariyani, Y.-S.; Cho, Y.-H.; Park, C. Automated White Blood Cell Counting in Nailfold Capillary Using Deep Learning Segmentation and Video Stabilization. Sensors 2020, 20, 7101. https://doi.org/10.3390/s20247101

AMA Style

Kim B, Hariyani Y-S, Cho Y-H, Park C. Automated White Blood Cell Counting in Nailfold Capillary Using Deep Learning Segmentation and Video Stabilization. Sensors. 2020; 20(24):7101. https://doi.org/10.3390/s20247101

Chicago/Turabian Style

Kim, Byeonghwi, Yuli-Sun Hariyani, Young-Ho Cho, and Cheolsoo Park. 2020. "Automated White Blood Cell Counting in Nailfold Capillary Using Deep Learning Segmentation and Video Stabilization" Sensors 20, no. 24: 7101. https://doi.org/10.3390/s20247101

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop