You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

16 April 2020

Intelligent Computer-Aided Diagnostic System for Magnifying Endoscopy Images of Superficial Esophageal Squamous Cell Carcinoma

,
,
,
and
1
Division of Computer Software Engineering, Silla University, Busan 46958, Korea
2
Department of Internal Medicine, Pusan National University School of Medicine, and Biomedical Research Institute, Pusan National University, Busan 49241, Korea
3
Department of Computer Games, Yong-in Songdam College, Yongin 17145, Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Machine Learning in Medical Image Processing

Abstract

Predicting the depth of invasion of superficial esophageal squamous cell carcinomas (SESCCs) is important when selecting treatment modalities such as endoscopic or surgical resections. Recently, the Japanese Esophageal Society (JES) proposed a new simplified classification for magnifying endoscopy findings of SESCCs to predict the depth of tumor invasion based on intrapapillary capillary loops with the SESCC microvessels classified into the B1, B2, and B3 types. In this study, a four-step classification method for SESCCs is proposed. First, Niblack’s method was applied to endoscopy images to select a candidate region of microvessels. Second, the background regions were delineated from the vessel area using the high-speed fast Fourier transform and adaptive resonance theory 2 algorithm. Third, the morphological characteristics of the vessels were extracted. Based on the extracted features, the support vector machine algorithm was employed to classify the microvessels into the B1 and non-B1 types. Finally, following the automatic measurement of the microvessel caliber using the proposed method, the non-B1 types were sub-classified into the B2 and B3 types via comparisons with the caliber of the surrounding microvessels. In the experiments, 114 magnifying endoscopy images (47 B1-type, 48 B2-type, and 19 B3-type images) were used to classify the characteristics of SESCCs. The accuracy, sensitivity, and specificity of the classification into the B1 and non-B1 types were 83.3%, 74.5%, and 89.6%, respectively, while those for the classification of the B2 and B3 types in the non-B1 types were 73.1%, 73.7%, and 72.9%, respectively. The proposed machine learning based computer-aided diagnostic system could obtain the objective data by analyzing the pattern and caliber of the microvessels with acceptable performance. Further studies are necessary to carefully validate the clinical utility of the proposed system.

1. Introduction

Initial changes in superficial esophageal squamous cell carcinomas (SESCCs) begin from the morphological alteration of intrapapillary capillary loops (IPCLs) in the esophageal mucosa. In the normal esophageal mucosa, IPCLs with regular morphologies are observed in magnifying endoscopy; however, the IPCLs are dilated with irregular morphologies and their calibers change at the esophageal cancer site. These changes in IPCLs are known to differ depending on the invasion depth of esophageal cancer [1]. In clinical practice, the prediction of the invasion depth of SESCCs is a critical factor to determine the appropriate treatment modality such as endoscopic resection or surgery. Recently, the Japan Esophageal Society (JES) proposed a simplified classification method based on the IPCL patterns for estimating the invasion depth of SESCCs [2,3]. The classification criteria for the microvessels in SESCCs are based on the IPCL morphology (Figure 1). The A type indicates normal IPCLs, while the B-types indicate abnormal IPCLs which are suggestive of esophageal cancer. The B1 type consists of abnormal microvessels with a conserved loop-like formation; the B2 type consists of stretched and markedly elongated transformations without a loop-like formation; and the B3 type consists of highly dilated, irregular vessels with a caliber that appears more than three times greater than that of the B2-type vessels. The accuracy of the JES classification ranged from 78.6%–90.5% in recent research studies [2,3].
Figure 1. Schema of intrapapillary capillary loops (IPCLs) according to the Japan Esophageal Society (JES) magnifying endoscopy classification in superficial esophageal squamous cell carcinomas. Adapted from Oyama et al. [2].
However, determining the morphological changes of IPCLs with the JES classification has non-negligible interobserver variability, as such decisions depend on the knowledge and experience of the endoscopist. To exclude this subjectivity, automatic software that analyzes the features of IPCLs to predict the invasive depth of SESCCs is required. Hence, if the index of the feature patterns automatically extracted by such software matches with the proven clinical results, more accurate prediction of the invasive depth of SESCCs can be provided to the endoscopists.
In this study, a method is proposed for classifying and analyzing SESCCs using image processing and machine learning techniques on the IPCL patterns related to the invasion depth of the tumor obtained from the magnifying endoscopy images of patients with SESCCs. To achieve the proposed goal, several machine learning algorithms, such as the adaptive resonance theory 2 (ART2) algorithm, were applied to isolate the vessel area from the image and the support vector machine (SVM) to classify the IPCL patterns according to the JES classification in the magnifying endoscopy images. An algorithm for blood vessel thickness measurement is also proposed. The morphological features of IPCLs extracted from the magnifying endoscopy images of SESCCs are first classified into the B1 and non-B1 types using the learning patterns of the SVM algorithm with the non-B1 type subsequently classified into the B2 and B3 types based on the thicknesses of the IPCLs.
The overall structure of this paper is organized as follows. Section 2 describes related studies such as the SVM and conventional vessel thickness measurement algorithms employed in this study. Section 3 describes the proposed method. Section 4 describes the experiments and analysis results of the magnifying endoscopy images of SESCCs. Finally, Section 5 presents the discussion of this study and some future research directions.

3. Materials and Methods

The proposed IPCL classification method for SESCCs can be summarized as shown in Figure 3.
Figure 3. Proposed IPCL feature extraction and measurement process for analysis of magnifying endoscopy images of superficial esophageal squamous cell carcinoma.
On magnifying endoscopy images of the esophagus, it is sometimes impossible to distinguish the microvessels from the background with a single threshold value due to the reflections and shadows that occur in the images. Thus, the ROI is extracted by removing unnecessary information, such as the image capture conditions, from the endoscopy images. The fuzzy stretching technique is applied to the extracted ROI to emphasize the contrast between the microvessel (low brightness value) and background area (high brightness value). Niblack’s binarization technique is applied to extract a candidate microvessel area from the ROI with stretched contrast. Subsequently, noise regions are removed from the extracted candidate microvessel area. A fast Fourier high-frequency filter is applied to the ROI with stretched contrast to extract the microvessel boundaries and remove them from the candidate microvessel area. In addition, the ART2-based quantization technique is applied to the ROI with stretched contrast to extract the background area. The final microvessel area is then extracted by removing the background from the candidate microvessel area. Efficient and accurate noise removal is essential as the proposed method relies on pixel clustering to separate the target organ area. Fuzzy stretching and ART2-based quantization procedures are applied to separate the microvessel area accurately as explained in Section 3.1 and Section 3.2.
Eight morphological features such as the x-axis, y-axis, perimeter, area, row dispersion, column dispersion, mixed dispersion, and roundness derived from the extracted microvessel area are input to the SVM algorithm as learning data to classify the data into the B1 and non-B1 types (i.e., B2 and B3 types) as explained in Section 3.4. In addition, as the thickness of the B-type microvessels is more than three times that of the B2-type microvessels, which is the criterion suggested by the JES, the B2 and B3 types were classified from the non-B1 type of previously learned data. To use this morphological characteristic, five microvessels of the B2 type are selected based on the ratio of their area to perimeter, and the thickness measurement method is applied, as explained in Section 3.5. After measuring the microvessel thickness, the largest thickness value among 10 representative microvessels is compared with the average thickness of the selected B2-type microvessel. If this ratio is ≥ 3.7, the microvessel is classified as the B3 type.

3.1. Contrast Enhancement of the ROI Using the Fuzzy Stretching Technique

Endoscopy images exhibit irregular distributions of brightness values for the microvessels and background due to the reflections and shadows that occur during the examination; IPCLs cannot be extracted sometimes due to the low contrast between the objects and background. Therefore, in this study, the contrast of endoscopy images is emphasized/enhanced using a stretching technique. Among the stretching techniques, fuzzy stretching determines the brightness adjustment rate by calculating the distance between the lowest and highest pixel values based on the average brightness value of the images. Using these brightness adjustment rates, the upper and lower limits of the triangular membership function are calculated to set the membership function of the fuzzy set, which is then applied to the image for contrast stretching [12].
X m r = i = 0 255 X i r 1 M N , X m g = i = 0 255 X i g 1 M N , X m b = i = 0 255 X i b 1 M N
In Equation (8), X i r ,   X i g ,   and   X i b are the RGB values, and M and N are the horizontal and vertical lengths of the image, respectively. The middle brightness values X m r ,   X m g ,   and   X m b can be determined by applying Equation (8).
D m a x r = | X h r X m r | , D m i n r = | X m r X l r | D m a x g = | X h g X m g | , D m i n g = | X m g X l g | D m a x b = | X h b X m b | , D m i n b = | X m b X l b |
The distance values of the brightest pixels, D m a x r , D m a x g , and D m a x b , and those of the darkest pixels, D m i n r , D m i n g , and D m i n b , can be calculated from Equation (9) using the middle brightness values, X m r , X m g , and X m b , determined from Equation (8).
i f ( X m r > 128 )     a d j u s t m e n t r = 255 X m r     e l s e   i f ( X m r < D m i n r )     a d j u s t m e n t r = D m i n r     e l s e   i f ( X m r > D m a x r )     a d j u s t m e n t r = D m a x r i f ( X m g > 128 )     a d j u s t m e n t g = 255 X m g     e l s e   i f ( X m g < D m i n g )     a d j u s t m e n t g = D m i n g     e l s e   i f ( X m g > D m a x g )     a d j u s t m e n t g = D m a x g i f ( X m b > 128 )     a d j u s t m e n t b = 255 X m b   e l s e   i f ( X m b < D m i n b )   a d j u s t m e n t b = D m i n b           e l s e   i f ( X m b > D m a x b )   a d j u s t m e n t b = D m a x b
The D m i n r , D m i n g , D m i n b , D m a x r , D m a x g , and D m a x b values determined from Equation (9) are then applied to Equation (10) to determine the brightness adjustment rate ( a d j u s t m e n t r ,   a d j u s t m e n t g ,   a d j u s t m e n t b ).
I m a x r = X m r + a d j u s t m e n t r , I m i m r = X m r a d j u s t m e n t r I m a x g = X m g + a d j u s t m e n t g , I m i m g = X m g a d j u s t m e n t g I m a x b = X m b + a d j u s t m e n t b , I m i m b = X m b a d j u s t m e n t b
The middle brightness value ( X m r , X m g , X m b ) and brightness adjustment rate ( a d j u s t m e n t r ,   a d j u s t m e n t g ,   a d j u s t m e n t b ) are applied to Equation (11) to determine the maximum brightness value ( I m a x r ,   I m a x g ,   I m a x b ) and minimum brightness value ( I m i m r , I m i m g , I m i m b ). Furthermore, the maximum brightness value ( I m a x r ,   I m a x g ,   I m a x b ) and minimum brightness value ( I m i m r , I m i m g , I m i m b ) are applied to the triangular membership function shown in Figure 4.
I m i d r = I m a x r + I m i n r 2 , I m i d g = I m a x g + I m i n g 2   , I m i d b = I m a x b + I m i n b 2
Figure 4. Triangular membership function.
The middle brightness value ( I m i d r ,   I m i d g ,   I m i d b ) at which the membership degree becomes 1 is determined by applying the maximum brightness value ( I m a x r ,   I m a x g ,   I m a x b ) and minimum brightness value ( I m i m r , I m i m g , I m i m b ) of the membership function shown in Figure 4 to Equation (12).
i f ( X m r I m i n r ) o r ( X m r I m a x r ) t h e n   μ ( X r ) = 0 i f ( X m r > I m i d r )   t h e n   μ ( X r ) = I m a x r X m r I m a x r I m i d r i f ( X m r < I m i d r )   t h e n   μ ( X r ) = X m r I m i n r I m i d r I m i n r i f ( X m r = I m i d r )   t h e n   μ ( X r ) = 1 i f ( X m g I m i n g )   o r   ( X m g I m a x g )   t h e n   μ ( X g ) = 0 i f ( X m g > I m i d g )   t h e n   μ ( X g ) = I m a x g X m g I m a x g I m i d g i f ( X m g < I m i d g )   t h e n   μ ( X g ) = X m g I m i n g I m i d g I m i n g i f ( X m g = I m i d g )   t h e n   μ ( X g ) = 1 i f ( X m b I m i n b )   o r   ( X m b I m a x b )   t h e n   μ ( X b ) = 0 i f ( X m b > I m i d b )   t h e n   μ ( X b ) = I m a x b X m b I m a x b I m i d b i f ( X m b < I m i d b )   t h e n   μ ( X b ) = X m b I m i n b I m i d b I m i n b i f ( X m b = I m i d b )   t h e n   μ ( X b ) = 1
In addition, ( I m i n r , I m a x r ) , ( I m i n g , I max g ) , and ( I m i n b , I m a x b ) to be applied to the membership function are determined using Equation (13), and the membership degree is determined using the membership function of each channel (RGB).
i f ( I m i n r < > 0 ) c u t r = I m i n r I m a x r , e l s e   c u t r = 0.5 i f ( I m i n g < > 0 ) c u t g = I m i n g I m a x g , e l s e   c u t g = 0.5 i f ( I m i n b < > 0 ) c u t b = I m i n b I m a x b , e l s e   c u t b = 0.5
The upper ( β ) and lower ( α ) limits can be determined by applying the membership degree μ ( x ) obtained from Equation (13) to c u t r ,   c u t g ,   and   c u t b which were determined from Equation (14). The image is stretched by applying each upper limit ( β r ,   β g ,   β b ) and lower limit ( α r ,   α g ,   α b ) to Equation (15). The X r ,   X g ,   and   X b values in Equation (15) represent the previous brightness values, while X n e w r , X n e w g ,   and   X n e w b represent the new brightness values obtained after stretching.
X n e w r = X r α r β r α r 255 , X n e w g = X g α g β g α g 255 , X n e w b = X b α b β b α b 255

3.2. Vein Area Extraction and Background Area Removal

3.2.1. Extraction of the Microvessel Candidate Region Using Niblack’s Binarization

Microvessels might not be extracted from the endoscopy images owing to the small brightness contrast between the microvessels and background, of which can be attributed to the reflections and shadows in the captured images. However, unlike computers, human eyes can easily distinguish the microvessels from the background in the endoscopy images. This is called the color constancy phenomenon which is illustrated in Figure 5. In this figure, A and B are of the same color, but appear different due to the influence of the background. Humans can perceive color differences relatively consistently even when the illumination and observation conditions are changed by color constancy [13].
Figure 5. Optical illusion attributable to color constancy.
In the endoscopy images, microvessels have lower chroma values than the background. Thus, using the principle of color constancy, a local area that has a lower brightness value than the average is defined as a candidate microvessel area. Niblack’s binarization algorithm, which is a traditional binarization technique, is used to extract the candidate vessel area [14]. Niblack’s binarization sets the threshold using the average and standard deviation values in a predefined window. This technique is widely employed to document images. In this study, the window size in the endoscopy image was divided into 10 × 10 sections. Figure 6 shows the result of applying the Niblack’s binarization method to an ROI subject to fuzzy stretching.
Figure 6. Extraction of the candidate microvessel areas after binarization.

3.2.2. Noise Elimination in the Vascular Boundary Region Using a Fast Fourier High-Frequency Filter

The candidate microvessel areas extracted via binarization include noise and microvessel areas. Therefore, background areas not containing the microvessels should be removed to extract the microvessel areas accurately. First, a fast Fourier transform (FFT) is applied to separate the microvessel and background areas from the ROI with enhanced contrast. The FFT is the most general method of transforming an image into frequency space with subsequent filtering of the images. In general, a continuous Fourier transform cannot be applied, as the images contain digital data. Thus, the digital signals are transformed into the frequency domain using the discrete Fourier transform [15]. The Fourier transform for two-dimensional images is given by Equation (16), where j = 1 , u = 0 ,   1 ,   2 ,   ,   M 1 , v = 0 ,   1 ,   2 ,   ,   N 1 , f ( x ,   y ) is the original image, and M and N are the width and length of the image, respectively. F ( u ,   v ) , which is the transformed function in frequency space, is a complex-valued function. F ( u ,   v ) can be transformed back into the original image using the inverse Fourier transform as shown in Equation (17).
F ( u , v ) = x = 0 M 1 y = 0 N 1 f ( x , y ) e j 2 π ( u x / M + v y / N )
f ( x ,   y ) = 1 M N u = 0 M 1 v = 0 N 1 F ( u ,   v ) e j 2 π ( u x M + u y / N )
M 2 N 2 multiplication operations are required when the discrete Fourier transform is applied to an image with width M and length N, resulting in a computational complexity of O ( k N 4 ) , which requires significant computation time. Therefore, the FFT, which has low computational complexity, was applied in this study. The fast discrete Fourier transform algorithm has improved speed compared with the discrete Fourier transform and is applicable only when the number of input data is a multiple of 2. As the computational complexity is N log 2 N , the FFT is more effective as the number of data increases.
W N = e j 2 π / N N = 2 K
Equation (18) defines W N in Equation (17), and the number of input data N as a multiple of 2.
F ( u ) = x = 0 2 k 1 f ( x ) W 2 K u x = x = 0 K 1 [ f ( 2 x ) W 2 K u ( 2 x ) + f ( 2 x + 1 ) W 2 K u ( 2 x + 1 ) ] = x = 0 K 1 [ f ( 2 x ) W K u x ] + x = 0 K 1 [ f ( 2 x + 1 ) W K u x ] W 2 K u
The fast discrete Fourier transform calculates the even and odd data separately by default. Thus, Equation (19) can be separately defined for the even and odd data as shown as Equation (20).
F e v e n ( u ) = x = 0 K 1 f ( 2 x ) W K u x   , F o d d ( u ) = x = 0 K 1 f ( 2 x + 1 ) W K u x
In Equation (20), F e v e n ( u ) and F o d d ( u ) represent the calculation results for the even and odd data, respectively. F ( u ) can be expressed as shown in Equation (21).
F ( u ) = F e v e n ( u ) + F o d d ( u ) W 2 K u
The complex coefficient W 2 K u has periodicity and symmetry as shown in Equation (22).
W K u + K = W K u ,     W 2 K u + K = W 2 K u
I f   u + K is substituted in Equation (21) instead of u , then this satisfies F e v e n ( u + K ) = F e v e n ( u ) and F o d d ( u + K ) = F o d d ( u ) . Therefore, F ( u ) can be expressed as shown in Equation (23).
F ( u + K ) = F e v e n ( u ) + F o d d ( u ) W 2 K u
In Equation (23), if K Fourier transforms are applied to the even and odd data, the results from the K Fourier transforms can be easily obtained without a complex computational process. Four layers are created when a Gaussian high-pass filter D with cut-off frequencies D 0 = {1, 10, 20, 30} is applied to the frequency domain image obtained after an FFT. The high-pass filter only passes the high-frequency region and blocks/filters the low-frequency region. The resulting frequency domain image can then be obtained by subtracting the low-frequency region from 1 as shown in Equation (24):
H ( u , v ) = 1 e D 2 ( u , v ) / 2 D 0 2
The application of a high-pass filter tends to sharpen the resulting image. A high-pass filter can effectively distinguish the boundary between the microvessels and background, regardless of the morphologies of the microvessels. Figure 7 shows the extraction results for the boundary area in each image and the microvessel to which the high-pass filter was applied. Based on the characteristic feature of the background around the microvessels having a high brightness in the images processed with a high-pass filter, the microvessel boundary area can be extracted by setting the threshold as the average pixel value in the high-pass filtered image.
Figure 7. Binarization results for various high-pass filter cut-off frequencies ( D 0 ).
The binarization results for the cut-off frequencies shown in Figure 7b,d,f,h are combined to first extract the microvessel boundary areas and then remove the boundary areas of the extracted microvessels as shown in Figure 8.
Figure 8. Microvessel boundary areas extracted using the fast Fourier transform (FFT).

3.2.3. Removal of the Background Area Using the ART2 Algorithm

Although the FFT can effectively remove noise at the microvessel boundaries and background areas via a high-pass filter, the entire background area cannot be removed. Therefore, the ART2-based quantization technique was employed to extract the candidate background areas [16]. The ART2 is an unsupervised learning algorithm that builds clusters of similar features by autonomously learning the input patterns. The ART2 algorithm has stability because it memorizes the already learned patterns and has adaptability, as it can learn new patterns. Furthermore, the high speed of the algorithm permits real-time learning. Figure 9 shows the flowchart of the ART2 algorithm.
Figure 9. Flowchart of the adaptive resonance theory 2 (ART2) algorithm.
In this study, the vigilance parameter ( ρ ) of the ART2 algorithm was set to 0.1. Among the clusters, the pixels that belong to a cluster with high brightness are set as the background area, and the image is binarized to distinguish the background and other areas. Figure 10 shows the quantization and binarization results obtained using the ART2 algorithm.
Figure 10. Results of the ART2-based quantization and binarization.
The final microvessel area is extracted by removing the background area, shown in Figure 10b, from the candidate microvessel area from which the boundary of the microvessel has been removed.

3.3. Extraction of Object Information from the Microvessel Area

Contour Tracing Algorithm

Contour tracing is a method used to search for contour coordinates while moving along the contour of an object. Contour tracing is a widely applied effective method for acquiring object information from binary images. Table 1 presents the contour tracing algorithm [17].
Table 1. Contour tracing algorithm.
Figure 11. Contour tracing directions.
The contours of the extracted microvessels shown in Figure 12b can be obtained by applying the contour tracing algorithm to the image of the microvessel area shown in Figure 12a.
Figure 12. Contour extraction images.
Figure 13 shows the magnifying endoscopy images of SESCCs according to the JES classification. As shown in Figure 13b,c, the endoscopy images of SESCCs exhibit various types of microvessels combined. For example, the B1 and B2 microvessels coexist in the B2-type image shown in Figure 13b. Therefore, in this study, as the B2- and B3-type microvessels are slenderer and longer than the B1 type, this morphological feature was exploited to extract the microvessel contours. The proposed microvessel extraction method extracts the 10 largest microvessels in the aligned contour following rearrangement of the microvessel contours. Subsequently, these 10 microvessels are set as the representative microvessels in the endoscopy images.
Figure 13. B-type images according to the JES classification.

3.4. Extraction of Morphological Information from the Microvessels

To classify the different types of microvessels as the B1 and non-B1 types, the learning data input to the SVM are composed of eight data points from the microvessels as shown in Figure 14. The objects can be largely divided into an enclosing box. The enclosing box method is used to enclose an object within ( X m i n ,   Y m i n ) and ( X m a x ,   Y m a x ) as shown in Figure 14a, with the bounding box used to find the long and short axes of the object and surround the object with minimum area. To measure the aspect ratio of the microvessels accurately, the microvessel information is extracted using the bounding box, as shown in Figure 14b.
Figure 14. Morphological features of the microvessels.
The center point ( C x ,   C y ) , shown in Figure 14, can be calculated using the following equation:
C x = i = 1 n P x i / n ,   C y = i = 1 n P y i / n
P in Equations (25) and (26) is the perimeter of object ③ in Figure 14, and n is the pixel count along the object perimeter. P x and P y are the x and y coordinates of the perimeter, respectively.
① The x-axis is the object length, ② the y-axis is the object width, ③ the perimeter is the pixel count along the object contour, ④ the area is the pixel count enclosed by the boundary of the object, ⑤ the row dispersion, ⑥ column dispersion, ⑦ mixed dispersion, and ⑧ roundness are calculated from Equation (26) with a denoting the area of the object.
  Row   dispersion = i = 1 n ( P x i C x ) 2 / a     Column   dispersion = i = 1 n ( P y i C y ) 2 / a   Mixed   dispersion = i = 1 n ( P x i C x ) ( P y i C y ) / a   Roundness = 4 π × a P 2

3.5. Microvessel Thickness Measurement

3.5.1. Extraction of the Microvessel Center Axis

As the first step in measuring the microvessel thickness, the center axis of a microvessel is extracted by applying the thinning algorithm to the candidate microvessel area. As the microvessels in SESCCs have irregular morphologies, every point passing through the center axis of the microvessel is measured to obtain an accurate thickness measurement. Therefore, the microvessel to which the thinning algorithm is applied has branch points that form a branch shape. The branch points are removed because it is difficult to determine the search direction based on the center axis.
The mask shown in Figure 15, which was proposed by Olsen [18], is applied to determine the microvessel end points and branch points. Figure 16 shows the operation process for thinning the image by applying the mask shown in Figure 15. Figure 17 shows all possible patterns of the branch points in the thinned image and the operation results. Figure 18 shows all the end points and the corresponding operation results in the thinned image.
Figure 15. Feature detection mask.
Figure 16. Operation process using the feature detection mask.
Figure 17. Branch point patterns.
Figure 18. End point patterns.
Figure 19a shows the thinned image of the microvessels, while Figure 19b shows the microvessel center axes extracted by removing the branch points using the mask proposed by Olsen.
Figure 19. Microvessel center axes extracted by removing branch points from the thinned image.

3.5.2. Microvessel Thickness Measurement

In the microvessel object shown in Figure 20, the red, solid line represents the microvessel center axis, and the blue points represent the pixels along the center axis. For example, the pixels along the center axis are indicated by 1, …, m , where P 1 and P m represent the end points of the microvessel object.
Figure 20. Schematic view of a microvessel.
The microvessel thickness at point P n is calculated using the angle θ n defined by P n and P n + 1 . In Figure 21b, the microvessel thickness is given by the length of the straight line passing through θ n + 90 ° and θ n 90 ° . To calculate the angle θ n , the points of each center axis are arranged by distance. Thus, the distance from P 1 to P m along the center axis of the microvessel can be calculated using Equation (27). In this study, the Euclidean distance method was applied to calculate the distance.
Distance = ( x n x n + 1 ) 2 + ( y n y n + 1 ) 2
Figure 21. Thickness measurement method using the angle θ n defined by P n and P n + 1 .
Assuming that the adjacent points P n and P n + 1 exist among the points arranged around the center point, θ n is defined as shown in Figure 22 and can be calculated using Equation (28).
θ n = tan 1 y n y n + 1 x n x n + 1
Figure 22. Angle between any two points.
The microvessel thickness measurement method using θ n is illustrated in Figure 23. The green line in Figure 23 represents D , which is the length to be measured at θ n + 90 ° , while the yellow line represents D 2 which is the line to be measured at θ n 90 ° .
Figure 23. Microvessel thickness measurement method.
T ( x ,   y ) is the coordinate of the point at a distance D from P n , and T ( x ,   y ) is the coordinate of the point at a distance D from P n . D and D are increased by 1 until the microvessel wall is encountered. T ( x ,   y ) and T ( x ,   y ) can be obtained using Equation (29).
x = c o s ( θ n + 90 ° ) D x = c o s ( θ n 90 ° ) D y = s i n ( θ n + 90 ° ) D y = s i n ( θ n 90 ° ) D
The thickness measurement method shown in Figure 23 is described in Table 2.
Table 2. Flowchart of the thickness measurement method.
The thickness measurement results are shown in Figure 24. The red points represent the pixels along the center axis of the microvessel. The green and yellow lines indicate the length measurements at θ + 90 ° and θ 90 ° from the pixels along the microvessel center axis, respectively.
Figure 24. Microvessel thickness measurement results.

4. Results

4.1. Environment

Herein, experiments were conducted with 114 images (47 B1 type, 48 B2 type, and 19 B3 type images) to classify features in the magnifying endoscopy images of SESCCs. The magnifying endoscopy images were collected using a video endoscopy system with narrow-band imaging function (EVIS-LUCERA SPECTRUM system; Olympus, Tokyo, Japan) and a magnifying endoscope (GIF-H260Z; Olympus, Tokyo, Japan). The control and analysis software was developed using Visual Studio 2015 C# on a PC with an Intel Core i7-3500U CPU and 12.0 GB RAM.

4.2. Experiment Results

4.2.1. Classification Experiment Results for B1 and Non-B1 Types

In this study, the linear kernel was selected from among the various types of SVM kernels for simplicity. The SVM learning model was combined with the OAA method which is a multi-class learning technique [19]. The SVM learning was performed by setting one class of data as positive and the other classes as negative. Herein, the B1 type was selected as the positive class, while the B2 and B3 types were the negative classes. The fivefold validation method was adopted to verify the learning data. Accuracy, sensitivity, and specificity were employed as performance indices, as shown in Equation (30), to determine the pattern classification performance for SESCCs. The TP in Equation (30) represents a true positive (i.e., an actual positive was correctly predicted), while TN represents a true negative (i.e., an actual negative was correctly analyzed). Furthermore, FP represents a false negative (i.e., an actual negative was analyzed as positive), while FN represents a false negative (i.e., an actual positive was predicted as negative).
Accuracy = TP + TN TP + TN + FN + FP × 100 Sensitivity = TP TP + FN × 100 Specificity = TN FP + TN × 100
Table 3 shows the pattern classification results obtained using the proposed method where the ground truth was provided by a domain expert. Substituting the TP, TN, FP, and FN values from Table 3 into Equation (30) resulted in an accuracy of 83.3%, with a sensitivity and specificity of 74.5% and 89.6%, respectively. The graph showing the receiver operating characteristic (ROC) curve in Figure 25 was drawn using the R software.
Table 3. Classification performance results for the B1 versus non-B1 types (114 cases).
Figure 25. Receiver operating characteristic (ROC) curve for the classification results of the B1 and non-B1 types.

4.2.2. Classification Experiment Results for the B2 and B3 Types

The JES classification was used to classify the B2 and B3 types, based on the morphological criterion that the thickness of the B3 type microvessel is more than three times that of the B2 type microvessels. The JES classification was applied based on the average measured thickness value of the microvessels. The largest thickness value among the measurements was not used, as this might lead to inaccurate results in the presence of an outlier. Thus, the average of the five largest measured thickness values was set as the representative thickness of a thick microvessel. Next, the B2 type microvessels were extracted to facilitate comparisons with the representative microvessel. The B2 type microvessels have long and slender morphological features. Therefore, five microvessels with the smallest area to perimeter ratio were selected as the B2 type among 20 microvessels with an object size ranking between 11 and 30, excluding the representative microvessels. The ratio of the thickness of the representative thick microvessels to the average thickness of the five extracted B2 type microvessels was compared and analyzed. The optimal cut-off value was 3.694. The microvessel classification performance of the B2 and B3 types is shown in Figure 26.
Figure 26. The ROC curve corresponding to the classification results of the B2 and B3 types.
The classification results for the B2 and B3 types with the optimal cut-off value of 3.694 are shown in Table 4. Substituting the TP, TN, FP, and FN values from Table 4 into Equation (30) resulted in an accuracy of 71.6% with a sensitivity and specificity of 73.7% and 70.8%, respectively.
Table 4. Classification results of the B2 and B3 types with the optimal cut-off value.
The clinical classification results for the B2 and B3 types with a cut-off-value of 3.7 instead of 3.694 are shown in Table 5. The TP, TN, FP, and FN values from Table 5 resulted in an accuracy of 73.1% with a sensitivity and specificity of 73.7% and 72.9%, respectively.
Table 5. Feature classification results for the B2 and B3 types with a cut-off value of 3.7.

5. Discussion

In this study, an automatic classification method for IPCL patterns based on the JES classification was proposed to estimate the invasion depth of a tumor by extracting microvessels from the magnifying endoscopy images of SESCCs. The microvessels were extracted using the features of IPCLs in the endoscopy images, and the SVM algorithm and proposed thickness measurement method were applied to the extracted microvessels to classify them into the B1, B2, and B3 types satisfying the JES classification. The SVM algorithm employed for classification into the B1 and non-B1 types was trained using a linear kernel. To evaluate the performance of the proposed method, experiments were conducted with 114 magnifying endoscopy images of SESCCs. The results showed an accuracy of 83.3%, with a sensitivity and specificity of 74.5% and 89.6%, respectively. Furthermore, the microvessel thickness was measured and compared to facilitate the classification of the non-B1 type into the B2 and B3 types. When the cut-off ratio of the largest thickness of the representative microvessels to the average thickness of five B2-type microvessels was set to 3.7, the experiment results showed an accuracy of 73.1% with a sensitivity and specificity of 73.7% and 72.9%, respectively. This indicated that the proposed method can provide an objective interpretation of the magnifying endoscopy images of SESCCs. Thus, this study demonstrates the feasibility of the proposed method as a computer-aided diagnosis system in clinical practice.
A recent study on artificial intelligence for the classification of IPCL patterns in the endoscopic diagnosis of SESCCs analyzed 7046 magnifying endoscopy images from 17 patients (10 SESCC, 7 normal) using a convolutional neural network (CNN) [20]. The accuracy, sensitivity, and specificity for abnormal IPCL patterns were 93.7%, 89.3%, and 98%, respectively. Another recent study introduced a computer-aided diagnosis system for the real-time automated diagnosis of precancerous lesions and SESCCs in 6473 narrow-band imaging images [21]. The sensitivity and specificity for diagnosing precancerous lesions and SESCCs were 98.0% and 95.0%, respectively. These outcomes were also observed in video datasets. However, the main output/outcome of these two studies was to only differentiate between the normal and abnormal. Detailed classification into the B1, B2, and B3 types, as presented in this study, was not reported. In addition, the non-B1 type was sub-classified into the B2 and B3 types using the microvessel thickness measurement method developed in this study.
This study has several limitations. First, the evaluated dataset had a relatively small number of images (114), and all magnifying endoscopy images were retrospectively obtained from a single center. The age-standardized incidence rate was 5.9 per 100,000 population in 2013, and the proportion of SESCCs in whole esophageal cancers was less than 10% in Korea [22]. Furthermore, only high-quality magnifying endoscopy images were employed. The proposed system might produce erroneous results for images that are blurred or out of focus. Therefore, only a small number of SESCC images were included in this study. Therefore, the ability of the proposed automatic classification method to evaluate poor-quality endoscopy images should be investigated after validating its performance on a large dataset of high-quality images.
Despite these limitations, the results demonstrated the ability of the proposed machine learning based computer-aided diagnostic system to obtain objective data by analyzing the pattern and caliber of the microvessels with acceptable performance. Future work will focus on a large-scale, multi-center study to validate the diagnostic ability of the proposed system to predict the invasion depth of the tumor in SESCCs.
Furthermore, to stabilize the performance of the proposed system and validate its clinical efficacy, further research is necessary to enhance the processing speed and improve diagnostic consistency between the endoscopists’ evaluation and the proposed method. This can potentially be achieved using the deep Boltzmann machine algorithm to increase the accuracy of the pattern classification of SESCCs.

Author Contributions

Conceptualization, K.B.K. and G.H.K.; methodology, K.B.K., G.Y.Y., and G.H.K.; software, K.B.K. and G.Y.Y.; validation, G.H.K and H.K.J.; formal analysis, G.Y.Y.; resources, G.H.K.; data curation, G.Y.Y. and G.H.K.; writing—Original draft preparation, K.B.K., G.Y.Y., and G.H.K.; writing—Review and editing, K.B.K., G.Y.Y., G.H.K., D.H.S., and H.K.J.; visualization, D.H.S.; supervision, K.B.K. and G.H.K.; project administration, K.B.K.; funding acquisition, G.H.K. and H.K.J. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by Biomedical Research Institute Grant (2017B004), Pusan National University Hospital.

Conflicts of Interest

The authors declare no conflict of interest regarding the publication of this paper.

References

  1. Kumagai, Y.; Toi, M.; Kawada, K.; Kawano, T. Angiogenesis in superficial esophageal squamous cell carcinoma: Magnifying endoscopic observation and molecular analysis. Dig. Endosc. 2010, 22, 259–267. [Google Scholar] [CrossRef] [PubMed]
  2. Oyama, T.; Inoue, H.; Arima, M.; Omori, T.; Ishihara, R.; Hirasawa, D.; Takeuchi, M.; Tomori, A.; Goda, K. Prediction of the invasion depth of superficial squamous cell carcinoma based on microvessel morphology: Magnifying endoscopic classification of the Japan Esophageal Society. Esophagus 2017, 14, 105–112. [Google Scholar] [CrossRef] [PubMed]
  3. Kim, S.J.; Kim, G.H.; Lee, M.W.; Jeon, H.K.; Baek, D.H.; Lee, B.H.; Song, G.A. New magnifying endoscopic classification for superficial esophageal squamous cell carcinoma. World J. Gastroenterol. 2017, 23, 4416–4421. [Google Scholar] [CrossRef] [PubMed]
  4. Sato, H.; Inoue, H.; Ikeda, H.; Sato, C.; Onimaru, M.; Hayee, B.; Phlanusi, C.; Santi, E.G.R.; Kobayashi, Y.; Kudo, S. Utility of intrapapillary capillary loops seen on magnifying narrow-band imaging in estimating invasive depth of esophageal squamous cell carcinoma. Endoscopy 2015, 47, 122–128. [Google Scholar] [CrossRef] [PubMed]
  5. Arima, M.; Tada, T.; Arima, H. Evaluation of microvascular patterns of superficial esophageal cancers by magnifying endoscopy. Esophagus 2005, 2, 191–197. [Google Scholar] [CrossRef]
  6. Cortes, C.; Vapnik, V. Support-Vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  7. Ebrahimi, M.A.; Khoshtaghaza, M.H.; Minaei, S.; Jamshidi, B. Vision-Based pest detection based on SVM classification method. Comput. Electron. Agric. 2017, 137, 52–58. [Google Scholar] [CrossRef]
  8. Sun, F.; Xu, Y.; Zhou, J. Active learning SVM with regularization path for image classification. Multimed. Tools Appl. 2016, 75, 1427–1442. [Google Scholar] [CrossRef]
  9. Yang, X.; Yu, Q.; He, L.; Guo, T. The one-against-all partition based binary tree support vector machine algorithms for multi-class classification. Neurocomputing 2013, 113, 1–7. [Google Scholar] [CrossRef]
  10. Gutierrez, M.A.; Pilon, P.E.; Lage, S.G.; Kopel, L.; Carvalho, R.T.; Furuie, S.S. Automatic measurement of carotid diameter and wall thickness in ultrasound images. Comput. Cardiol. 2002, 29, 359–362. [Google Scholar]
  11. Lowell, J.; Hunter, A.; Steel, D.; Basu, A.; Ryder, R.; Lee Kennedy, R. Measurement of retinal vessel widths from fundus images based on 2-D modeling. IEEE Trans. Med. Imaging 2004, 23, 1196–1204. [Google Scholar] [CrossRef] [PubMed]
  12. Kim, K.B.; Song, Y.S.; Park, H.J.; Song, D.H.; Choi, B.K. A fuzzy C-means quantization based automatic extraction of rotator cuff tendon tears from ultrasound images. J. Intell. Fuzzy Syst. 2018, 35, 149–158. [Google Scholar] [CrossRef]
  13. Saxena, L.P. Niblack’s binarization method and its modifications to real-time applications: A review. Artif. Intell. Rev. 2019, 51, 673–705. [Google Scholar] [CrossRef]
  14. Samorodova, O.A.; Samorodov, A.V. Fast implementation of the Niblack binarization algorithm for microscope image segmentation. Pattern Recognit. Image Anal. 2016, 26, 548–551. [Google Scholar] [CrossRef]
  15. Broughton, S.; Bryan, K. Discrete Fourier Analysis and Wavelets: Applications to Signal and Image Processing; Wiley-Interscience: Hoboken, NJ, USA, 2008. [Google Scholar]
  16. Park, J.; Song, D.H.; Han, S.S.; Lee, S.J.; Kim, K.B. Automatic extraction of soft tissue tumor from ultrasonography using ART2 based intelligent image analysis. Curr. Med. Imaging 2017, 13, 447–453. [Google Scholar] [CrossRef]
  17. Kim, K.B.; Kim, S. A passport recognition and face verification using enhanced fuzzy ART based RBF network and PCA algorithm. Neurocomputing 2008, 71, 3202–3210. [Google Scholar] [CrossRef]
  18. Olsen, M.A.; Hartung, D.; Busch, C.; Larsen, R. Convolution approach for feature detection in topological skeletons obtained from vascular patterns. In Proceedings of the 2011 IEEE Workshop on Computational Intelligence in Biometrics and Identity Management (CIBIM), Paris, France, 11–15 April 2011; pp. 163–167. [Google Scholar]
  19. Kang, C.; Huo, Y.; Xin, L.; Tian, B.; Yu, B. Feature selection and tumor classification for microarray data using relaxed Lasso and generalized multi-class support vector machine. J. Theor. Biol. 2019, 463, 77–91. [Google Scholar] [CrossRef] [PubMed]
  20. Everson, M.; Herrera, L.; Li, W.; Luengo, I.M.; Ahmad, O.; Banks, M.; Magee, C.; Alzoubaidi, D.; Hsu, H.M.; Graham, D.; et al. Artificial intelligence for the real-time classification of intrapapillary capillary loop patterns in the endoscopic diagnosis of early oesophageal squamous cell carcinoma: A proof-of-concept study. United Eur. Gastroenterol. J. 2019, 7, 297–306. [Google Scholar] [CrossRef] [PubMed]
  21. Guo, L.; Xiao, X.; Wu, C.; Zeng, X.; Zhang, Y.; Du, J.; Bai, S.; Xie, J.; Zhang, Z.; Li, Y.; et al. Real-Time automated diagnosis of precancerous lesions and early esophageal squamous cell carcinoma using a deep learning model (with videos). Gastrointest. Endosc. 2020, 91, 41–51. [Google Scholar] [CrossRef]
  22. Jung, H.K. Epidemiology of and risk factors for esophageal cancer in Korea. Korean J. Helicobacter Up. Gastrointest. Res. 2019, 19, 145–148. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.