Next Article in Journal
Integrated Bioinformatics-Based Subtractive Genomics Approach to Decipher the Therapeutic Drug Target and Its Possible Intervention against Brucellosis
Next Article in Special Issue
Feature–Classifier Pairing Compatibility for sEMG Signals in Hand Gesture Recognition under Joint Effects of Processing Procedures
Previous Article in Journal
Utilization of Nanotechnology to Improve Bone Health in Osteoporosis Exploiting Nigella sativa and Its Active Constituent Thymoquinone
Previous Article in Special Issue
Machine Learning for Aiding Blood Flow Velocity Estimation Based on Angiography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PHF3 Technique: A Pyramid Hybrid Feature Fusion Framework for Severity Classification of Ulcerative Colitis Using Endoscopic Images

1
Department of Digital Medicine, School of Biomedical Engineering and Imaging Medicine, Army Medical University (Third Military Medical University), Chongqing 400038, China
2
Department of Gastroenterology, Daping Hospital, Army Medical University (Third Military Medical University), Chongqing 400042, China
3
Department of Gastroenterology, Sir Run Run Shaw Hospital, Zhejiang University School of Medicine, Hangzhou 310016, China
*
Authors to whom correspondence should be addressed.
Bioengineering 2022, 9(11), 632; https://doi.org/10.3390/bioengineering9110632
Submission received: 2 September 2022 / Revised: 27 October 2022 / Accepted: 31 October 2022 / Published: 1 November 2022
(This article belongs to the Special Issue Machine Learning for Biomedical Applications)

Abstract

:
Evaluating the severity of ulcerative colitis (UC) through the Mayo endoscopic subscore (MES) is crucial for understanding patient conditions and providing effective treatment. However, UC lesions present different characteristics in endoscopic images, exacerbating interclass similarities and intraclass differences in MES classification. In addition, inexperience and review fatigue in endoscopists introduces nontrivial challenges to the reliability and repeatability of MES evaluations. In this paper, we propose a pyramid hybrid feature fusion framework ( PHF 3 ) as an auxiliary diagnostic tool for clinical UC severity classification. Specifically, the PHF 3 model has a dual-branch hybrid architecture with ResNet50 and a pyramid vision Transformer (PvT), where the local features extracted by ResNet50 represent the relationship between the intestinal wall at the near-shot point and its depth, and the global representations modeled by the PvT capture similar information in the cross-section of the intestinal cavity. Furthermore, a feature fusion module (FFM) is designed to combine local features with global representations, while second-order pooling (SOP) is applied to enhance discriminative information in the classification process. The experimental results show that, compared with existing methods, the proposed PHF 3 model has competitive performance. The area under the receiver operating characteristic curve (AUC) of MES 0, MES 1, MES 2, and MES 3 reached 0.996, 0.972, 0.967, and 0.990, respectively, and the overall accuracy reached 88.91%. Thus, our proposed method is valuable for developing an auxiliary assessment system for UC severity.

Graphical Abstract

1. Introduction

Ulcerative colitis (UC) is a chronic inflammatory bowel disease characterized by mucosal inflammation, which begins in the rectum and extends proximally into the colon in a continuous manner. Bloody diarrhea is the most common early symptom of UC, and other clinical symptoms include abdominal pain, faecal urgency, tenesmus, and vomiting [1,2]. In recent years, although the incidence of UC has stabilized in developed regions, the disease burden remains high [3]. Moreover, in developing regions, with the acceleration of urbanization, the incidence of UC continues to increase, reaching 5.41 cases per 100,000 persons in India [4]. Endoscopy plays a fundamental role in the diagnosis, treatment, and management of UC, especially in monitoring disease activity and responses to treatment [5]. Endoscopic mucosal remission is an important therapeutic goal for UC, as well as the basis for evaluating future colorectal cancer risk and improving the prognostic quality of life [6]. Therefore, accurately assessing UC activity and the overall severity of the disease is critical for selecting the best management strategy for patients [7].
At present, the most commonly used evaluation index for assessing the severity of UC in clinical practice is the Mayo score [8], and the Mayo endoscopic subscore (MES) is the most important component of the overall Mayo score [6,9]. The MES evaluates the degree of damage to the intestinal mucosa. As shown in Figure 1, the MES classifies mucosal injury into four levels: normal or inactive, mild disease, moderate disease, and severe disease. However, the use of the MES for endoscopic evaluation is difficult and requires that endoscopists be trained. The reliance on subjective interpretations by endoscopists also hinders the reliability and repeatability of MES classification [10,11]. In addition, inexperience and review fatigue may lead endoscopists to misjudge the severity of UC, which may result in delayed treatment and missing the best time to change treatment decisions. However, artificial intelligence technology has been used to assist endoscopists in the rapid and accurate determination of UC severity classification.
In recent years, convolutional neural networks (CNNs) have made substantial progress in the field of computer vision and are widely used in medical image classification [12,13], segmentation, registration, reconstruction, and object detection [14,15,16]. Due to the powerful feature extraction ability of CNNs, CNN-based deep learning models have been applied to colonoscopies to identify a variety of diseases in the small intestine [17] and detect polyps [18], significantly reducing the workload of endoscopists. The excellent remote dependency modeling capabilities of the Vision Transformer (ViT) have popularized this approach in the field of computer vision [19]. Moreover, the number of technical reports on the ViT in medical image analysis has increased exponentially [20]. While the ViT can compensate for the CNN’s inability to capture global representations, the input mode of patch embedding ignores local details and lacks local inductive bias and an overall hierarchical structure. As a result, many excellent studies that combine the ViT and CNN to utilize their complementary advantages have emerged [21,22].
Compared with other medical imaging methods, colonoscopy images are closer to natural images and have three color channels. The progressive shooting characteristics of colonoscopy cause colonoscopy images to appear diversified, and many shooting points contain not only the features of the same intestinal lumen cross-section, but also the depth features of the intestinal cavity. In one colonoscopy image, although the upper left corner may be far from the lower right corner, the characteristics of these regions may be similar in the bowel lumen sectional space. Therefore, remote relationship modeling may be critical for extracting information from colonoscopy images. Inspired by the pyramid vision Transformer (PvT) proposed by Wang et al. [23,24] and rich CNN-Transformer feature aggregation networks [25], we propose a pyramid hybrid feature fusion framework ( PHF 3 ) for UC severity classification. Compared with the ViT, the PvT can learn high-resolution representations while taking into account the problem of computational consumption and has a more nuanced local feature extraction process, but its relationship modeling is still global and lacks the inductive bias unique to convolution. The pyramidal hierarchical structure of the PvT creates a unique condition for its fusion with the CNN, which avoids the problem of feature dimension mismatch when the CNN is combined with Transformer. The design of the PHF 3 dual-branch stream hybrid architecture ensures that the local features and the global representation are relatively independent while complementing each other, which provides a simple and effective way of feature fusion. The main contributions of this article can be summarized as follows:
(1)
The dual-branch pyramid hybrid architecture combines two feature extractors, namely a CNN and a PvT, to extract the deep features and cross-sectional spatial features of the intestinal cavity in colonoscopy images.
(2)
A feature fusion module was designed to integrate the local features extracted by the CNN and the global dependencies modeled by the PvT, thereby improving the classification accuracy by enhancing the feature representation ability.
(3)
At the output of the model, the second-order aggregation of the features was applied to enhance discriminative representations, which is effective for classifying the UC severity. In addition, an iterative method for covariance normalization was utilized to accelerate network training.
The remainder of this article is organized as follows. In Section 2, related work on UC severity classification and the hybrid architecture of the CNN and ViT is summarized. The proposed PHF 3 model for UC severity classification is described in detail in Section 3. The performance of the PHF 3 model is evaluated in Section 4 and compared with that of conventional deep models. A discussion and some conclusions are presented in Section 5 and Section 6, respectively.

2. Related Work

2.1. Deep Learning for UC Severity Classification

In recent years, deep-learning-based algorithms have replaced traditional machine learning methods due to superior recognition ability and end-to-end training strategies and have shown promise in gastroenteroscopy image diagnosis applications [26,27]. However, few studies have assessed the severity of UC. Ozawa et al. [28] constructed a computer-aided diagnosis system based on GoogleNet to identify MES 0 and MES 0-1, which was the first study that explored the performance of CNNs in evaluating different disease activity levels in UC. Subsequently, Stidham [29] showed that the deep learning model performed similarly to experienced human reviewers in grading the endoscopic severity of UC. In the recognition of MES 0-1 and MES 2-3, an Inception V3-based image classification architecture achieved an area under the receiver operation curve (AUC) of 0.970 (95% CI, 0.967–0.972). Bhambhvani et al. [6] developed a deep learning model based on ResNeXt101, with the aim of automatically classifying the MES of individuals with UC. However, this study included only three categories (MES 1-3), and the number of samples was small. Thus, recent deep-learning-based UC severity classification strategies are based on CNNs, and there have been few reports on four-level MES assessments. In a new study, Luo et al. [30] designed an efficient attention mechanism network (EAM-Net), and fed the features extracted from convolutional neural networks into EAM-Net and recurrent neural networks, respectively, achieving advanced results in the UC severity classification task, with an overall accuracy of 0.906 and 0.916 on two datasets, respectively. However, its DenseNet-based backbone may be difficult to complete the global modeling of the overall relationship [31].

2.2. Dual-Branch Stream Hybrid Architecture of CNN and ViT

Since the advantages and disadvantages of the CNN and ViT have been revealed, the combination of a CNN and ViT to develop a model with better performance has become a popular research topic. In general, these diverse works can be divided into three categories: conv-like Transformers, Transformer-like ConvNets, and conv-Transformer hybrids [32]. Among them, the conv-Transformer hybrid takes advantage of the CNN and ViT in a more direct and simpler way, and the dual-branch stream structure is one example of this type of model. In this kind of structure, an effective feature fusion module is critical. Peng et al. [31] proposed Conformer, which utilizes a convolution operation and a self-attention mechanism to enhance representation learning. The feature coupling unit of Conformer integrates local features and global representations at different resolutions in an interactive manner. Chen et al. [33] presented Mobile-Former, which adopts a parallel lightweight bidirectional bridge design between MobileNet and Transformer. Yoo et al. [25] designed a more concise feature aggregation method in which the flat features of Transformer linear embeddings are rearranged and concatenated and combined with CNN features. Due to the mismatch between the intermediate feature dimensions of the CNN and ViT, the design of the feature fusion module in these studies was relatively complex. Liu et al. [34] developed a hybrid architecture named CVM-Cervix, which does not include any interactions or fusion between the CNN and ViT branches, with a multilayer perceptron applied only at the output to combine the features of the two branches. The excellent performance of CVM-Cervix in cervical cancer classification tasks suggests that effective fusion at the output may be indispensable.

2.3. Higher-Order Statistics in Deep Learning

Since Lin et al. [35] proposed bilinear CNNs, many studies [36,37] have found that high-order pooling representations and deep CNN integration introduce promising improvements in challenging fine-grained visual classification tasks. Li et al. [37,38] conducted global covariance pooling for convolution features, achieving better improvements than those achieved by first-order pooling, and proposed a covariance iterative normalization method. Dai et al. [39], inspired by the work of Li et al., considered learning the feature interdependencies through the second-order statistics of the features and designed a second-order channel attention module for single-image super-resolution. Fang et al. [40] introduced a novel bilinear attention block for person retrieval, adopting the bilinear pooling method to model local feature interactions in each channel while preserving spatial structure information. Chen et al. [41] developed a new approach, fitting higher-order statistics with linear polynomials, and constructed a higher-order attention module for person re-identification, which can be simply realized by 1 × 1 convolution and an element-level addition/product. These encouraging studies demonstrate that higher-order statistics play a significant role in deep learning in enhancing the representations of discriminative features.

3. Materials and Methods

3.1. Dataset Details

This study was approved by the Ethics Committee of Daping Hospital affiliated with Army Medical University and was performed according to the Declaration of Helsinki. A total of 15,120 colonoscopy images with high quality of 768 cases were collected from the Daping Hospital affiliated with Army Medical University and Sir Run Run Shaw Hospital of Zhejiang University from January 2018 to December 2021. Each colonoscopy image was independently annotated by two endoscopic experts; when their labels were inconsistent, a third expert assisted in the discussion and they made the final decision together. Finally, the whole dataset included 4124 MES 0 images, 6669 MES 1 images, 1773 MES 2 images, and 2554 MES 3 images. Table 1 illustrates the specific data distribution. The whole dataset was randomly divided into training and test datasets at a ratio of 8:2, with the training dataset containing 12090 images and the test dataset containing 3030 images. Details on the datasets are presented in Table 2.

3.2. Overview of the Framework

The proposed PHF 3 model is illustrated in Figure 2. The PHF 3 model has a dual-branch structure, which consists of a PvT branch, a CNN branch, a feature fusion module (FFM), two dual-branch classifiers, and a fusion classifier based on second-order pooling (SOP). The structure of the PvT is described in detail in Section 3.2, while the CNN branch is based on the ResNet50 architecture. The PvT and ResNet50 model were both pretrained on the ImageNet dataset. In a colonoscopy image, the local features extracted by the CNN represent the relationship between the intestinal wall at the near-shot point and the coaxial extension line, while the global representations modeled by the PvT capture similar information in the cross-section space of the intestinal cavity. The FFM enhances the visual representation ability by combining local features with global representations in an interactive manner. The PvT and CNN branches both contain four stages, and the FFM performs feature fusion on the output from Stages 1 to 3. Then, the fused features are transmitted back to the two main branches. The PHF 3 model has three outputs: the auxiliary outputs of the dual branches adopt average pooling, while the outputs of the fourth stage are concatenated and spliced at the channel level and serve as the main output of the model after SOP and the fully connected (FC) layers. Therefore, the total loss of the model includes the sum of three losses, L o s s a l l = α L o s s p v t + β L o s s c n n + γ L o s s c o m b i n e , and the cross-entropy loss function with label smoothing is applied for all losses. α , β , and γ are the weight coefficients of the three losses, respectively, and their proportions are discussed in Section 4.1. Suppose that the true label corresponding to the n-th sample is y n { 1 , 2 , , K } , and v = ( v 1 , v 2 , v K ) is the final output of the network, that is the prediction result of sample n. The calculation is expressed as follows:
L o s s Ω = ( 1 ε ) · [ 1 N n = 1 N log ( e v y n m = 1 K e v m ) ] + ε · [ 1 N K n = 1 N k = 1 K log ( e v k m = 1 K e v m ) ]
where N is the number of samples, K is the number of classification categories, and ε is the coefficient of label smoothing. Ω can stand for pvt, cnn, and combine.

3.3. Pyramid Vision Transformer

The PvT was originally proposed for dense prediction tasks, such as semantic segmentation and target detection, and is a pure Transformer backbone [23]. The progressive shrinking pyramid and spatial-reduction attention layer in the PvT can learn high-resolution representations, reducing the computational costs. Although the architecture of the PvT is favorable for dense prediction tasks, it does not show a strong advantage for image classification tasks. The improved version of the PvT [24] includes overlapping patch embedding (OPE) and a convolutional feed-forward network, thereby reducing the computational costs and exhibiting excellent performance in classification tasks. Each stage of the PvT consists of one OPE, one block, and one normalization layer, with the block containing N basic component blocks.
Figure 3 shows the components of the i-th stage of the PvT in our study. Let H i , W i , and d i be the height, width, and embedding dimensions of the features in the i-th stage, respectively. The flattened token output in the (i-1)-th stage is reshaped, and then, OPE is carried out. In contrast to linear embeddings in the ViT, OPE is realized mainly by convolution operations with a kernel size larger than the stride size. When i = 1, the convolution kernel size in OPE is 7 and the stride is 3, while when i = 2,3,4, the convolution kernel size in OPE is 3 and the stride is 2. In the basic component block, spatial reduction (SR) is performed first; then, multihead attention (MHA) is implemented. The MHA mechanism receives a query Q, key K, and value V as the input, and the SR operation greatly reduces the scale of K and V, effectively reducing the computational overhead and strongly encouraging the model to learn higher-resolution representations. The MHA operation can be formulated as follows:
MHA ( Q , K , V ) = Concat ( h e a d 1 , , h e a d h i ) W O
head j = Attention ( Q W j Q , SR ( K ) W j K , SR ( V ) W j V )
where h i is the number of heads in the attention layer at stage i and Concat(·) is the concatenation operation. W O R d i × d i and W j Q , W j K , W j V R d i × ( d i / h i ) are linear projection parameters. The SR(·) operation can be formulated as follows:
SR ( x ) = LN ( RP 2 ( Conv ( RP 1 ( x ) , R s i ) ) W S )
where x R H i W i × d i denotes the input sequence and RP 1 (·) and RP 2 (·) are reshape operations. x R d i × H i × W i denotes the feature after RP 1 (·), and R s i represents the spatial reduction ratio, which is also the size of the kernel and stride in the convolution operation Conv(·).
At the end of the spatial reduction operation, RP 2 (·) changes x R d i × ( H i / R s i ) × ( W i / R s i ) to x R ( H i W i / ( R s i ) 2 ) × d i and LN(·) refers to layer normalization, while W S R d i × d i is a linear projection. Note that the Attention(·) calculation is consistent with the original paper [42]:
Attention ( q , k , v ) = Softmax ( q k d i / h i ) v
Depthwise convolution is introduced into the convolution feed-forward network to capture the local continuity of the input tensor. The dimensional decay factor between the two fully connected (FC) layers at the i-th stage is R m i . In our study, the settings of d, h, R s , and R m in the four stages were [64, 128, 320, 512], [1, 2, 5, 8], [8, 4, 2, 1], and [8, 8, 4, 4], while the numbers of basic component blocks in the four stages were 3, 8, 27, and 3.

3.4. Feature Fusion Module

As shown in Figure 2, we developed a complementary design between the two branches, namely the feature fusion module (FFM). The structure of the FFM is illustrated in Figure 4. The FFM receives feature maps from the PvT and CNN branches, and after the fusion at the channel scale, the output is sent back to the two main branches to enhance the complementary representation. Concretely, F p v t i R C p v t i × H i × W i and F c n n i R C c n n i × H i × W i are intermediate feature mappings from the PvT and CNN in stage i, which are aggregated by G f u s e :
M p v t i , M c n n i = Split ( G f u s e ( Concat ( F p v t i , F c n n i ) ) )
where Concat(·) is the concatenation operation and Split(·) is the tensor split operation. G f u s e consists of 1 × 1 convolutions and ReLU activation functions and is designed for channel-level fusion. The M p v t i R C p v t i × H i × W i and M c n n i R C c n n i × H i × W i obtained by splitting along the channel dimension are followed by G p v t and G c n n . Then, the fused features are transmitted back to each branch and added to the original input features F p v t i and F c n n i . It is worth noting that the FFM aggregates local features and global representations only in Stages 1 to 3; in Stage 4, the outputs of the two branches are concatenated, followed by SOP and the final classification.

3.5. Second-Order Pooling

In typical CNN structures, global average pooling implements first-order data statistics on the extracted features to determine the final classification. However, compared with the complex learning process of the CNN, the first-order statistics are relatively crude. Inspired by [37,38], in our study, second-order pooling was applied for the final abstraction of the features obtained by each branch. Specifically, F p v t 4 R C p v t 4 × H 4 × W 4 and F c n n 4 R C c n n 4 × H 4 × W 4 are the output features of the PvT and CNN branches in Stage 4, and F f i n a l R ( C p v t 4 + C c n n 4 ) × H 4 × W 4 is obtained after the concatenation operation along the channel dimension. We reshaped F f i n a l to a feature matrix X R C × S , where C = C p v t 4 + C c n n 4 and S = H 4 W 4 . The covariance matrix is calculated as follows:
Σ = X I ¯ X
I ¯ = 1 S ( I 1 S 1 )
Here, I and 1 are identity and all-ones matrices of size S × S, respectively. Covariance normalization is beneficial for discriminative representations [39], and this normalization often relies on eigenvalue decomposition (EIG) or singular-value decomposition of the matrices. However, since graphics processing units (GPUs) are not ideal for EIG implementations, Newton–Schulz iterations were adopted to accelerate the covariance normalization process. To ensure the convergence of the Newton–Schulz iteration, Σ is first normalized as follows:
A = 1 tr ( Σ ) Σ
Here, tr ( Σ ) = i C λ i denotes the trace of Σ . Given Y 0 = A and Z 0 = I, for l = 1,…, L, the Newton–Schulz iteration is given as follows:
Y l = 1 2 Y l 1 ( 3 I Z l 1 Y l 1 ) Z l = 1 2 ( 3 I Z l 1 Y l 1 ) Z l 1
Note that the pre-normalization process has an adverse effect on the network since it nontrivially changes the magnitude of the data. After the Newton–Schulz iteration, post-compensation is applied to produce the final normalized covariance matrix:
Y ^ = tr ( Σ ) Y L
where Y ^ is a symmetric matrix, and we extracted the upper triangular elements of this matrix to use as the input to the final fully connected layer.

3.6. Implementation Details

During training, randomly clipped 224 × 224 areas were fed into the network. During testing, the colonoscopy images were resized to 256 × 256, and center clipped 224 × 224 areas were used for prediction (for the InceptionV4 model, the image size was resized to 320 × 320 and the input size was 300 × 300). For data enhancement, random horizontal and vertical flips were adopted with a probability of 0.3. In addition, contrast-limited adaptive histogram equalization (Clahe) was applied to enhance the representation of the color features. The work of Mokter et al. [43] suggests that the vascular pattern is important characteristic information. As shown in Figure 5, the Clahe algorithm improved the image contrast and could highlight the vascular texture or white ulcers, while random flipping was conducive to increasing image diversity. The PHF 3 model was trained using the stochastic gradient descent (SGD) optimizer with a weight decay of 1 × 10 5 and a momentum of 0.9. The learning rate was initialized to 0.001 and decreased by 0.1 every 10 epochs in the 50 training epochs, with 128 images included in each small batch. The cross-entropy loss with label smoothing was adapted, and the coefficient ε was set to 0.1. Our work was implemented using PyTorch with two NVIDIA Quadro GV100 32G graphics cards.
For the classification task, we evaluated the performance of all methods according to the overall accuracy and the accuracy (ACC), sensitivity (SEN), specificity (SPE), positive predictive value (PPV), negative predictive value (NPV), and F1-score (F1) of each individual class. These metrics can be calculated as follows:
t o t a l a c c u r a c y = T N a l l / N
a c c u r a c y = ( T P + T N ) / N
s e n s i t i v i t y = T P / ( T P + F N )
s p e c i f i c i t y = T N / ( T N + F P )
p o s i t i v e p r e d i c t i v e v a l u e = T P / ( T P + F P )
n e g a t i v e p r e d i c t i v e v a l u e = T N / ( T N + F N )
F 1 s c o r e = 2 × r e c a l l × p r e c i s i o n r e c a l l + p r e c i s i o n
Here, N is the number of samples, while T N a l l denotes the number of all samples with correct predictions. TP, TN, FP and FN represent the number of true positive, true negative, false positive, and false negative samples in each prediction category, respectively.

4. Experiment and Results

4.1. Preliminary Study

In the final feature fusion process, the Newton–Schulz iterative procedure was applied to achieve fast covariance normalization, where the number of iterations was a tunable hyperparameter. Therefore, we first explored the impact of the number of Newton–Schulz iterations L on model performance. The results are shown in Figure 6, where L = 0 indicates that SOP was not adopted. When L = 1–8, the accuracy fluctuated to some extent, but the overall trend was rising. After L = 8, the overall accuracy decreased, indicating that increasing the number of iterations is not conducive to improving accuracy, which is consistent with the discussion in [37]. Therefore, in our final model, L was set to 8.
In addition, we also explored the proportion of weight coefficients α , β , and γ for the three losses: L o s s p v t , L o s s c n n , and L o s s c o m b i n e . Intuitively, we believed that L o s s c o m b i n e was more important, and the experimental results are shown in Table 3. To a certain extent, increasing the weight of L o s s c o m b i n e can effectively improve the performance of the model. Therefore, we applied α : β : γ = 1:1:2 in our final model. On the training set, 5-fold cross-validation was applied to select various parameters for model training, and the area under the receiver operating characteristic curve (AUC) values of each class in the cross-validation analysis were 0.986, 0.951, 0.957, and 0.981 (Figure 7), while the overall accuracy was 85.45%.

4.2. The Ablation Experiments and Comparison Experiments with Classical CNNs

The results of the ablation experiments are shown in Table 4. Compared with the branch structure alone, the PHF 3 model was able to combine the local features extracted by ResNet50 with the global representations modeled by the PvT to achieve better classification performance. The overall accuracy of ResNet50, the PvT, and the PHF 3 model were 86.01%, 87.29%, and 88.91%, respectively. A performance comparison of some representative CNN models is shown in Table 5. Compared with ResNet50, ResNet101 had no significant improvement in terms of the recognition performance, suggesting that increasing the network depth did not enhance the representations of discriminant features. Considering the balance between performance and efficiency, our proposed PHF 3 model adopted ResNet50 and the PvT as the two main branches and outperformed both individual models. The ROC curves of the five models on the test set are shown in Figure 8, and the AUC values of the PHF 3 model for MES 0, MES 1, MES 2, and MES 3 reached 0.996, 0.972, and 0.990, respectively.

4.3. Comparative Experiments with Advanced Models

Even the basic ViT (ViT-B) exhibited better prediction accuracy than most CNNs. Moreover, compared to the ViT-B, the the PvT, which learns higher-resolution representations, improved the overall accuracy by approximately 1.0 percentage points and exhibited performance comparable to that of the basic Swin Transformer (Swin-B) [44]. In order to highlight the advantages of the PHF 3 model, we also compared it with advanced models such as the VAN [45] and MViT [46], and Conformer [31], as a representative of the two-branch CNN-Transformer structure, was also included in our comparison scope. The results are shown in Table 6, from which we can find that the VAN did not achieve exciting performance. In comparison, the Swin-B , MViT, and PvT performed better, which implies that the effectiveness of the pyramid structure and the overlap and transformation of patches may be beneficial for learning diffuse lesions. The confusion matrices of the six models are shown in Figure 9, visually illustrating the differences in their predictions. Compared with the Swin-B, the PHF 3 model had slightly higher false negatives for MES 1 and false positives for MES 3, but the Swin-B was more likely to predict MES 0 as MES 1 and MES 3 as MES 2. As a trade-off between recall and precision, the F1-score considers the false positive and false negative rates, and the PHF 3 model had better F1-scores than the comparison models in all categories.

4.4. Visualization of Feature Maps and Heat Maps

The feature maps in the model inference process helped in understanding the feature capture characteristics of the CNN and PvT branches and the effectiveness of the FFM. As shown in Figure 10, we plotted partial feature maps before and after fusion for each stage. Overall, the extracted features of both branches became increasingly abstract as the model deepened, with the CNN branch focusing on local features and highlighting local details, while the PvT branch focused on global representations and the overall performance was chaotic. The FFM clearly introduced global information into the CNN branch, while the local details introduced into the PvT branch were more difficult to notice due to the chaotic feature representations. Furthermore, as shown in Figure 11, to enhance the interpretability of the model, Grad-CAM [47] was used to draw the heat maps. Some local ulcers and bleeding were highlighted, which are important features when the model makes decisions.

5. Discussion

5.1. A Novel Deep Learning Framework for UC Severity Assessment

In clinical practice, evaluating the severity of UC through MES is of great significance for understanding patient conditions and providing effective treatment. Ensuring the reliability and reproducibility of UC severity classification remains a nontrivial challenge, and previous works have been limited to convolutional neural networks, thus ignoring the global dependencies of features in the intestinal lumen. Our work provides a novel solution to this challenge. In colonoscopy images, we considered not only the local relationship between the depth features of the intestinal cavity, but also the global dependencies of the features in the same intestinal cavity cross-section. The proposed hybrid architecture combines local features and global representations in a simple and effective manner, achieving better performance than the baseline models. In the ablation study, the overall accuracy of the PHF 3 model was 2.90 and 1.62 percentage points higher than that of ResNet50 and PvT, respectively. The feature fusion process of the two-branch hybrid framework can be observed obviously by the visualization of the feature maps. Compared with the classical CNNs, the accuracy of the proposed method was improved by 2.38–4.26 percentage points, and the AUC of the proposed method was the highest in all categories. Even compared to advanced models based on the Transformer architecture and CNN-Transformer combined framework, the PHF 3 model still had the upper hand. Our approach highlighted the importance of the fusion of local features and global representations in the feature capture of diffuse lesions and the effectiveness of second-order information in enhancing fused discriminative features. This exciting result prompted us to believe that our study will further advance the application of deep learning as an auxiliary diagnostic tool in intestinal digestive diseases.

5.2. Multi-Branch Hybrid Architectures May Be Irreplaceable

The ViT and CNN are two mainstream models of deep learning at present. However, they all have their own shortcomings, such as the difficulty of global relationship modeling for the CNN and the lack of local inductive bias for the ViT. Therefore, a natural idea is to combine them to complement each other’s advantages, such as introducing convolution operators for the ViT or adding global attention mechanism for the CNN. Actually, the convolution operation is introduced into the PvT structure, which is a conv-like Transformer. Our experiments suggest that, even if convolutions are introduced to capture local relations in Transformers, the effects may be limited, and there is still room for improvement. Similarly, it is worth exploring whether Transformer-like ConvNets can be enhanced further by introducing the ViT. In studies combining the CNN and ViT, a hybrid structure with multiple branches may be a fusion method that is difficult to replace. Another issue worth noting is how to weight losses incurred by multiple branches. Although we tested weighting with different proportions, this artificial setup may not be optimal. If possible, designing adaptive weighting methods may yield surprising results. In addition, due to the complex attention mechanism and model design, the training and reasoning of the ViT are not as fast as those of the CNN, which increases the computational costs of the CNN-ViT fusion structure. The SR strategy in the PvT effectively reduced the amount of computations. In a recent report, inspired by the SR in the PvT, Li et al. [48] proposed the Next-ViT, a new paradigm that fuses convolutional and Transformer modules during every stage, aiming to improve model efficiency and achieve industrial-scale deployment of the CNN-Transformer hybrid architecture. Therefore, how to achieve efficient computation in a multi-branch hybrid framework is also a problem that needs more research and experiments.

5.3. Higher-Order Statistics Require More Exploration

It is well known that first-order statistics may limit modeling capabilities; however, second-order statistics are difficult to apply in GPUs because they introduce additional computations. In our study, covariance normalization realized by Newton–Schultz iterations achieved end-to-end training with acceptable computational costs, providing a new approach for higher-order statistics applications. In a preliminary study, we explored the influence of the number of Newton–Schultz iterations. Although the more iterations, the better the fitting, the experimental results showed that the existence of a certain fitting bias is beneficial to the generalization performance of the model. Our research may show that second-order information has higher discriminative representation power for the fusion features of two branches. A very smooth thought is whether statistics larger than second-order can perform better or whether it is an effective combination of first-order and second-order information that needs more research to explore. In addition, when the computational costs are no longer a hindrance, attention mechanisms based on higher-order statistics are an exciting concept. At present, some studies have used second-order statistical information to construct attention modules [39,40]. Furthermore, the kinds of feature representations that can be enhanced by higher-order information should be investigated.

5.4. Limitations and Future Work

Despite the excellent and encouraging classification results, our research has several limitations. First, the number of samples in each category in our dataset was not balanced, which is a characteristic of most clinical diseases: there are always more remission samples than severe case samples. UC severity classification is a more refined identification, and simple upsampling or downsampling of the data cannot provide effective improvements. Although class weighting based on loss can improve the classification accuracy of small sample categories, this process inevitably reduces the overall accuracy. Second, due to differences between the equipment and endoscopist manipulations, it is challenging to construct a model with strong generalizability. In our study, the data were obtained from only two large centers, and multicenter verification was lacking. In the future, we will consider collecting multisource data and applying our model to colonoscopy video processing. Moreover, we will explore the non-substitutability of the dual-branch hybrid architecture and applications of higher-order attention mechanisms.

6. Conclusions

In this paper, we proposed PHF 3 , a novel dual-branch hybrid feature fusion framework with the PvT and ResNet50 for UC severity classification in endoscopic images. Compared with the ViT, the PvT can learn higher-resolution representations, which is beneficial for learning the diffuse lesion features of UC. Moreover, the PvT has high operation efficiency. The designed FFM structure solves the feature fusion problem of the CNN and PvT as the feature resolution pyramid changes, thereby effectively combining local features and global representations, while the SOP module enhances discriminant information by using second-order statistics. The comparative ablation studies were encouraging, showing that the PHF 3 model exhibited better performance than the comparison methods and can, thus, be used as an auxiliary tool for UC severity assessment. We hope that our work provides new ideas for combining convolution and Transformer and its application in assisted recognition in colonoscopy images.

Author Contributions

Conceptualization, J.Q.; data curation, G.R. and Q.C.; formal analysis, J.L. and Y.N.; funding acquisition, J.Q. and Y.W.; investigation, J.Q. and J.L.; methodology, J.Q.; project administration, Y.Y. and Y.W.; software, J.Q.; supervision, Y.W. and Y.N.; validation, J.Q.; visualization, J.Q.; writing—original draft, J.Q.; writing—review and editing, J.Q. and Y.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Chongqing Postgraduate Research Innovation Project (CYS22746) and the Key Project of Chongqing Social Livelihood (cstc2018jscx-mszdX0026).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of Army Medical Center of PLA affiliated with Army Medical University (Third Military Medical University) (No. 2021-285, date: 14 December 2021).

Informed Consent Statement

No applicable.

Data Availability Statement

The data cannot be made available due to privacy restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UCUlcerative colitis
MESMayo endoscopic subscore
CNNsConvolutional neural networks
ViTVision Transformer
PvTPyramid vision Transformer
PHF 3 Pyramid hybrid feature fusion framework
FFMFeature fusion module
SOPSecond-order pooling
FCFully connected layer
OPEOverlapping patch embedding
SRSpatial reduction
MHAMultihead attention
SGDStochastic gradient descent
ROCThe receiver operating characteristic curve
AUCThe area under the receiver operating characteristic curve
ClaheContrast-limited adaptive histogram equalization
GPUsGraphics processing units

References

  1. Kobayashi, T.; Siegmund, B.; Le Berre, C.; Wei, S.C.; Ferrante, M.; Shen, B.; Bernstein, C.N.; Danese, S.; Peyrin-Biroulet, L.; Hibi, T. Ulcerative colitis. Nat. Rev. Dis. Prim. 2020, 6, 74. [Google Scholar] [CrossRef] [PubMed]
  2. Feuerstein, J.D.; Moss, A.C.; Farraye, F.A. Ulcerative Colitis. Mayo Clin. Proc. 2019, 94, 1357–1373. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Ng, S.C.; Shi, H.Y.; Hamidi, N.; Underwood, F.E.; Tang, W.; Benchimol, E.I.; Panaccione, R.; Ghosh, S.; Wu, J.; Chan, F.; et al. Worldwide incidence and prevalence of inflammatory bowel disease in the 21st century: A systematic review of population-based studies. Lancet 2017, 390, 2769–2778. [Google Scholar] [CrossRef]
  4. Ng, S.C.; Kaplan, G.G.; Tang, W.; Banerjee, R.; Adigopula, B.; Underwood, F.E.; Tanyingoh, D.; Wei, S.C.; Lin, W.C.; Lin, H.H.; et al. Population Density and Risk of Inflammatory Bowel Disease: A Prospective Population-Based Study in 13 Countries or Regions in Asia-Pacific. Am. J. Gastroenterol. 2019, 114, 107–115. [Google Scholar] [CrossRef]
  5. Spiceland, C.M.; Lodhia, N. Endoscopy in inflammatory bowel disease: Role in diagnosis, management, and treatment. World J. Gastroenterol. 2018, 24, 4014–4020. [Google Scholar] [CrossRef]
  6. Bhambhvani, H.P.; Zamora, A. Deep learning enabled classification of Mayo endoscopic subscore in patients with ulcerative colitis. Eur. J. Gastroenterol. Hepatol. 2021, 33, 645–649. [Google Scholar] [CrossRef]
  7. Pabla, B.S.; Schwartz, D.A. Assessing Severity of Disease in Patients with Ulcerative Colitis. Gastroenterol. Clin. N. Am. 2020, 49, 671–688. [Google Scholar] [CrossRef]
  8. Pagnini, C.; Di Paolo, M.C.; Mariani, B.M.; Urgesim, R.; Pallotta, L.; Vitale, M.A.; Villotti, G.; d’Alba, L.; De Cesare, M.A.; Di Giulio, E.; et al. Mayo Endoscopic Score and Ulcerative Colitis Endoscopic Index Are Equally Effective for Endoscopic Activity Evaluation in Ulcerative Colitis Patients in a Real Life Setting. Gastroenterol. Insights 2021, 12, 217–224. [Google Scholar] [CrossRef]
  9. Gutierrez Becker, B.; Arcadu, F.; Thalhammer, A.; Gamez Serna, C.; Feehan, O.; Drawnel, F.; Oh, Y.S.; Prunotto, M. Training and deploying a deep learning model for endoscopic severity grading in ulcerative colitis using multicenter clinical trial data. Ther. Adv. Gastrointest Endosc. 2021, 14, 2631774521990623. [Google Scholar] [CrossRef]
  10. Mohammed Vashist, N.; Samaan, M.; Mosli, M.H.; Parker, C.E.; MacDonald, J.K.; Nelson, S.A.; Zou, G.Y.; Feagan, B.G.; Khanna, R.; Jairath, V. Endoscopic scoring indices for evaluation of disease activity in ulcerative colitis. Cochrane Database Syst. Rev. 2018, 1, Cd011450. [Google Scholar] [CrossRef]
  11. Takenaka, K.; Ohtsuka, K.; Fujii, T.; Negi, M.; Suzuki, K.; Shimizu, H.; Oshima, S.; Akiyama, S.; Motobayashi, M.; Nagahori, M.; et al. Development and Validation of a Deep Neural Network for Accurate Evaluation of Endoscopic Images From Patients With Ulcerative Colitis. Gastroenterology 2020, 158, 2150–2157. [Google Scholar] [CrossRef] [PubMed]
  12. Dutta, A.K.; Mageswari, U.R.; Gayathri, A.; Dallfin Bruxella, M.J.; Ishak, M.K.; Mostafa, S.M.; Hamam, H. Barnacles Mating Optimizer with Deep Transfer Learning Enabled Biomedical Malaria Parasite Detection and Classification. Comput. Intell. Neurosci. 2022, 2022, 7776319. [Google Scholar] [CrossRef] [PubMed]
  13. Raza, A.; Ayub, H.; Khan, J.A.; Ahmad, I.; Salama, A.S.; Daradkeh, Y.I.; Javeed, D.; Rehman, A.U.; Hamam, H. A Hybrid Deep Learning-Based Approach for Brain Tumor Classification. Electronics 2022, 11, 1146. [Google Scholar] [CrossRef]
  14. Cai, L.; Gao, J.; Zhao, D. A review of the application of deep learning in medical image classification and segmentation. Ann. Transl. Med. 2020, 8, 713. [Google Scholar] [CrossRef] [PubMed]
  15. Torres-Velázquez, M.; Chen, W.J.; Li, X.; McMillan, A.B. Application and Construction of Deep Learning Networks in Medical Imaging. IEEE Trans. Radiat. Plasma Med. Sci. 2021, 5, 137–159. [Google Scholar] [CrossRef]
  16. Yousef, R.; Gupta, G.; Yousef, N.; Khari, M. A holistic overview of deep learning approach in medical imaging. Multimed. Syst. 2022, 28, 881–914. [Google Scholar] [CrossRef]
  17. Ding, Z.; Shi, H.; Zhang, H.; Meng, L.; Fan, M.; Han, C.; Zhang, K.; Ming, F.; Xie, X.; Liu, H.; et al. Gastroenterologist-Level Identification of Small-Bowel Diseases and Normal Variants by Capsule Endoscopy Using a Deep-Learning Model. Gastroenterology 2019, 157, 1044–1054. [Google Scholar] [CrossRef]
  18. Ahmad, O.F.; Soares, A.S.; Mazomenos, E.; Brandao, P.; Vega, R.; Seward, E.; Stoyanov, D.; Chand, M.; Lovat, L.B. Artificial intelligence and computer-aided diagnosis in colonoscopy: Current evidence and future directions. Lancet Gastroenterol. Hepatol. 2019, 4, 71–80. [Google Scholar] [CrossRef] [Green Version]
  19. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. In Proceedings of the 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, 3–7 May 2021. [Google Scholar]
  20. Shamshad, F.; Khan, S.; Zamir, S.W.; Khan, M.H.; Hayat, M.; Khan, F.S.; Fu, H. Transformers in Medical Imaging: A Survey. arXiv 2022, arXiv:2201.09873. [Google Scholar]
  21. Guo, J.; Han, K.; Wu, H.; Xu, C.; Tang, Y.; Xu, C.; Wang, Y. CMT: Convolutional Neural Networks Meet Vision Transformers. arXiv 2021, arXiv:2107.06263. [Google Scholar]
  22. Wu, H.; Xiao, B.; Codella, N.; Liu, M.; Dai, X.; Yuan, L.; Zhang, L. CvT: Introducing Convolutions to Vision Transformers. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 22–31. [Google Scholar]
  23. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 548–558. [Google Scholar]
  24. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. PVTv2: Improved Baselines with Pyramid Vision Transformer. arXiv 2021, arXiv:2106.13797. [Google Scholar]
  25. Yoo, J.; Kim, T.; Lee, S.; Kim, S.H.; Lee, H.; Kim, T.H. Rich CNN-Transformer Feature Aggregation Networks for Super-Resolution. arXiv 2022, arXiv:2203.07682. [Google Scholar]
  26. Javaid, A.; Shahab, O.; Adorno, W.; Fernandes, P.; May, E.; Syed, S. Machine Learning Predictive Outcomes Modeling in Inflammatory Bowel Diseases. Inflamm. Bowel. Dis. 2022, 28, 819–829. [Google Scholar] [CrossRef] [PubMed]
  27. Min, J.K.; Kwak, M.S.; Cha, J.M. Overview of Deep Learning in Gastrointestinal Endoscopy. Gut Liver 2019, 13, 388–393. [Google Scholar] [CrossRef]
  28. Ozawa, T.; Ishihara, S.; Fujishiro, M.; Saito, H.; Kumagai, Y.; Shichijo, S.; Aoyama, K.; Tada, T. Novel computer-assisted diagnosis system for endoscopic disease activity in patients with ulcerative colitis. Gastrointest Endosc. 2019, 89, 416–421. [Google Scholar] [CrossRef]
  29. Stidham, R.W.; Liu, W.; Bishu, S.; Rice, M.D.; Higgins, P.D.R.; Zhu, J.; Nallamothu, B.K.; Waljee, A.K. Performance of a Deep Learning Model vs Human Reviewers in Grading Endoscopic Disease Severity of Patients With Ulcerative Colitis. JAMA Netw. Open 2019, 2, e193963. [Google Scholar] [CrossRef] [Green Version]
  30. Luo, X.; Zhang, J.; Li, Z.; Yang, R. Diagnosis of ulcerative colitis from endoscopic images based on deep learning. Biomed. Signal Process. Control 2022, 73, 103443. [Google Scholar] [CrossRef]
  31. Peng, Z.; Huang, W.; Gu, S.; Xie, L.; Wang, Y.; Jiao, J.; Ye, Q. Conformer: Local Features Coupling Global Representations for Visual Recognition. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 357–366. [Google Scholar]
  32. Li, J.; Chen, J.; Tang, Y.; Wang, C.; Landman, B.A.; Zhou, S.K. Transforming medical imaging with Transformers? A comparative review of key properties, current progresses, and future perspectives. arXiv 2022, arXiv:2206.01136. [Google Scholar]
  33. Chen, Y.; Dai, X.; Chen, D.; Liu, M.; Dong, X.; Yuan, L.; Liu, Z. Mobile-Former: Bridging MobileNet and Transformer. arXiv 2021, arXiv:2108.05895. [Google Scholar]
  34. Liu, W.; Li, C.; Xu, N.; Jiang, T.; Rahaman, M.M.; Sun, H.; Wu, X.; Hu, W.; Chen, H.; Sun, C.; et al. CVM-Cervix: A hybrid cervical Pap-smear image classification framework using CNN, visual Transformer and multilayer perceptron. Pattern Recognit. 2022, 130, 108829. [Google Scholar] [CrossRef]
  35. Lin, T.Y.; RoyChowdhury, A.; Maji, S. Bilinear CNN Models for Fine-Grained Visual Recognition. In Proceedings of the 2015 IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, 7–13 December 2015; pp. 1449–1457. [Google Scholar]
  36. Lin, T.Y.; Maji, S. Improved Bilinear Pooling with CNNs. In Proceedings of the British Machine Vision Conference 2017, BMVC 2017, London, UK, 4–7 September 2017. [Google Scholar]
  37. Li, P.; Xie, J.; Wang, Q.; Gao, Z. Towards Faster Training of Global Covariance Pooling Networks by Iterative Matrix Square Root Normalization. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 947–955. [Google Scholar]
  38. Li, P.; Xie, J.; Wang, Q.; Zuo, W. Is Second-Order Information Helpful for Large-Scale Visual Recognition? In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 2089–2097. [Google Scholar]
  39. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.T.; Zhang, L. Second-Order Attention Network for Single Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 11065–11074. [Google Scholar]
  40. Fang, P.; Zhou, J.; Roy, S.K.; Petersson, L.; Harandi, M. Bilinear Attention Networks for Person Retrieval. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea, 27 October–2 November 2019; pp. 8029–8038. [Google Scholar] [CrossRef]
  41. Chen, B.; Deng, W.; Hu, J. Mixed High-Order Attention Network for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea, 27 October–2 November 2019; pp. 371–381. [Google Scholar] [CrossRef] [Green Version]
  42. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  43. Mokter, M.F.; Oh, J.H.; Tavanapong, W.; Wong, J.; Groen, P.C. Classification of Ulcerative Colitis Severity in Colonoscopy Videos Using Vascular Pattern Detection. Mach. Learn. Med. Imaging 2020, 12436, 552–562. [Google Scholar] [CrossRef]
  44. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
  45. Guo, M.H.; Lu, C.; Liu, Z.N.; Cheng, M.M.; Hu, S. Visual Attention Network. arXiv 2022, arXiv:2202.09741. [Google Scholar]
  46. Li, Y.; Wu, C.Y.; Fan, H.; Mangalam, K.; Xiong, B.; Malik, J.; Feichtenhofer, C. MViTv2: Improved Multiscale Vision Transformers for Classification and Detection. arXiv 2021, arXiv:2112.01526. [Google Scholar]
  47. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
  48. Li, J.; Xia, X.; Li, W.; Li, H.; Wang, X.; Xiao, X.; Wang, R.; Zheng, M.; Pan, X. Next-ViT: Next Generation Vision Transformer for Efficient Deployment in Realistic Industrial Scenarios. arXiv 2022, arXiv:2207.05501. [Google Scholar]
Figure 1. Representative images of Mayo endoscopy score. (a) MES 0: normal or inactive; (b) MES 1: mild, erythema, decreased vascular pattern, mild friability; (c) MES 2: moderate, marked erythema, absent vascular pattern, friability, erosions; (d) MES 3: severe, spontaneous bleeding, ulcerations.
Figure 1. Representative images of Mayo endoscopy score. (a) MES 0: normal or inactive; (b) MES 1: mild, erythema, decreased vascular pattern, mild friability; (c) MES 2: moderate, marked erythema, absent vascular pattern, friability, erosions; (d) MES 3: severe, spontaneous bleeding, ulcerations.
Bioengineering 09 00632 g001
Figure 2. Illustration of the pyramid hybrid feature fusion framework ( PHF 3 ). The stem consists of a convolution, a batch normalization, a ReLU activation function, and maximum pooling; FFM: feature fusion module, SOP: second-order pooling, LayerNorm: layer normalization, FC: fully connected layer.
Figure 2. Illustration of the pyramid hybrid feature fusion framework ( PHF 3 ). The stem consists of a convolution, a batch normalization, a ReLU activation function, and maximum pooling; FFM: feature fusion module, SOP: second-order pooling, LayerNorm: layer normalization, FC: fully connected layer.
Bioengineering 09 00632 g002
Figure 3. The basic component block of the pyramid vision Transformer (PvT). OPE: overlapping patch embedding, SR: spatial reduction, CFF: convolutional feed-forward.
Figure 3. The basic component block of the pyramid vision Transformer (PvT). OPE: overlapping patch embedding, SR: spatial reduction, CFF: convolutional feed-forward.
Bioengineering 09 00632 g003
Figure 4. The structure of the feature fusion module (FFM).
Figure 4. The structure of the feature fusion module (FFM).
Bioengineering 09 00632 g004
Figure 5. Data enhancement presentation. (a) The original image; (b) contrast-limited adaptive histogram equalization (Clahe); (c) image clipping; (d) horizontal flipping; (e) vertical flipping.
Figure 5. Data enhancement presentation. (a) The original image; (b) contrast-limited adaptive histogram equalization (Clahe); (c) image clipping; (d) horizontal flipping; (e) vertical flipping.
Bioengineering 09 00632 g005
Figure 6. Performance comparison of PHF 3 under different Newton–Schulz iterations L.
Figure 6. Performance comparison of PHF 3 under different Newton–Schulz iterations L.
Bioengineering 09 00632 g006
Figure 7. ROC curves for 5-fold cross-validation.
Figure 7. ROC curves for 5-fold cross-validation.
Bioengineering 09 00632 g007
Figure 8. ROC curves for five models. (a) MES 0, (b) MES 1, (c) MES 2, and (d) MES 3.
Figure 8. ROC curves for five models. (a) MES 0, (b) MES 1, (c) MES 2, and (d) MES 3.
Bioengineering 09 00632 g008
Figure 9. Confusion matrix for the six models.
Figure 9. Confusion matrix for the six models.
Bioengineering 09 00632 g009
Figure 10. Partial feature maps before and after fusion at each stage. before: feature maps before fusion, fusion: feature maps after fusion, add: the blended features are superimposed on the original image and are also the input for the next stage.
Figure 10. Partial feature maps before and after fusion at each stage. before: feature maps before fusion, fusion: feature maps after fusion, add: the blended features are superimposed on the original image and are also the input for the next stage.
Bioengineering 09 00632 g010
Figure 11. Visualization of the attention maps obtained by Grad-CAM.
Figure 11. Visualization of the attention maps obtained by Grad-CAM.
Bioengineering 09 00632 g011
Table 1. The specific data distribution.
Table 1. The specific data distribution.
MES 0MES 1MES 2MES 3Total
Daping Hospital3299247995010107738
Sir Run Run Shaw Hospital825419082315447382
Total412466691773255415,120
Table 2. Distribution of experimental datasets.
Table 2. Distribution of experimental datasets.
MES 0MES 1MES 2MES 3Total
Training dataset329853321419204112,090
Test dataset82613373545133030
Total412466691773255415,120
Table 3. Exploration of the ratio of α , β , and γ .
Table 3. Exploration of the ratio of α , β , and γ .
α : β : γ 1:1:11:1:21:1:31:1:41:1:5
Total ACC88.5588.9188.7888.4588.05
Table 4. The ablation study.
Table 4. The ablation study.
ModelMetricsMES 0MES 1MES 2MES 3
ResNet50ACC95.4889.0493.0094.49
SEN89.5988.6362.7189.47
SPE97.6989.3797.0195.51
PPV93.5586.8173.5180.24
NPV96.1690.8795.1697.80
F191.5387.7167.6884.61
PvTACC95.8190.1793.3795.25
SEN88.1491.3269.7787.52
SPE98.6889.2596.4996.82
PPV96.1787.0372.4384.88
NPV95.6992.8796.0297.44
F191.9889.1271.0886.18
PHF 3 ACC96.7791.2293.9995.84
SEN90.1991.1773.1691.81
SPE99.2391.2696.7596.66
PPV97.7789.1774.8684.86
NPV96.4392.9096.4698.30
F193.8390.1674.0088.20
Table 5. Comparison of the proposed PHF 3 with some representative CNN models.
Table 5. Comparison of the proposed PHF 3 with some representative CNN models.
VGG19ResNet101DenseNet121InceptionV4 PHF 3
Total ACC85.8186.5385.4584.6588.91
MES 0
      ACC95.4895.8494.8894.6596.77
      SEN89.2388.9887.2988.2690.19
      SPE97.8298.4197.7397.0599.23
      PPV93.8995.4593.5191.8197.77
      NPV96.0495.9795.3595.6696.43
      F191.5092.1190.2990.0093.83
MES 1
      ACC88.7589.7088.2587.6291.22
      SEN86.9189.9090.2888.1191.17
      SPE90.1989.5586.6587.2491.26
      PPV87.5087.1684.2384.5189.17
      NPV89.7291.8291.8690.2892.90
      F187.2088.5187.1586.2790.16
MES 2
      ACC93.0092.4892.6792.4893.99
      SEN69.7768.0859.6060.7373.16
      SPE96.0895.7097.0596.6796.75
      PPV70.1767.7072.7670.7274.86
      NPV96.0095.7794.7894.9096.46
      F169.9767.8965.5365.3574.00
MES 3
      ACC94.3995.0595.0894.5595.84
      SEN88.5086.5587.7286.3591.81
      SPE95.5996.7896.5896.2396.66
      PPV80.3584.5783.9682.3484.86
      NPV97.6197.2597.4797.1998.30
      F184.2385.5585.8084.3088.20
Table 6. Comparison of the proposed PHF 3 with some advanced Transformer models.
Table 6. Comparison of the proposed PHF 3 with some advanced Transformer models.
VAN [45]MViT-B [46]ViT-B [42]Swin-B [44]Conformer [31] PHF 3
Total ACC85.6887.2686.2787.4687.1688.91
MES 0
      ACC94.9596.1195.8495.8795.4896.77
      SEN85.9688.6287.7786.9285.8490.19
      SPE98.3298.9198.8799.2399.0999.23
      PPV95.0596.8396.6797.6997.2697.77
      NPV94.9295.8795.5795.2994.9296.43
      F190.2792.5492.0191.9991.1993.83
MES 1
      ACC88.5190.0389.2790.2689.9091.22
      SEN89.9891.6290.9591.7091.5591.17
      SPE87.3688.7887.9589.1388.6091.26
      PPV84.9086.5785.6386.9586.3889.17
      NPV91.6993.0792.4893.1592.9992.90
      F187.3689.0388.2189.2688.8990.16
MES 2
      ACC92.9492.9492.5493.2393.3393.99
      SEN65.8263.2866.3870.0668.9373.16
      SPE96.5296.8696.0096.3096.5696.75
      PPV71.4772.7368.7171.4772.6274.86
      NPV95.5395.2295.5796.0595.9296.46
      F168.5367.6767.5370.7670.7274.00
MES 3
      ACC94.9595.4594.8895.5495.6195.84
      SEN87.7290.2585.3889.2890.4591.81
      SPE96.4296.5096.8296.8296.6696.66
      PPV83.3384.0384.5685.1384.6784.86
      NPV97.4797.9897.0197.7998.0398.30
      F185.4787.0384.9787.1687.4688.20
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qi, J.; Ruan, G.; Liu, J.; Yang, Y.; Cao, Q.; Wei, Y.; Nian, Y. PHF3 Technique: A Pyramid Hybrid Feature Fusion Framework for Severity Classification of Ulcerative Colitis Using Endoscopic Images. Bioengineering 2022, 9, 632. https://doi.org/10.3390/bioengineering9110632

AMA Style

Qi J, Ruan G, Liu J, Yang Y, Cao Q, Wei Y, Nian Y. PHF3 Technique: A Pyramid Hybrid Feature Fusion Framework for Severity Classification of Ulcerative Colitis Using Endoscopic Images. Bioengineering. 2022; 9(11):632. https://doi.org/10.3390/bioengineering9110632

Chicago/Turabian Style

Qi, Jing, Guangcong Ruan, Jia Liu, Yi Yang, Qian Cao, Yanling Wei, and Yongjian Nian. 2022. "PHF3 Technique: A Pyramid Hybrid Feature Fusion Framework for Severity Classification of Ulcerative Colitis Using Endoscopic Images" Bioengineering 9, no. 11: 632. https://doi.org/10.3390/bioengineering9110632

APA Style

Qi, J., Ruan, G., Liu, J., Yang, Y., Cao, Q., Wei, Y., & Nian, Y. (2022). PHF3 Technique: A Pyramid Hybrid Feature Fusion Framework for Severity Classification of Ulcerative Colitis Using Endoscopic Images. Bioengineering, 9(11), 632. https://doi.org/10.3390/bioengineering9110632

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop