Next Article in Journal
Vibration Energy Coupling Behavior of Rolling Mills under Double Disturbance Conditions
Next Article in Special Issue
Low Complexity Speech Enhancement Network Based on Frame-Level Swin Transformer
Previous Article in Journal
A DDoS Attack Detection Method Based on Natural Selection of Features and Models
Previous Article in Special Issue
FESSD: Feature Enhancement Single Shot MultiBox Detector Algorithm for Remote Sensing Image Target Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Feature Pyramid Network and Feature Fusion Single Shot Multibox Detector for Real-Time Prostate Capsule Detection

1
Electronic Information School, Wuhan University, Wuhan 430072, China
2
School of Information Engineering, Wuhan Business University, Wuhan 430056, China
3
Department of Urology, Zhongnan Hospital, Wuhan 430072, China
*
Authors to whom correspondence should be addressed.
Electronics 2023, 12(4), 1060; https://doi.org/10.3390/electronics12041060
Submission received: 20 November 2022 / Revised: 1 February 2023 / Accepted: 10 February 2023 / Published: 20 February 2023

Abstract

:
In the process of feature propagation, the low-level convolution layers of the forward feature propagation network lack semantic information, and information loss occurs when fine-grained information is transferred to higher-level convolution; therefore, multi-stage feature fusion networks are needed to solve the interaction between low-level convolution layers and high-level convolution layers. Based on a two-way feature feedback network and feature fusion mechanism, we created a new object detection network called Feature Pyramid Network (FPN)-based Feature Fusion Single Shot Multibox Detector (FFSSD). A bottom-up and top-down architecture with lateral connections enhances the detector’s ability to extract features, then high-level multi-scale semantic feature maps are utilized to generate a feature pyramid network. The results show that the proposed network the mAP for prostate capsule image detection reaches 83.58%, providing real-time detection ability. The context interaction mechanism can transfer high-level semantic information to low-level convolution, and the resulting convolution after low-level and high-level fusion contains richer location and semantic information.

1. Introduction

Early object detection techniques did not use feature fusion technology [1,2,3]. Following the emergence of deep learning, more and more feature fusion techniques have been used in object detection networks, including image pyramid, feature fusion for R-CNN series, Single Shot Multibox Detector (SSD), Feature Pyramid Network (FPN) in FoveaBox, Sparse R-CNN and Task-aligned One-stage Object Detection (TOOD), Feature Fusion Single Shot Multibox Detector (FSSD), Weighted Bidirectional feature pyramid Network (BiFPN) in You Only Look Once Version 4 (YOLOv4) and EfficientDet, Path Aggregation Network (PANet) in YOLOv4, and more [4,5,6,7,8,9,10,11,12,13,14]. An image pyramid is composed of multiple multi-resolution and multi-scale images, such as a Gauss/Laplacian pyramid. Ziming Liu proposed IPG-Net using image pyramid on 2020 CVPR to solve the problem of feature misalignment and small target deep feature loss in object detection [15]. To an extent, an image pyramid makes the results of object detection more accurate. The disadvantage is that as the pyramid level increases, the amount calculation increases and the detection speed slows down. The feature fusion work of R-CNN series makes prediction on the last layer. Taking Faster R-CNN as an example, the algorithm relies on multi-stage training to improve the detection precision, and the precision improvement of feature fusion for Faster R-CNN is limited. Each feature layer of the SSD series is detected separately. Its disadvantage is that there is less convolution at the lower level involved in detection (only one) and the location information may not be accurate enough. The feature fusion of FPN adopts the methods of bottom-up, lateral connection, and top-down, which improves the accuracy of object detection. Its disadvantage is that the network is more complex, with accuracy improved at the expense of speed. The feature fusion strategy of FSSD is to construct the feature pyramid through bottom-up and multi-scale fusion. FSSD fuses the features of three scales and uses bilinear interpolation and 1 × 1 convolution to adjust the size of the feature map and the number of channels. FSSD follows the idea that the small-scale convolution layer is not useful; thus, part of the size and number of channels are discarded during feature fusion. The above works use a single pyramid network to complete feature extraction, and the results of object detection may not be accurate. YOLOv4 and EfficientDet ultized BiFPN for feature fusion, which introduces learnable weights to learn the importance of different input features while repeatedly applying top-down and bottom-up multi-scale feature fusion. PANet is based on the proposed instance segmentation framework enhances the entire feature hierarchy by using accurate low-level positioning signals, thereby shortening the information path between the low-level and top-level features; however, experiments have shown that such complex networks may not work well with small data sets.
Fusing different networks to extract features has recently become very popular. Ibtissam Bakkouri et al. proposed a multistage fusion network for skin injury recognition [16]. The most discriminant features were extracted based on the VGG, ResNet, and DenseNet networks, with three features output from the three networks fused by the feature fusion layer. Xiaojing Qiu proposed a multi-stage feature fusion network for medical image segmentation which included a basic feature expression layer, intermediate feature fusion layer, deep feature fusion layer, and feature fusion module [17]. The middle layer and the fusion layer were used to complement the feature information of the lower layer by means of convolution, channel superposition, and corresponding element addition. In this approach, the deep fusion layer uses the supplementary information of deep and shallow features while the three-level feature representation layer completes the feature aggregation process. Lingzhi Kong used DenseNet and VGG16 for feature extraction and fusion [18]. In this approach, the input of DenseNet is the original image, and the input of VGG16 is the image processed by ResNet. Their aim was to quickly achieve COVID-19 detection using an AI approach. Their proposed feature fusion network integrates different backbone networks; however, it neglects the feedback and semantic transfer of high-level convolution to low-level convolution.
The current prostate capsule research includes the surgical level, contour recognition, thickness calculation, and volume calculation. At the surgical level, Zia et al. explored the U-shaped incision of the prostate capsule in laparoscopic prostatectomy, and Alon Z. Weizer evaluated the risk of urothelial and prostate cancers in a series of patients undergoing radical cystoprostatectomy with preserved prostate capsule [19,20]. Rania Hussein used prostate shape and generalized Hough transform to automatically recognize the contour of the prostate capsule [21]. K.G. Holder et al. studied the relationship between prostate volume, prostate capsule thickness, and gland epithelial cell density in the peripheral zone (PZ) using tissue anatomical imaging technology [22]. In this paper, we utilize FPN to complete the reconstruction of low-level convolution, then take the convolution reconstructed by fusion as the basis to perform 2×, 4×, 8×, 16× and 32× downsampling. To complete the real-time detection of the prostate capsule, we construct a new network, FFSSD, based on the FPN and FSSD networks.
This article’s innovations include the following:
(1)
To reduce the impact of salt and pepper noise on the object detection network, a salt and pepper noise reduction method based on edge feature preservation is proposed. Compared with fourteen other methods, this method has the highest peak signal-to-noise ratio. The proposed image denoising methods can improve the mAP of Faster R-CNN, YOLOv4, TOOD, SSD, and FSSD.
(2)
Based on FPN and FSSD, a multistage bidirectional feature fusion network called FFSSD is proposed. Compared with other algorithms such as Faster R-CNN, YOLOv4, TOOD, SSD, FSSD, OWOD, Foveabox, Sparse R-CNN, and Efficientdet, the proposed algorithm has the highest mAP on the prostate capsule detection task.

2. Materials and Methods

2.1. Dataset

The images of the prostate capsule were extracted from surgical videos. Due to the interference of gastrointestinal peristalsis and other noise, salt and pepper noise was easily introduced into the images. Using the Composed Fusion (CF) algorithm to denoise pictures with noise can ensure the smooth development of subsequent object detection work. The CF algorithm is described in Algorithm 1.
Algorithm 1: CF algorithm
Input: Noisy_img v, noisy level x
Output: p ( R ) , s ( R ) , R, d e n o i s e d _ i m g { R }
  1: Function u = compose(v, z)
  2: Initialize d = {}, p = ones(0,6), s = ones(0,6), R = 0, M A X = 0, D = {2,6}
  3: set D e n o i s e d _ i m g { 1 } = AMF(v,x)
  4: set D e n o i s e d _ i m g { 2 } = NAMF(v,x)
  5: set D e n o i s e d _ i m g { 3 } = ANN(v,x)
  6: set D e n o i s e d _ i m g { 4 } = RUN_ADF( D e n o i s e d _ i m g { 1 } , D e n o i s e d _ i m g { 2 } ,0)
  7: set D e n o i s e d _ i m g { 5 } = RUN_ADF( D e n o i s e d _ i m g { 1 } , D e n o i s e d _ i m g { 3 } ,0)
  8: set D e n o i s e d _ i m g { 6 } = RUN_ADF( D e n o i s e d _ i m g { 2 } , D e n o i s e d _ i m g { 3 } ,0)
  9: MAX = PSNR(
  10: for all ( s p ) D
  11:   p ( k ) = PSNR( d e n o i s e d _ i m g { k } ,o)
  12:   s ( k ) = COMPUTE_FOM( d e n o i s e d _ i m g { k } ,o)
  13:  IF( p ( k ) > M A X )
  14:  { M A X = p ( k ) ; R = k}
  15: end for
  16: return p ( R ) , s ( R ) , R, d e n o i s e d _ i m g { R }
In the CF algorithm, Adaptive Median Filter (AMF), Anisotropic Diffusion Fusion (ADF), Non-local Adaptive Mean Filter (NAMF), and Artificial Neural Network (ANN) algorithms are metioned, while Peak Signal-to-Noise Ratio (PSNR) and Figure of Merit (FoM) are the criteria used to evaluate image denoising. The CF algorithm steps are as follows:
(1)
The image with noise is denoised by AMF, NAMF, and ANN.
(2)
Anisotropic diffusion fusion is used to combine the image denoising results of AMF, NAMF, and ANN in pairs.
(3)
The noise reduction results of the three algorithms and the pin-two fusion results of the three algorithms are combined, then the maximum value of the combination according to the PSNR is used to obtain the final image denoising cascade optimization results.
In Figure 1, the CF is closer to the original and performs better than individual algorithms (Tristate Filter (TF), AMF, Median Filter (MF), Adaptive Total Variation L1 Regularization (ATVL1), Decision-Based Algorithms (DBA), Noise Adaptive Fuzzy Switching Median (NAFSM), NAMF, ANN, Convolutional Neural Network (CNN)-based methods, ADF-based methods, Adaptive Centre Weight Median Filter (ACMF), and Noise Adaptive Edge-Preserving Filter (NAPEF)). The areas shown in detail after image denoising are indicated by the red rectangles [23,24,25,26,27,28,29,30,31].
Compared to ANN and NAMF, CF performs the best, maintaining better edge detail after noise reduction; see Figure 2.
The proposed network was verified on three datasets, all of which contained 597 images split into 478 training and 119 test images. The datasets used consisted of the Primary Component Analysis (PCA)-based dataset, PCA+CF-based dataset, and original dataset. For the first two datasets, preprocessing was only performed on the training set, while the test set remained unchanged.
In addition, compared with other detection objects of fixed size, the prostate capsule has the characteristics of large-scale changes and different shapes. If the object detection algorithm is directly used to detect the prostate capsule from an image, the mAP of the resulting detection may not be high enough (Figure 3).

2.2. The Proposed Network

Based on FPN and FSSD, we propose a new network called FFSSD. The feature pyramid network consists of three parts, two feature forward propagation parts and one backward propagation part. A new feature pyramid was formed by three feature propagation parts.

2.2.1. First Feature Forward Propagation

Feature forward propagation consists of two parts. The first feature forward propagation is based on the feature pyramid of the SSD network, and the feature propagation ends at convolution layer conv7_2 (Figure 4). Fine-grained position information is passed from the lower convolution to the higher convolution. This forward propagation belongs to the VGG16 feature extraction backbone network.

2.2.2. Reverse Feature Propagation

After the first forward feature is passed to conv7_2, reverse feature propagation begins. After the first forward feature information completes forward propagation of position information, reverse feature propagation can transmit the semantic information of the prostate capsule to the lower convolution layer. In this way, the higher-level convolution has the fine-grained information transmitted by the lower-level convolution, and the lower-level convolution has the semantic information returned by the higher-level convolution (Figure 5).
As shown in Figure 5, we adjusted the conv7_2 through 1 × 1 convolution layers to obtain conv7_2_reduce; p1 was formed by conv7_2_reduce (2x upsampling) and fc7, while fea_concat was formed by conv4_3 and p1 (2× upsampling).

2.2.3. New Feature Pyramid Network

After completion of the reverse propagation, the new feature pyramid is generated by the second feature forward propagation. The second feature forward propagation builds on the first forward propagation and reverse feature propagation. The forward and reverse feature propagation of conv4_3, fc7, conv7_2 constitute an FPN feature pyramid network. Bidirectional passing (forward propagation and reverse propagation) constitutes the context interaction mechanism, which integrates and supplements the low-level location information and high-level semantic information. The three convolution layers are fused by the FPN feature pyramid to form a new convolution layer. On this basis, second feature forward propagation is carried out to generate a new feature pyramid network (Figure 6 and Figure 7).
On the basis of the new fea_concat layer generated by FPN fusion, downsampling of 2×, 4×, 8×, 16×, and 32× was implemented to carry out the second forward propagation of features. Compared with SSD, which is only downsampled on a single convolution layer, conv4_3, the details of the shallow convolution layer of the multistage feature fusion method are richer thanks to bidirectional feature fusion, which enables the high level to learn more information about the prostate capsule. In Figure 7, the orange dotted line box represents the FPN feature fusion network and the blue dotted box indicates the feature fusion network of the FSSD. The small dotted line box in black represents the fusion process of the lower-layer feature graphs by one transverse connection ( 1 × 1 conv) and 2× upsampling with higher convolution. BatchNorm accelerates the convergence process of the convolutional network, and the 1 × 1 conv eliminates the effect caused by upsampling. The fusion details are shown in the large dotted black box.

2.2.4. FFSSD Network

Based on the new feature pyramid network, the FFSSD network is designed as shown Figure 8.
The FFSSD backbone network is the VGG16 network. The feature fusion network consists of one FPN network and one FSSD fusion module. The FPN network consists of conv4_3, fc7, and conv7_2 through feature forward propagation and reverse propagation. The first feature forward propagation transmits the position information of the model from the bottom layer to the top layer; the position information gradually decreases while the semantic information gradually increases. Through this initialization, the network ensures the preliminary determination of the category of the object by high-level convolution. In the second reverse propagation, the semantic (type) information is transferred from the upper level (conv7_2) to the lower level (conv4_3) and the semantic information in the lower level convolution is increased, that is, the lower-level convolution (conv4_3) now carries the location information of the prostatic envelope and has rough knowledge of its semantic information. After the fusion of semantic and positional information through bi-directional transmission of the FPN feature pyramid network, the conv4_3 fusion of positional and semantic information carries the dual information for the third feature propagation, that is, the second feature forward propagation. This time, based on fusion convolution, the fusion information is transmitted from the lower level to the upper level by bidirectional transmission, and the final category determination and location calibration are carried out.
The VGG16 backbone network of FFSSD is the same as that of FSSD. Unlike SSD, the convolution of feature maps less than 10 × 10 is abandoned. An FPN network consists of two feature fusion modules. There are three convolution layers involved in FPN feature fusion, namely, conv4_3, fc7, and conv7_2; the dimensions of the three convolutional layers are 38 × 38 × 512 , 19 × 19 × 1024 , and 10 × 10 × 512 , respectively. The convolution layer with the smallest scale is fused to form the first module through the second smallest convolution layer with twice the upsampling and adjusted channels, and the second module is fused with the feature degree of the maximum scale with twice the upsampling and adjusted channels. The two feature fusion modules of FPN were fused with conv4_3 through double upsampling and transverse connection, which together constitute the feature fusion module of FSSD.
The feature extraction layer involved in detection in the network structure is based on the result of two-stage fusion, and the result layer of two-stage fusion is obtained by 2×, 4×, 8×, 16×, and 32× downsampling.

3. Results

The framework used for training the deep learning networks was Caffe/pytorch. The hardware environment for network training was an Intel Core-i7-8700 (Microsoft Corporation, Redmond, WA, USA), the main frequency of the CPU was 3.2 GHz, the running memory was 8 GB, the graphics card model was an NVIDIA GTX 1070/Tesla v100 (Microsoft Corporation, Redmond, WA, USA), and the operating system was Linux Ubuntu 64-bit (Red Hat, Raleigh, NC, USA). The learning rate of the first 500 iterations was 0.0001, and the batch_size used for training was 1.

3.1. Criteria for Evaluation

In object detection techniques, algorithms are generally evaluated on their mAP (mean Average Precision) and FPS (Frame Per Second). The mAP indicates the Average Precision A P of all detection classes, while the A P refers to the area enclosed under the Precison–Recall (PR) curve, with a larger area being better. Its formula is defined as follows:
P r e c i s i o n = T P T P + F P R e c a l l = T P T P + F N
P r e c i s i o n generally refers to the number of correct classes in all results, while R e c a l l generally refers to the recall ratio, which represents the proportion of positive categories in all the results. For object detection, if the Intersection over Union (IOU) is larger than a given threshold and positive samples are detected, this is considered a True Positive T P , which means that the correct sample prediction is positive; in other words, the IOU is larger than the threshold. When the IOU of the prediction box is less than the threshold provided by the system, the correct sample prediction is marked as negative, that is, as a False Positive F P ; a positive class predicted as a negative class is a False Negtive F N , while a negative class predicted correctly is a True Negative T N .
In order to draw the Precision–Recall (PR) curve, the detection results are first sorted by the system will sort according to the descending order of the confidence score. After assigning a value (rank), only accuracy and recall scores higher than this value are calculated. Changing the rank value changes the final recall value. We set eleven recall values ( [ 0 , 0.1 , , 0.9 , 1.0 ] ) , equal to eleven rank values, and selected eleven confidence thresholds. Generally speaking, to calculate mAP it is necessary to first calculate the average precision A P of each sample. Here, A P is defined as the mean value of the precision after eleven recalls, with the mean value of the eleven maximum precision values taken as the mean precision. The formula is as follows:
A P = 1 11 r 0.0 . 1 , . . . , 1 p i n t e r p ( r ) w h e r e p i n t e r p ( r ) = max r ˜ : r ˜ > = r p ( r ˜ )
In the formula, p ( r ˜ ) refers to the measurement accuracy when it is recalled to r ˜ . Here, p i n t e r p ( r ) takes only the maximum precision where the recall rate exceeds r. When calculating A P , the prediction box is first calculated and sorted. Second, a value of k is set to calculate the respective F P and T P values to ensure that the recall is equal to the above eleven values, then the calculation accuracy is determined. Third, the eleven accuracies are averaged to obtain the A P . The mAP is the sum of all A P values divided by the total number of A P values. FPS stands for the frame rate per second. FPS represents the real-time speed of the object detection model, and represents the total number of images that can be detected per second. In addition, the time taken to detect an image has been used by many researchers to evaluate the detection speed of object detection models.

3.1.1. mAP and Loss

The mAP curve transformation during SSD, FSSD, and FFSSD training (a total of 5000 training times) is shown in Figure 9.
Because of the participation of the two levels, namely, feature fusion and pre-training weight, the starting mAP of FFSSD is high. When the batch size is equal to 1, the mAP of the FSSD fluctuates greatly.
The transformation of the loss curve (5000 times in total) during SSD, FSSD, and FFSSD training is shown in Figure 10.
Because of the participation of two-stage feature fusion and pre-training weight, the loss of FFSSD decreases rapidly.

3.1.2. Feature Visualization

A feature visualization comparison of the conv3_3, conv4_3, conv5_3, and fc7 features of the convolution layer of FSSD and FFSSD is shown in Figure 11.
Compared with FSSD, the convolution layer in FFSSD can obtain more and richer features due to the addition of FPN and PCA. FSSD performs multi-fold downsampling based on the fusion of three convolution layers, then performs multi-scale detection. FFSSD performs multifold downsampling on the basis of three-times fusion of three convolution layers. In FPN, the upper convolution is fused with the lower convolution by double upsampling. The convolution layer after fusion is convolved by BatchNorm and 1 × 1 , which makes the model converge quickly during training, eliminates the noise caused by the upper convolution, and provides more nonlinear factors.

3.1.3. Speed and Precision Comparison

We first show a speed and precision comparison of FSSD, FSSD, and PCA+FFSSD on NVIDIA 1070 (Table 1). The speed is expressed as img/s, that is, the time required to detect an image.
The proposed FSSD includes a new feature fusion model based on SSD, and on this basis, mAP improved by 1.9%, with a 1.62% increase when FPN is used on top of FSSD. In the case of small samples, PCA was used to accelerate the convergence, pre-training weight, and multi-stage feature fusion, achieving a final the detection precision of 82.35%. In terms of time, FSSD takes 4 s to detect all test images, SSD takes 3 s to detect all test images, and PCA+FFSSD takes 5 s to detect all test images. Another very important reason why PCA+FFSSD has such a high precision is that we use the weight of FSSD+FPN training 50,000 times as the initial pre-training weight of PCA+FFSSD, meaning that the loss and mAP are in relatively low and relatively high positions, respectively. While FFSSD is nearly twice as fast as SSD, mAP has results of 82.35% in real-time. Compared with FSSD, the addition of FPN slows down FFSSD, while the detection precision of FFSSD is improved due to the top-down connection, the lateral connection between convolutional layers, the bottom-up connection, and the addition of pre-training weights. FFSSD is able to transmit information from both high-level to low-level and from low-level to high-level. The pre-training weight of FSSD+FPN makes the starting point of FFSSD very high, and the loss value is very low; thus, the detection precision is greatly improved.
Next, we provide a comparison with other object detection algorithms such as Faster R-CNN (VGG16), Faster R-CNN (ResNet 50+FPN), SSD (VGG16), SSD (ResNet-101), Efficient-D0, Efficient-D1, FSSD (VGG16), FoveaBox (ResNet+FPN), TOOD (ResNet 50+FPN), YOLOv4 (CSPDarknet-53), and Sparse R-CNN (ResNet 50+FPN) (Table 2). Their backbone networks compared algorithmically by mAP and FPS are shown in parentheses. An FPS is greater than 12.5 can be considered real-time. Faster R-CNN is only approximately real-time, and Faster R-CNN (ResNet 50+FPN) and Efficient-D7 do not reach real-time; thus, we use a short horizontal line to indicate the threshold.
The proposed method has the highest detection precision and real-time detection speed. As the scale and shape of medical images can vary greatly, a general object detection algorithm is not necessarily suitable for detection of the prostate capsule, requiring verification by experiment. Image denoising can improve the mAP of algorithms such as SSD, FSSD, Faster R-CNN, YOLOv4, and TOOD. Other algorithms, such as EfficientDet, FoveaBox, Sparse R-CNN, and OWOD, do not require image denoising.

3.1.4. Detection Results

A comparison of the detection results of SSD, FSSD, and PCA+FFSSD is shown in Figure 12.
PCA+FFSSD has better detection results and more accurate positioning. This accuracy is mainly due to the effect of blending the feature fusion mechanism of FPN and FSSD. PCA provides the effect of small sample accelerated convergence.
The mAP of FFSSD can be improved to 83.58% using PCA+FFSSD trained weights and using the PCA+CF dataset for fine-tuning during training. Because the network is unchanged, the detection speed is unchanged as well. The mAP results for ten training epochs are compared in Figure 13.
Because of the pre-training weight, the mAP of the PCA+CF+FFSSD is relatively high; FoveaBox is second (Figure 14).
Compared with the original model, PCA+CF+FFSSD detects the location of the prostate capsule more accurately.

4. Discussion and Conclusions

Although FSSD integrates the features of all three scales, it is easy to lose semantic details after multiple downsampling. However, as SSD only has one single-layer convolution to participate in subsampling, much of the location information is ignored and semantic information from the high level cannot be transferred to the low level. This paper proposes a new feature fusion mechanism that combines the feature fusion techniques of FPN and FSSD and uses top-down, bottom-up, and lateral connection methods to construct a new feature pyramid through two forward transfers and one reverse transfers. Feature visualization and ablation experiments prove the effectiveness of the proposed approach. The combination of FPN and FSSD improves the accuracy of prostate capsule detection and is able to extract more abundant features. In the later stage, the accuracy of the dataset with noise reduction is further improved. The reason for this is that during training, the pre-training weights which had been trained 50,000 times by PCA+FFSSD were used for transfer learning; thus, the accuracy of the model was further improved after fine-tuning the noise removal. In terms of detection, the multi-level network mixed method can improve the detection precision of the network. Next, we intend to try to find a method that can achieve a good balance of speed and precision. Typical object detection algorithms such as YOLOv4, Sparse R-CNN, TOOD, and FoveaBox fail to produce excellent performance in this case due to the unique deformation and multi-scale characteristics of the prostatic envelope and the sparsity of the datasets.

Author Contributions

Conceptualization, X.W.; methodology, S.W.; validation, S.W. and X.W.; formal analysis, S.W.; investigation, C.G.; resources, X.W.; data curation, X.W.; writing—original draft preparation, S.W.; writing—review and editing, X.W.; visualization, S.W.; supervision, C.G.; project administration, S.W.; funding acquisition, S.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Scientific Research Project of Hubei Education Department, grant number B202287. Financial and technical support was provided by National Natural Science Foundation of China (grant number 2016YFC0106301)and the establishment of “5G+ Artificial intelligence” remote treatment and diagnosis platform for major aortic diseases (grant number 2022BCA035).

Institutional Review Board Statement

This study was approved by the Institutional Review Board Ethics Committee of Zhongnan Hospital of Wuhan University (approval number 2015029, project number 2016YFC0106301).

Data Availability Statement

The data presented in this study are available on request from the corresponding authors. The data are not publicly available due to the patients’ privacy protection.

Acknowledgments

This paper was supported by the Scientific Research Project of Hubei Education Department (Detection of Medical Images Based on Feature Fusion Hierarchy Deep Network, grant number B202287).Financial and technical support was provided by National Natural Science Foundation of China (grant number 2016YFC0106301)and the establishment of “5G+ Artificial intelligence” remote treatment and diagnosis platform for major aortic diseases (grant number 2022BCA035).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Papageorgiou, C.; Poggio, T. A Trainable System for Object Detection. Int. J. Comput. Vis. 2000, 38, 15–33. [Google Scholar] [CrossRef]
  2. Viola, P.A.; Jones, M.J. Rapid Object Detection using a Boosted Cascade of Simple Features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, Kauai, HI, USA, 8–14 December 2001. [Google Scholar]
  3. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
  4. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  5. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Computer Vision—ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2016. [Google Scholar]
  6. Li, Z.; Zhou, F. FSSD: Feature Fusion Single Shot Multibox Detector. arXiv 2017, arXiv:1712.00960. [Google Scholar]
  7. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  8. Liu, J.; Cao, L.; Akin, O.; Tian, Y. 3DFPN-HS2: 3D Feature Pyramid Network Based High Sensitivity and Specificity Pulmonary Nodule Detection. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 513–521. [Google Scholar]
  9. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C. Sparse R-CNN: End-to-End Object Detection With Learnable Proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  10. Bochkovskiy, A.; Wang, C.Y.; Liao, H.y. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  11. Kong, T.; Sun, F.; Liu, H.; Jiang, Y.; Shi, J. FoveaBox: Beyound Anchor-Based Object Detection. IEEE Trans. Image Process. 2020, 29, 7398. [Google Scholar] [CrossRef]
  12. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. TOOD: Task-aligned One-stage Object Detection. arXiv 2021, arXiv:2108.07755. [Google Scholar]
  13. Joseph, K.; Khan, S.; Khan, F.; Balasubramanian, V. Towards Open World Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  14. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. arXiv 2019, arXiv:1911.09070. [Google Scholar]
  15. Liu, Z.; Gao, G.; Sun, L. IPG-Net: Image Pyramid Guidance Network for Object Detection. arXiv 2019, arXiv:1912.00632. [Google Scholar]
  16. Bakkouri, I.; Karim, A. Computer-aided diagnosis (CAD) system based on multi-layer feature fusion network for skin lesion recognition in dermoscopy images. Multimed. Tools Appl. 2020, 79, 20483–20518. [Google Scholar] [CrossRef]
  17. Haekal, M.; Septiawan, R.R.; Haryanto, F.; Arif, I. A comparison on the use of Perlin-noise and Gaussian noise based augmentation on X-ray classification of lung cancer patient. J. Phys. Conf. Ser. 2021, 1951, 012064. [Google Scholar] [CrossRef]
  18. Kong, L.; Cheng, J. Classification and detection of COVID-19 X-Ray images based on DenseNet and VGG16 feature fusion. Biomed. Signal Process. Control. 2022, 77, 103772. [Google Scholar] [CrossRef]
  19. Zia, H.; Khatami, F.; Aghamir, S. U-shape incision on prostate capsule: New intraperitoneal laparoscopic technique in simple prostatectomy: A case report. Ann. Med. Surg. 2021, 69, 102787. [Google Scholar] [CrossRef]
  20. Weizer, A.; Shah, R.; Lee, C.; Gilbert, S.; Daignault, S.; Montie, J.; Wood, D. Evaluation of the prostate peripheral zone/capsule in patients undergoing radical cystoprostatectomy: Defining risk with prostate capsule sparing cystectomy. Urol. Oncol. 2007, 25, 460–464. [Google Scholar] [CrossRef]
  21. Hussein, R.Y. An Adaptive Algorithm to Identify Ambiguous Prostate Capsule Boundary Lines for Three-Dimensional Reconstruction and Quantitation. Ph.D. Thesis, Old Dominion University, Norfolk, VA, USA, 2006. [Google Scholar]
  22. Holder, K.G.; Galvan, B.; Knight, A.S.; Ha, F.; Riese, W. Possible clinical implications of prostate capsule thickness and glandular epithelial cell density in benign prostate hyperplasia. Investig. Clin. Urol. 2021, 62, 423. [Google Scholar] [CrossRef]
  23. Lim, J.S. Two-Dimensional Signal and Image Processing; Prentice-Hall, Inc.: Upper Saddle River, NJ, USA, 1990. [Google Scholar]
  24. Hwang, H.; Haddad, R.A. Adaptive median filters: New algorithms and results. IEEE Trans. Image Process. 1995, 4, 499–502. [Google Scholar] [CrossRef] [Green Version]
  25. Chen, T.; Wu, H.R. Adaptive impulse detection using center-weighted median filters. IEEE Signal Process Lett. 2001, 8, 1–3. [Google Scholar] [CrossRef]
  26. Xu, G.; Tan, J. A Universal Impulse Noise Filter with an Impulse Detector and Nonlocal Means. Circuits Syst. Signal Process. 2014, 33, 421–435. [Google Scholar] [CrossRef]
  27. Lei, H.; Tan, J.; Gan, K. Non-local Means Image Denoising Algorithm Based on Edge Detection. In Proceedings of the 2014 5th International Conference on Digital Home (ICDH), Guangzhou, China, 28–30 November 2014; pp. 117–121. [Google Scholar]
  28. Zhang, H.; Zhu, Y.; Zheng, H. NAMF: A Nonlocal Adaptive Mean Filter for Removal of Salt-and-Pepper Noise. Math. Probl. Eng. 2021, 2021, 4127679. [Google Scholar] [CrossRef]
  29. Thanh, D.N.H.; Thanh, L.T.; Hien, N.N.; Prasath, S. Adaptive total variation L1 regularization for salt and pepper image denoising. Optik 2020, 208, 163677. [Google Scholar] [CrossRef]
  30. Pang, S.; Ding, T.; Qiao, S.; Meng, F.; Wang, S.; Li, P.; Wang, X. A novel YOLOv3-arch model for identifying cholelithiasis and classifying gallstones on CT images. PLoS ONE 2019, 14, e0217647. [Google Scholar] [CrossRef]
  31. Turan, B. ANN Based Removal for Salt and Pepper Noise. In Proceedings of the Global Conference on Engineering Research, Online, 2–5 June 2021; pp. 558–566. [Google Scholar]
Figure 1. CF results for denoising, compared with fourteen different methods.
Figure 1. CF results for denoising, compared with fourteen different methods.
Electronics 12 01060 g001
Figure 2. Edge detection comparision.
Figure 2. Edge detection comparision.
Electronics 12 01060 g002
Figure 3. The prostate capsule has the characteristics of large-scale changes and different shapes.
Figure 3. The prostate capsule has the characteristics of large-scale changes and different shapes.
Electronics 12 01060 g003
Figure 4. First feature forward propagation.
Figure 4. First feature forward propagation.
Electronics 12 01060 g004
Figure 5. Reverse feature propagation.
Figure 5. Reverse feature propagation.
Electronics 12 01060 g005
Figure 6. Reverse feature transfer.
Figure 6. Reverse feature transfer.
Electronics 12 01060 g006
Figure 7. Multi-level feature pyramid network in FFSSD.
Figure 7. Multi-level feature pyramid network in FFSSD.
Electronics 12 01060 g007
Figure 8. FFSSD Network.
Figure 8. FFSSD Network.
Electronics 12 01060 g008
Figure 9. The mAP curve transformation during SSD, FSSD, and PCA+FFSSD training (a total of 5000 training instances).
Figure 9. The mAP curve transformation during SSD, FSSD, and PCA+FFSSD training (a total of 5000 training instances).
Electronics 12 01060 g009
Figure 10. Loss curve transformation during SSD, FSSD, and FFSSD training (a total of 5000 training times).
Figure 10. Loss curve transformation during SSD, FSSD, and FFSSD training (a total of 5000 training times).
Electronics 12 01060 g010
Figure 11. Feature visualization comparison of conv3_3, conv4_3, conv5_3, and fc7 features of the convolution layer of FSSD and FFSSD.
Figure 11. Feature visualization comparison of conv3_3, conv4_3, conv5_3, and fc7 features of the convolution layer of FSSD and FFSSD.
Electronics 12 01060 g011
Figure 12. Detection result comparison for SSD, FSSD, and FFSSD.
Figure 12. Detection result comparison for SSD, FSSD, and FFSSD.
Electronics 12 01060 g012
Figure 13. Comparison of mAP for ten training epochs.
Figure 13. Comparison of mAP for ten training epochs.
Electronics 12 01060 g013
Figure 14. Detection result comparison for PCA+CF+FFSSD, TOOD, Sparse R-CNN, and FoveaBox.
Figure 14. Detection result comparison for PCA+CF+FFSSD, TOOD, Sparse R-CNN, and FoveaBox.
Electronics 12 01060 g014
Table 1. Speed and Precision comparison on NVIDIA 1070.
Table 1. Speed and Precision comparison on NVIDIA 1070.
ModelBackbone NetworkmAPimg/s
SSDVGG1671.90%0.027
FSSDVGG1673.82%0.037
FSSD+FPNVGG1675.45%0.046
PCA+FFSSDVGG1682.39%0.046
Table 2. Comparison of various methods.
Table 2. Comparison of various methods.
ModelBackbone NetworkmAPFPS
Faster R-CNN [7]VGG1662.67%5 (K40)
Faster R-CNN [7]ResNet 50+FPN74.41%-
PCA+CF+Faster R-CNN (ours)ResNet 50+FPN77.10%-
SSD [5]VGG1671.90%46
SSD [5]ResNet-101 [5]74.39%15 (NVIDIA GTX 1070)
PCA+CF+SSD (ours)VGG1677.30%46
EfficientDet-D0 [14]B053.38%97 (Telsa v100)
EfficientDet-D1 [14]B156.58%74 (Telsa v100)
EfficientDet-D2 [14]B259.23%57 (Telsa v100)
EfficientDet-D3 [14]B361.14%35 (Telsa v100)
EfficientDet-D4 [14]B458.81%23 (Telsa v100)
EfficientDet-D5 [14]B558.09%10 (Telsa v100)
EfficientDet-D7 [14]B678.37%
FSSD [6]VGG1673.82%65.8 (NVIDIA 1080Ti)
PCA+CF+FSSD (ours)VGG1675.58%65.8 (NVIDIA 1080Ti)
FoveaBox [11]ResNet50+FPN81.10%25 (NVIDIA RTX 2060)
TOOD [12]ResNet50+FPN73.08%20 (NVIDIA RTX 2060)
PCA+CF+TOOD (ours)ResNet50+FPN76.50%20 (NVIDIA RTX 2060)
YOLOv4 [10]CSPDarknet-5370.29%45 (NVIDIA RTX 2060)
PCA+CF+YOLOv4 (ours)CSPDarknet-5372.47%45 (NVIDIA RTX 2060)
Sparse R-CNN [9]ResNet50+FPN75.68%17 (NVIDIA RTX 2060)
OWOD [13]ResNet-5071.30%62 (NVIDIA RTX 2060)
PCA+CF+FFSSD (ours)VGG1683.58%27 (NVIDIA GTX 1070)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, S.; Wang, X.; Guo, C. Application of Feature Pyramid Network and Feature Fusion Single Shot Multibox Detector for Real-Time Prostate Capsule Detection. Electronics 2023, 12, 1060. https://doi.org/10.3390/electronics12041060

AMA Style

Wu S, Wang X, Guo C. Application of Feature Pyramid Network and Feature Fusion Single Shot Multibox Detector for Real-Time Prostate Capsule Detection. Electronics. 2023; 12(4):1060. https://doi.org/10.3390/electronics12041060

Chicago/Turabian Style

Wu, Shixiao, Xinghuan Wang, and Chengcheng Guo. 2023. "Application of Feature Pyramid Network and Feature Fusion Single Shot Multibox Detector for Real-Time Prostate Capsule Detection" Electronics 12, no. 4: 1060. https://doi.org/10.3390/electronics12041060

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop