Next Article in Journal
Effects of a Subanesthetic Ketamine Infusion on Inflammatory and Behavioral Outcomes after Closed Head Injury in Rats
Next Article in Special Issue
A Synthesizing Semantic Characteristics Lung Nodules Classification Method Based on 3D Convolutional Neural Network
Previous Article in Journal
Three-Dimensional Preoperative Planning Software for Hip Resurfacing Arthroplasty
Previous Article in Special Issue
Semi-Supervised Medical Image Segmentation with Co-Distribution Alignment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Soft-Reference Breast Ultrasound Image Quality Assessment Method That Considers the Local Lesion Area

1
School of Mechanical Engineering and Automation, Harbin Institute of Technology, Shenzhen 518055, China
2
Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
3
The Second School of Clinical Medicine, Southern Medical University, Guangzhou 510515, China
4
Department of Ultrasound, The Third Medical Centre of Chinese PLA General Hospital, Beijing 100039, China
*
Authors to whom correspondence should be addressed.
Bioengineering 2023, 10(8), 940; https://doi.org/10.3390/bioengineering10080940
Submission received: 30 June 2023 / Revised: 23 July 2023 / Accepted: 3 August 2023 / Published: 7 August 2023
(This article belongs to the Special Issue Recent Advance of Machine Learning in Biomedical Image Analysis)

Abstract

:
The quality of breast ultrasound images has a significant impact on the accuracy of disease diagnosis. Existing image quality assessment (IQA) methods usually use pixel-level feature statistical methods or end-to-end deep learning methods, which focus on the global image quality but ignore the image quality of the lesion region. However, in clinical practice, doctors’ evaluation of ultrasound image quality relies more on the local area of the lesion, which determines the diagnostic value of ultrasound images. In this study, a global–local integrated IQA framework for breast ultrasound images was proposed to learn doctors’ clinical evaluation standards. In this study, 1285 breast ultrasound images were collected and scored by experienced doctors. After being classified as either images with lesions or images without lesions, they were evaluated using soft-reference IQA or bilinear CNN IQA, respectively. Experiments showed that for ultrasound images with lesions, our proposed soft-reference IQA achieved PLCC 0.8418 with doctors’ annotation, while the existing end-to-end deep learning method that did not consider the local lesion features only achieved PLCC 0.6606. Due to the accuracy improvement for the images with lesions, our proposed global–local integrated IQA framework had better performance in the IQA task than the existing end-to-end deep learning method, with PLCC improving from 0.8306 to 0.8851.

1. Introduction

Breast disease is the most serious problem for women’s health today and has become a serious global public health problem [1]. Studies have shown the importance of early screening for the prognosis and treatment of breast cancer [2]. Ultrasound (US) examination is widely used for the diagnosis of internal organs and superficial organs due to its non-invasiveness, non-radiation use and convenience, compared with CT, MRI or X-ray [3]. In particular, it has been proved to be especially effective in the diagnosis of breast cancer [4]. Meanwhile, ultrasound examinations rely on clinical experience, which affects the quality of ultrasound images and the accuracy of diagnosis. Therefore, objective quality evaluation of ultrasound images has clinical significance to guarantee the standardization of ultrasound examination. In addition, to achieve automation of ultrasound examination, the ultrasound scanning robot has emerged as a research trend. In our previous research, an autonomous ultrasound scanning robot was designed [5]. The quality assessment of ultrasound images is considered as a visual servo that can be used as control feedback for autonomous scanning robots. There are many reasons for the poor quality of ultrasound images, such as noise in signal transmission, artifacts, low echoes due to poor contact between the probe and the skin, and dragging blurring due to probe movement. To address these issues, this study focuses on the quality assessment of breast ultrasound images.
Image quality assessment has been intensively studied in the field of nature images. Generally, IQA is divided into subjective assessment, which is judged by experts, and objective assessment, which is computed with mathematical algorithms. Subjective IQA can be time-consuming, labor-intensive, costly, error-prone, and inconsistent [6]. The objective assessment can be further divided into three categories based on the availability of reference images: (i) full reference (FR-IQA), where there is a perfect reference image for comparison with the test image; (ii) reduced reference (RR-IQA), which contains partial information of the reference image, and (iii) no reference (NR-IQA), where no information regarding the reference image is available for the assessment [7]. FR-IQA calculates the statistical difference between the sample image and the reference image, such as PSNR [8], RMSE [9], intensity [10] or SSIM [11]. However, FR-IQA and RR-IQA have limited practical applications since the reference image is usually not available, especially for medical images. In contrast, NR-IQA methods do not require any information from a reference image and are well-suited for practical applications.
Many researchers have focused on the quality assessment of medical images, including MRI [6,12,13], retinal fundus images [14], video endoscopy [15,16], CT [17], and ultrasound images. Medical ultrasound images are known to present with a lot of noise and artifacts, low resolution, blurred boundaries, and low contrast, which bring many difficulties to the quality assessment of images. Traditional NF-IQA methods for ultrasound images generally extract image features first and then adopt SVM, SVR, and other methods with machine learning for quality regression or classification [13]. In recent years, some researchers have introduced confidence mapping into autonomous robotic scanning tasks to evaluate the quality of ultrasound images. Confidence mapping is a method to quantitatively evaluate the uncertainty caused by the attenuation and shadow in ultrasound images [18,19]. However, the confidence map only evaluates the contact condition between the probe and human skin, indicating the reliability of the ultrasound imaging at each pixel, and does not consider the significance of the image content for diagnosis. Therefore, the confidence mapping method has a big gap between it and the clinical standard of ultrasound image quality assessment. Recently, end-to-end deep learning methods have been demonstrated to have great potential for IQA. Most researchers working on clinical ultrasound image analysis have proposed to assess the image quality based on whether key anatomical structures can be detected. Wu et al. [20] proposed a computer-assisted fetal ultrasound image quality assessment algorithm, which is composed of L-CNN and C-CNN. The function of L-CNN is to find the ROI (Region of Interest) in the fetal abdominal region of the ultrasound image, and C-CNN detects whether there are key structures, such as gastric vesicles and umbilical veins, in the obtained ROI to evaluate the quality of the ultrasound image. Lin et al. [21] proposed a MF R-CNN (Multi-task Fast Regional Convolutional Neural Network) for the quality assessments of fetal head ultrasound images, which can detect and identify the key anatomical structures of the fetal head, and then evaluate the quality of ultrasound images. Camps et al. [22] proposed a DenseNet classification to automatically assess the quality of trans-perineal ultrasound images of the male pelvic region based on the visibility of the prostate. Similarly, Antico et al. [23] trained a DenseNet to classify knee ultrasound images into two categories based on the visibility of cartilage. In breast ultrasound images, the anatomical structures such as the subcutaneous fat layer, breast glands, pleura and other anatomical structures are not clear, and the breast morphology of different patients is very different due to individuality. Therefore, it is difficult to assess the quality of breast ultrasound images through the extraction and detection of specific anatomical structures.
There are some other studies which have converted quality assessments to classification or regression by using doctors’ subjective evaluations of samples as labels. In Zhang’s work, ultrasound images were artificially added with noise, subjectively scored, and regressed with DCNN [24]. In our previous work, doctors classified breast ultrasound images into five levels according to subjective evaluation, and a BCNN assessment network was trained [25]. These works conducted global quality assessment of ultrasound images; however, for medical ultrasound images, doctors usually pay more attention to image content, especially those images with lesions. This research is an extension of our previous work [25], and a soft-reference-based global–local hybrid breast ultrasound IQA framework is proposed. The breast ultrasound images were first classified into two groups that were either with or without lesions, then for the ultrasound images without lesions, the global image quality was calculated, and for the ultrasound images with lesions, the lesion was segmented and the image quality of the local region was evaluated.
The main contributions of this paper are listed as follows:
  • An integrated global–local breast ultrasound IQA framework is established, in which the images without lesions are evaluated globally, and the images with lesions are evaluated locally (i.e., with a focus on the lesion area).
  • A soft-reference IQA method based on lesion segmentation is proposed, in which the segmented lesion is taken as a reference image and transforms the medical ultrasound image assessment problem, from no reference to reduced reference.
  • A comprehensive evaluation index for medical ultrasound images with lesions is proposed, including pixel-level features MSE (Mean Squared Error), PSNR (Peak Signal-to-Noise Ratio) and semantic-level features CR (Contrast Ratio) and SSIM (Structural Similarity).
The rest of this paper is organized as follows. Section 2 describes the dataset construction, which includes the image collection and data labeling and the IQA framework. Section 3 presents the implement details of the network and the results of the experiments. Finally, the conclusion of this study is presented in Section 4.

2. Materials and Methods

2.1. Data Collection

The ultrasound images used in this paper were collected by five doctors from the ultrasound department of PLA General Hospital, including ultrasound images with and without lesions. The lesions include benign cancers, nodules or cysts (fibroadenoma or fibrocystic breast disease). All the doctors participating in this study had more than five years of clinical experience. The collection period ranged from August 2020 to April 2021. The ultrasound examination equipment used was Siemens S2000, and the probe used was 9L4. Starting with the default parameter setting for breast examination, different doctors adjusted these parameters according to their habits and collected breast images. The collected images were captured by an image acquisition card and saved in PNG format. The images of healthy breasts and images with lesions were used in this research, and the images with missing breasts after mastectomy were excluded. Finally, there was a total of 1285 ultrasound images collected from 524 female patients, including 844 healthy breast ultrasound images (A) and 441 (B) breast ultrasound images with lesions. The age of patients ranged from 18 to 86 years, with an average of 49.5 years. The data collection was approved by the Institutional Review Board under YSB-2020-Y0902. The datasets A and B were divided randomly into the training set (A.1 and B.1), validation set (A.2 and B.2), and test set (A.3 and B.3) according to the ratio of 6:2:2, respectively.

2.2. Data Labeling

To improve the credibility of the data labeling, three other doctors with more than ten years of ultrasound examination experience were invited to conduct the image quality assessment. The score ranged from 1 to 4, with 4 indicating excellent image quality that can be used for diagnosis, with 3 indicating defective image quality but which does not affect the diagnosis, with 2 indicating poor image quality which affects the diagnosis, and with 1 indicating unacceptable image quality and cannot be used for diagnosis. The scoring is based on doctors’ clinical evaluation standards including the amount of noise, the definition, the contrast, and the vision field of the images. The final score of each image was the averaged value from different doctors and rounded to an integer. The numbers of breast ultrasound images with their different quality scores are shown in Table 1. Additionally, the 441 breast ultrasound images with lesions were manually segmented.

2.3. Proposed Framework for Breast Ultrasound Image Quality Assessment

Most of the existing breast ultrasound IQA methods focus on the global image quality, but ignore the image quality of the lesion area, which determines the diagnostic value of ultrasound images. Hence, in this section, a global–local integrated quality assessment framework for breast ultrasound images is presented. Figure 1 provides an overview of the process. The proposed IQA system has three major modules: firstly, a binary classification of breast ultrasound images is implemented to determine whether the input is negative (without lesions) or positive (with lesions), using a residual network containing an attention module. Secondly, for breast ultrasound images without lesions, a global image quality evaluation network based on BCNN [25] is used. Thirdly, for breast ultrasound images with lesions, a local IQA method based on breast lesion segmentation is proposed. For the local image quality assessment, the result of lesion segmentation is used as a mask to compare with the local lesion area of the breast ultrasound image, then IQA metrics are calculated to measure the pixel-level differences and structural differences between them; finally, these metrics are inputted to an SVM (Support Vector Machine) to obtain the image quality of the local lesion area. By taking the segmented lesion as a reference, the traditional medical ultrasound image assessment problem is transformed from no reference to reduced reference, so we define this method as soft reference (SR) in this research. Meanwhile, positive samples and negative samples were evaluated using different models to improve the diagnosability of ultrasound images in clinical applications.

2.4. Classification Network

ResNet is one of the most commonly utilized networks in image classification, whose main contribution is the residual learning module, which solves the network degradation problem caused by the gradient disappearance of CNN [26]. Considering the real-time performance of the network and the size of the existing training data, ResNet18 was selected as the basic component of the feature extractor in this research. In addition, the lesion identification of breast ultrasound images mainly relies on local features. Hence, to further improve the model performance, we introduce the convolutional block attention module (CBAM) [27] into our network to select the features containing more information and suppress less useful features. In this research, the breast lesion detection network (CBAM-ResNet18) contains 8 CBAM-Res modules inserted into the Residual Block, as shown in Figure 2a,c. CBAM includes a spatial attention module and a channel attention module. The spatial attention module can find out the areas that need to be paid attention to in the image. As shown in Figure 2, global average pooling (GAP) and global maximum pooling (GMP) are performed along the channel direction for the input feature matrix  F H W C , followed by a 7 × 7 convolution and activation function. The spatial attention vector can be expressed as Equation (1).
M s ( F ) = σ { f 7 × 7 [ A v g P o o l ( F ) ; M a x P o o l ( F ) ] }
where  σ  represents the activation function ReLu. The channel attention module can find out the features that need to be paid attention to in the image. GAP and GMP are performed by each feature plane, followed by a shared multilayer perception machine. The channel attention vector can be expressed as Equation (2).
M c ( F ) = σ { M L P [ A v g P o o l ( F ) ] + M L P [ M a x P o o l ( F ) ] }

2.5. BCNN Global IQA Network

The BCNN architecture has good performance in fine grained image classification task and is introduced in this work to improve the accuracy of breast ultrasound image quality assessment. The application scenario of fine-grained classification network is to distinguish objects from different subclasses within a general category such as identifying the species of a bird or the model of an aircraft. It is with large intra-class variance and small inter-class variance affected by the pose, background, and viewpoint [28]. These characteristics are similar to the quality assessment of medical ultrasound images. The difference between ultrasound images of different qualities is not significant, while ultrasound images of the same image quality may have great differences due to the location and angle of image acquisition. The BCNN architecture uses a quadratic linear pooling operation to combine the features extracted by the feature extractor A and feature extractor B, which comprise ResNet18, as shown in Figure 1. The quadratic linear pooling operation involves carrying out the outer product operation of the features extracted by the two networks. To reduce the number of parameters, we make feature extractor A and feature extractor B share parameters. Only one network model needs to be trained, which shortens the training time.

2.6. Lesion Segmentation Network

U-Net is the most successful network structure in the field of medical image segmentation [29]. It uses the classical encoder-decoder structure, consisting of compressed and extended paths. Because the lesion detection task in 2.4 has similar feature extraction process with the lesion segmentation task in this section, the encoder of U-Net is replaced by CBAM-ResNet18 to extract more effective features. The decoder of U-Net is replaced by the basic structure of ResNet18. The encoder uses the lesion detection network in 2.4 as the pre-trained model, and the decoder uses the model pre-trained on ImageNet. The modified U-Net network is named as U-Res-CBAM in this study, as shown in Figure 2b.

2.7. Parameters of Local Image Quality Assessment

As shown in Figure 3, for ultrasound images with lesions, we first segment the lesions using a deep learning-based method and an image processing-based method separately. If the local quality of the image is good, then the masks obtained by the two methods should be closer. The local quality can be reflected by calculating the difference between the masks obtained by the two methods. In this way, we convert the no-reference method into a soft-reference method. In this subsection, four parameters are adopted to calculate the image quality of the local lesion region in breast ultrasound images, at both the pixel level and the semantic level.

2.7.1. Mean Squared Error (MSE)

MSE is a common metric used in image processing to evaluate the mean value of the difference between two images, which can be expressed by Equation (3).
MSE = 1 M N x = 0 M 1 y = 0 N 1 [ A ( x , y ) B ( x , y ) ] 2
where A(x, y) and B(x, y) denote the grayscale values of pixels (x, y) in image A and image B, respectively. M and N denote the number of pixels in the length direction and width direction, respectively.

2.7.2. Peak Signal-to-Noise Ratio (PSNR)

PSNR measures the ratio of the intensity of the peak signal to the average intensity of the noise, and it is changed to decibels (dB) by taking the logarithm of the value, as expressed by Equation (4).
PSNR = 10 l o g 10 M a x V a l u e 2 M S E
where MaxValue indicates the maximum grayscale value.

2.7.3. Contrast Ratio (CR)

CR indicates the degree of distinguishability between the lesion region and the surrounding region of the image. The lesion region in the breast ultrasound image is usually darker than the surrounding region, since the echo of the lesion area is weaker than that of healthy tissue area, presenting a low echo state. The greater the difference in image gray value between the lesion area and the surrounding area, the easier the lesion region is to distinguish, which indicates that the image quality is high. The breast ultrasound image is firstly segmented to obtain its foreground (the lesion region) and background (the surrounding region), and the parameter CR is calculated by Equation (5).
CR = ( M e a n F M e a n B ) 255
where MeanF and MeanB represent the average grayscale value of the foreground and background, respectively.

2.7.4. Structural Similarity (SSIM)

SSIM is adopted to measure the similarity of structure between the lesion region and the reference image mask. In this study, SSIM is defined as the Dice between the segmentation results obtained using the OTSU threshold segmentation algorithm [30] and the reference image mask obtained with the segmentation network described in Section 2.6. Figure 4 shows the lesion segmentation results of high-quality and low-quality breast ultrasound images with the OTSU algorithm and our segmentation network U-Res-CBAM. It is observed that the OTSU algorithm has better segmentation result for a high-quality image, which is with a clear edge, a complete boundary and an obvious difference between the lesion region and the surrounding region. For a low-quality image with a blurred edge, the segmentation result of the OTSU algorithm is poor. Therefore, the closer the segmentation results between those obtained using OSTU algorithm and the reference image mask obtained with the U-Res-CBAM network, the higher the image quality. The parameter SSIM can be calculated by Equation (6).
Dice = 2 | A B | | A | + | B |
where A and B represent binary segmented images using the OSTU algorithm and the U-Res-CBAM network.

2.8. Support Vector Machine (SVM)

To create a quality model based on the above pixel-level features and semantic-level features, an optimizable SVM classifier is used. Firstly, these features are normalized into [0, 1]. Then the SVM maps the feature vector Fi = {f1, f2, f3, f4} of the images into a subjective score classes space {1, 2, 3, 4}. With training dataset {(F1, q1), (F2, q2), …, (FN, qN)}, the SVM classifier tries to find a hyperplane function that separates classes based on these features, as expressed by Equation (7).
f = ω T ϕ ( F ) + b = i = 1 N α i q i κ ( F , F i ) + b
where ω is the weight vector, b is a bias parameter,  α i  is the combination coefficient,  κ ( F , F i ) = exp ( γ F i F 2 )  is a Gaussian kernel function, and γ denotes a precision parameter. The determination of ω and b can be modeled as an optimization problem, as expressed by Equation (8).
min ω , b , ξ i 1 2 ω 2 + C i = 1 N ξ i s . t . { q i ( ω T ϕ ( F ) + b ) 1 ξ i ξ i 0 , i = 1 , 2 , , N
where  ξ i  is the slack variable and C is a penalty parameter. In this study, the parameters C and γ of the SVM model are determined using a Bayesian optimization search method [31], tuned by four-fold cross-validation, as shown in Figure 3. Once the hyperparameters are determined, the SVM model is then retrained with the whole training data of breast ultrasound images with lesions. Then, for each testing image, the SVM-based model is used for quality prediction.

3. Results

3.1. Implementation Details

In this section, we conduct experiments to evaluate the efficacy of the proposed US image quality assessment framework, including classification, global IQA, segmentation, and local IQA. Finally, the proposed approach is compared with our previous US image quality assessment methods. In this study, we used the open-source framework Pytorch for network implementation. The version of Pytorch was 1.7 and the version of Python was 3.7. All the experiments were performed on a workstation with NVIDIA GTX3090 (Intel i7-9700K, 32 G RAM, Windows 10).

3.2. Quantitative Evaluation Metrics

In order to quantitatively measure the correlation between the output results of quality assessment network and doctors’ ground-truth manual annotations, four quantitative evaluation metrics are used as the evaluation metrics, as shown in Section 3.2.1, Section 3.2.2, Section 3.2.3 and Section 3.2.4.

3.2.1. Pearson Linear Correlation Coefficient (PLCC)

PLCC is adopted to evaluate the accuracy of the IQA model [7], which can be calculated with Equation (9).
PLCC = i = 1 N ( X i X ¯ ) ( Y i Y ¯ ) i = 1 N ( X i X ¯ ) 2 i = 1 N ( Y i Y ¯ ) 2
where  X i  and  Y i  represent the score of doctors’ manual annotation and the score output by the network for the i-th image, N represents the number of images included in the comparison, and  X ¯  and  Y ¯  represent the averaged values for these images. The value of PLCC ranges from −1 to 1. If the absolute value of the PLCC value is close to 1, it indicates that the output results of quality assessment network and doctors’ ground-truth manual annotations have strong correlation.

3.2.2. Spearman Rank-Order Correlation Coefficient (SRCC)

SRCC is adopted to evaluate the monotonicity of the IQA model [7], which can be calculated with Equation (10).
SRCC = 1 6 i = 1 N d i 2 N ( N 2 1 )
where di denotes the difference between the subjective quality score ranking and the objective quality score ranking of the i-th image. The value of SRCC ranges from 0 to 1. If the absolute value of SRCC is close to 1, it indicates that the two sets of data are completely consistent.

3.2.3. Accuracy

Accuracy indicates the percentage of samples with completely correct predictions.

3.2.4. Root Mean Square Error (RMSE)

RMSE is used to evaluate the consistency of the IQA model with the labels, as shown in Equation (11). si and pi denote the subjective quality score and objective quality score of the i-th image, respectively.
R M S E = 1 n i = 1 N ( s i p i ) 2

3.3. Classification Experiment

In this sub-experiment, 1285 breast ultrasound images (dataset A and B) were used. The model was trained for 150 epochs with a batch size of 16. The Adam optimizer was adopted with an initial learning rate of 1 × 10−4. CrossEntropyLoss was chosen as the loss function, and the classification results were evaluated in term of accuracy. The performance of the model was tested on the validation set (A.2 and B.2) every four epochs, and the model with the best validation accuracy was retained and evaluated on the test set (A.3 and B.3). ResNet18 was initialized using a pre-trained model on ImageNet. To show the effectiveness of our proposed method, three groups of classification experiments were conducted. The first experiment was with the classification model ResNet18 without pre-training, the second experiment was with the classification model ResNet18 with pre-training, and the third experiment was with our proposed classification model CBAM-ResNet18 with pre-training. As shown in Table 2, the classification accuracy of our proposed model was 99.2% on the test set, which is the best among the three methods. Figure 5 shows the heat map of the classification network computed by Grad-CAM [32]. For ultrasound images with lesions, the neural network tends to focus on the lesion, while for images without lesions, the neural network does not have a clear focus.

3.4. Global IQA Experiment

For breast ultrasound images without lesions (dataset A), the global IQA network described in Section 3.3 was used. The model was trained for 30 epochs with a batch size of 24. The Adam optimizer was adopted with an initial learning rate of 1 × 10−4. MSE was chosen as the loss function. We used ResNet18 pre-trained in ImageNet as the initial state of the feature extractor. The validation set was used to optimize hyperparameters and retain the best model on a certain epoch. Table 3 shows that the global IQA method achieves good correlation with the doctor’s subjective evaluation for images without lesion.

3.5. Segmentation Experiment

In this subsection, we briefly describe the experiment of breast lesion segmentation, which is a necessary step for local IQA. A total of 441 breast ultrasound images with lesions (dataset B) were used. The model was trained for 100 epochs with a batch size of 16. The Adam optimizer was adopted with an initial learning rate of 1 × 10−4. Dice was chosen as the loss function. To show the effectiveness of our proposed segmentation network, contrast experiments were conducted to compare with other state-of-the-art segmentation models including VGG16 [32], PSPNet [33], DeepLab V3+ [34], U-Res [35], and ViT [36]. The initial weights of the encoder used the weights of the classification network in Section 3.3. As shown in Table 4, the U-Res-CBAM network achieved the best segmentation results, with Dice 0.9272 on the validation set and Dice 0.9126 on the test set. Figure 6 shows the lesion segmentation results of the ultrasound images with different segmentation networks.

3.6. Local IQA Experiment

For 441 breast ultrasound images with lesions (dataset B), the local soft-reference image quality assessment (SF-IQA) method was adopted. Firstly, four local IQA parameters were calculated. Then the values were normalized to [0, 1]. We took 1 minus the normalized result of MSE, so that all the four quality evaluation parameters were positively correlated with the quality degree of the images. A box-plot of the four parameters with respect to the different quality scores is shown in Figure 7. It is noted that the mean value of each parameter increases monotonically with image quality, which confirms the rationality and feasibility of adopting these four parameters to quantify the image quality. A fourfold cross-validation method was used to find the best SVM parameters and the test set was used to evaluate the performance of the local SF-IQA model. The results using the mask obtained with different lesion segmentation methods as reference images are shown in Table 5. For the U-Res-CBAM segmentation method used in this paper, the best performance was achieved on the test set, with PLCC 0.8418, SRCC 0.8462, Accuracy 0.8068, and RMSE 0.4395. Furthermore, the PLCC between the SVM IQA model and the doctor’s subjective quality evaluation can reach 0.9255 when the reference images of the four quality evaluation metrics are replaced with Ground-Truth manually segmented by the doctor. It is obvious that lesion segmentation accuracy has an important impact on our IQA method. The higher the segmentation accuracy, the better the consistency between the soft-reference IQA and the subjective evaluation of the physician.

3.7. Global–Local Integrated IQA Experiment

To evaluate the proposed global–local integrated IQA framework, a contrast experiment was conducted to compare with our previous work which applied only a global BCNN network to assess ultrasound images with and without lesions. For our proposed global–local integrated IQA framework, the global component and local component were trained with dataset A.1 and dataset B.1 separately, and the best hyperparameters were determined with validation datasets A.2 and B.2 separately, which were introduced in Section 3.4. and Section 3.6. Then the test images (A.3 + B.3) were classified and assessed. Meanwhile, for the global BCNN network without consideration of lesions [25], training took place with mixed dataset A.1 + B.1, and the best hyperparameters were determined with mixed dataset A.2 + B.2; the 257 test images (A.3 + B.3) were then assessed.
The experimental results are listed in Table 6. It can be noted that, for the test set consisting of images with lesions (B.3), our proposed SR-IQA achieved PLCC 0.8418 with doctors’ labelling, while the global BCNN network that did not consider the local lesion features only achieved PLCC 0.6606, proving the necessity to design different IQA networks for ultrasound images with and without lesions. Due to the accuracy improvement for the images with lesions, our proposed global–local integrated IQA framework had better performance than the global BCNN network only, with PLCC 0.8851, SRCC 0.8834, Accuracy 0.8288, and RMSE 0.4541, which showed the advantage of our proposed global–local integrated framework.
Figure 8 shows the confusion matrix of IQA experiments. It can be noted that for the global BCNN network only, the image quality of some ultrasound images with lesions were incorrectly scored, reducing the performance of the overall IQA task. The reason can be explained in the two aspects. Firstly, the lesions produced hypoechoic dark areas in the ultrasound image, which may have affected the assessment of the global quality evaluation. Secondly, some images containing lesions had poor global quality, but the lesion areas were clear enough to help doctors in making their diagnosis; the doctor would then classify them as high-quality images. Therefore, implementing image quality assessment separately according to the existence of lesions significantly improved the consistency of the evaluation method with the physician’s clinical judgement.

3.8. Real-Time Performance

In practice, ultrasound image quality evaluation methods can be used to evaluate images acquired by physicians, to provide guidance to new physicians, or as image feedback for ultrasound autonomous scanning robots. Therefore, we should consider its Real-Time Performance. We have conducted real-time tests with the trained network to evaluate the quality of 257 ultrasound images from the test dataset. The experimental results show that the total time cost is 55.51 s, and the average time to evaluate the quality of a single image is 0.216 s. This is acceptable for a quality evaluation application. Of course, this is also relevant to the equipment on which the application is deployed.

4. Discussion and Conclusions

This paper presented a novel breast US image quality assessment framework integrating global IQA for ultrasound images without lesions and local SF-IQA for ultrasound images with lesions. The proposed SF-IQA used the lesion segmentation mask as the reference and transformed the medical ultrasound image assessment problem from no reference to reduced reference. To assess the image quality of the local lesion area, a comprehensive evaluation index, including the pixel-level features MSE and PSNR and the semantic-level features CR and SSIM, is proposed. Thanks to the specially designed image features and classification network for the lesion area, our proposed SF-IQA achieved a large improvement for quality assessment of ultrasound images with lesions (PLCC 0.8418 vs. 0.6606), proving the necessity of designing different IQA networks for ultrasound images with and without lesions. Due to the accuracy improvement for the images with lesions, our proposed global–local integrated IQA framework had better performance in the IQA task than the global BCNN network that did not consider the local lesion features, with PLCC improving from 0.8306 to 0.8851. The soft-reference method we propose can be extended to images containing cystic lesions, and other medical images where a clear visualization of key anatomical structures is required. For example, the area around the carotid artery in a carotid ultrasound image, or the cartilage and surrounding area in a knee ultrasound image. However, The IQA framework proposed in this research can still be further improved. Infiltrating cancer has unclear boundaries and may be judged as low quality by the IQA network. In our future work, benign–malignant decisions will be added to the IQA framework, along with the development of ultrasound image IQA methods for malignant lesions.

Author Contributions

Conceptualization, Z.W., B.Z. and B.L.; methodology, Z.W. and B.Z.; software, Y.S.; validation, Y.S.; formal analysis, Z.W. and B.Z.; investigation, B.L.; data curation, Z.Z. and F.L.; writing—original draft preparation, Z.W. and B.Z.; writing—review and editing, Y.H.; visualization, L.Y.; supervision, B.Z. and B.L.; project administration, Y.H.; funding acquisition, Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China (2022YFB4703500), the National Natural Science Foundation of China (No. 62273328, U21A20489), Guangdong Fundamental Research Program (No. 2021B1515130003, 2021B1515120011, 2023A1515030130), Shenzhen Science and Technology Program (No. JCYJ20200109115201707, JCYJ20200109112818703, JCYJ20220818101408019). This work is also supported by Joint Fund of State Key Laboratory of Robotics (No. 2022-KF-22-07).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of Chinese PLA General Hospital (protocol code YSB-2020-Y0902).

Informed Consent Statement

The Ethics Committee of Chinese PLA General Hospital waived the need for informed consent.

Data Availability Statement

The datasets presented in this article are not readily available because the hospital’s research ethics board does not permit sharing of clinical images.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  2. Brinkley, D.; Haybittle, J.L. The Curability of Breast Cancer. Lancet 1975, 306, 95–97. [Google Scholar] [CrossRef] [PubMed]
  3. Bainbridge, D.; McConnell, B.; Royse, C. A review of diagnostic accuracy and clinical impact from the focused use of perioperative ultrasound. Can. J. Anesth. 2018, 65, 371–380. [Google Scholar] [CrossRef] [Green Version]
  4. Sencha, A.N.; Evseeva, E.V.; Mogutov, M.S.; Patrunov, Y.N. Breast Ultrasound; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  5. Wang, Z.W.; Zhao, B.L.; Zhang, P.; Yao, L.; Wang, Q.; Li, B.; Meng, M.Q.H.; Hu, Y. Full-Coverage Path Planning and Stable Interaction Control for Automated Robotic Breast Ultrasound Scanning. IEEE Trans. Ind. Electron. 2023, 70, 7051–7061. [Google Scholar] [CrossRef]
  6. Liu, S.Y.; Thung, K.H.; Lin, W.L.; Shen, D.G.; Yap, P.T. Hierarchical Nonlocal Residual Networks for Image Quality Assessment of Pediatric Diffusion MRI With Limited and Noisy Annotations. IEEE Trans. Med. Imaging 2020, 39, 3691–3702. [Google Scholar] [CrossRef] [PubMed]
  7. Chow, L.S.; Paramesran, R. Review of medical image quality assessment. Biomed. Signal Process. Control 2016, 27, 145–154. [Google Scholar] [CrossRef]
  8. Shiao, Y.H.; Chen, T.J.; Chuang, K.S.; Lin, C.H.; Chuang, C.C. Quality of compressed medical images. J. Digit. Imaging 2007, 20, 149–159. [Google Scholar] [CrossRef] [Green Version]
  9. Choong, M.K.; Logeswaran, R.; Bister, M. Improving diagnostic quality of MR images through controlled lossy compression using SPIHT. J. Med. Syst. 2006, 30, 139–143. [Google Scholar] [CrossRef]
  10. Abolmaesumi, P.; Salcudean, S.E.; Zhu, W.H.; Sirouspour, M.R.; DiMaio, S.P. Image-guided control of a robot for medical ultrasound. IEEE Trans. Robot. Autom. 2002, 18, 11–23. [Google Scholar] [CrossRef] [Green Version]
  11. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  12. Fantini, I.; Yasuda, C.; Bento, M.; Rittner, L.; Cendes, F.; Lotufo, R. Automatic MR image quality evaluation using a Deep CNN: A reference-free method to rate motion artifacts in neuroimaging. Comput. Med. Imaging Graph. 2021, 90, 101897. [Google Scholar] [CrossRef]
  13. Oszust, M.; Bielecka, M.; Bielecki, A.; Stepien, I.; Obuchowicz, R.; Piorkowski, A. Blind image quality assessment of magnetic resonance images with statistics of local intensity extrema. Inf. Sci. 2022, 606, 112–125. [Google Scholar] [CrossRef]
  14. Shen, Y.X.; Sheng, B.; Fang, R.G.; Li, H.T.; Dai, L.; Stolte, S.; Qin, J.; Jia, W.P.; Shen, D.G. Domain-invariant interpretable fundus image quality assessment. Med. Image Anal. 2020, 61, 101654. [Google Scholar] [CrossRef] [PubMed]
  15. Khan, Z.A.; Beghdadi, A.; Kaaniche, M.; Alaya-Cheikh, F.; Gharbi, O. A neural network based framework for effective laparoscopic video quality assessment. Comput. Med. Imaging Graph. 2022, 101, 102121. [Google Scholar] [CrossRef]
  16. Ali, S.; Zhou, F.; Bailey, A.; Braden, B.; East, J.E.; Lu, X.; Rittscher, J. A deep learning framework for quality assessment and restoration in video endoscopy. Med. Image Anal. 2021, 68, 101900. [Google Scholar] [CrossRef] [PubMed]
  17. Khanin, A.; Anton, M.; Reginatto, M.; Elster, C. Assessment of CT Image Quality Using a Bayesian Framework. IEEE Trans. Med. Imaging 2018, 37, 2687–2694. [Google Scholar] [CrossRef] [PubMed]
  18. Chatelain, P.; Krupa, A.; Navab, N. Confidence-Driven Control of an Ultrasound Probe. IEEE Trans. Robot. 2017, 33, 1410–1424. [Google Scholar] [CrossRef] [Green Version]
  19. Karamalis, A.; Wein, W.; Klein, T.; Navab, N. Ultrasound confidence maps using random walks. Med. Image Anal. 2012, 16, 1101–1112. [Google Scholar] [CrossRef]
  20. Wu, L.Y.; Cheng, J.Z.; Li, S.L.; Lei, B.Y.; Wang, T.F.; Ni, D. FUIQA: Fetal Ultrasound Image Quality Assessment with Deep Convolutional Networks. IEEE Trans. Cybern. 2017, 47, 1336–1349. [Google Scholar] [CrossRef]
  21. Lin, Z.H.; Li, S.L.; Ni, D.; Liao, Y.M.; Wen, H.X.; Du, J.; Chen, S.P.; Wang, T.F.; Lei, B.Y. Multi-task learning for quality assessment of fetal head ultrasound images. Med. Image Anal. 2019, 58, 101548. [Google Scholar] [CrossRef]
  22. Camps, S.M.; Houben, T.; Carneiro, G.; Edwards, C.; Antico, M.; Dunnhofer, M.; Martens, E.; Baeza, J.A.; Vanneste, B.G.L.; van Limbergen, E.J.; et al. Automatic quality assessment of transperineal ultrasound images of the male pelvic region, using deep learning. Ultrasound Med. Biol. 2020, 46, 445–454. [Google Scholar] [CrossRef]
  23. Antico, M.; Vukovic, D.; Camps, S.M.; Sasazawa, F.; Takeda, Y.; Le, A.T.H.; Jaiprakash, A.T.; Roberts, J.; Crawford, R.; Fontanarosa, D.; et al. Deep Learning for US Image Quality Assessment Based on Femoral Cartilage Boundary Detection in Autonomous Knee Arthroscopy. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2020, 67, 2543–2552. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Zhang, S.Y.; Wang, Y.F.; Jiang, J.Y.; Dong, J.X.; Yi, W.W.; Hou, W.G. CNN-Based Medical Ultrasound Image Quality Assessment. Complexity 2021, 2021, 9938367. [Google Scholar] [CrossRef]
  25. Song, Y.X.; Zhong, Z.M.; Zhao, B.L.; Zhang, P.; Wang, Q.; Wang, Z.W.; Yao, L.; Lv, F.Q.; Hu, Y. Medical Ultrasound Image Quality Assessment for Autonomous Robotic Screening. IEEE Robot. Autom. Lett. 2022, 7, 6290–6296. [Google Scholar] [CrossRef]
  26. He, F.X.; Liu, T.L.; Tao, D.C. Why ResNet Works? Residuals Generalize. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 5349–5362. [Google Scholar] [CrossRef] [Green Version]
  27. Woo, S.H.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  28. Lin, T.Y.; RoyChowdhury, A.; Maji, S. Bilinear CNN Models for Fine-Grained Visual Recognition. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1449–1457. [Google Scholar]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the 18th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  30. Bindu, C.H.; Prasad, K.S. An efficient medical image segmentation using conventional OTSU method. Int. J. Adv. Sci. Technol. 2012, 38, 67–74. [Google Scholar]
  31. Malkomes, G.; Schaff, C.; Garnett, R. Bayesian optimization for automated model selection. In Proceedings of the 30th Conference on Neural Information Processing Systems (NIPS), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  32. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  33. Zhao, H.S.; Shi, J.P.; Qi, X.J.; Wang, X.G.; Jia, J.Y. Pyramid Scene Parsing Network. In Proceedings of the 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar]
  34. Chen, L.C.E.; Zhu, Y.K.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 833–851. [Google Scholar]
  35. Zhang, Q.; Cui, Z.P.; Niu, X.G.; Geng, S.J.; Qiao, Y. Image Segmentation with Pyramid Dilated Convolution Based on ResNet and U-Net. In Proceedings of the 24th International Conference on Neural Information Processing (ICONIP), Guangzhou, China, 14–18 November 2017; pp. 364–372. [Google Scholar]
  36. Matsoukas, C.; Haslum, J.F.; Söderberg, M.; Smith, K. Is It Time to Replace CNNS with Transformers for Medical Images? arXiv 2021, arXiv:2108.09038. [Google Scholar]
Figure 1. Global–local integrated breast ultrasound IQA framework.
Figure 1. Global–local integrated breast ultrasound IQA framework.
Bioengineering 10 00940 g001
Figure 2. Classification and segmentation networks with shared weights. (a) CBAM-ResNet18 classification network. (b) U-Res-CBAM segmentation network. (c) Residual block with CBAM module. (d) Spatial attention module. (e) Channel attention module.
Figure 2. Classification and segmentation networks with shared weights. (a) CBAM-ResNet18 classification network. (b) U-Res-CBAM segmentation network. (c) Residual block with CBAM module. (d) Spatial attention module. (e) Channel attention module.
Bioengineering 10 00940 g002
Figure 3. The proposed soft-reference IQA method based on lesion segmentation and SVM classifier.
Figure 3. The proposed soft-reference IQA method based on lesion segmentation and SVM classifier.
Bioengineering 10 00940 g003
Figure 4. The original ultrasound image, the segmentation mask obtained by the neural network, and the segmentation result obtained by the OSTU algorithm of lesion ROI.
Figure 4. The original ultrasound image, the segmentation mask obtained by the neural network, and the segmentation result obtained by the OSTU algorithm of lesion ROI.
Bioengineering 10 00940 g004
Figure 5. Heat map of the classification network for different ultrasound images. (a,b) are images without lesion, (c,d) are images with lesion.
Figure 5. Heat map of the classification network for different ultrasound images. (a,b) are images without lesion, (c,d) are images with lesion.
Bioengineering 10 00940 g005
Figure 6. The lesion segmentation results with different networks.
Figure 6. The lesion segmentation results with different networks.
Bioengineering 10 00940 g006
Figure 7. Data distribution of the four parameters with respect to different quality scores.
Figure 7. Data distribution of the four parameters with respect to different quality scores.
Bioengineering 10 00940 g007
Figure 8. Confusion matrix of IQA experiments. (a) Result of global–local integrated framework for mixed data. (b) Result of BCNN method for mixed data. (c) Result of SR-IQA method for positive data. (d) Result of BCNN method for positive data. The colors on the diagonal of (a) is the darkest, which indicates that the method performs best.
Figure 8. Confusion matrix of IQA experiments. (a) Result of global–local integrated framework for mixed data. (b) Result of BCNN method for mixed data. (c) Result of SR-IQA method for positive data. (d) Result of BCNN method for positive data. The colors on the diagonal of (a) is the darkest, which indicates that the method performs best.
Bioengineering 10 00940 g008
Table 1. The numbers of breast ultrasound images with different quality scores.
Table 1. The numbers of breast ultrasound images with different quality scores.
LesionSamples
Score1234
NegativeImageBioengineering 10 00940 i001Bioengineering 10 00940 i002Bioengineering 10 00940 i003Bioengineering 10 00940 i004
Number98112345289
PositiveImageBioengineering 10 00940 i005Bioengineering 10 00940 i006Bioengineering 10 00940 i007Bioengineering 10 00940 i008
Number4749161184
Table 2. Result of classification experiment.
Table 2. Result of classification experiment.
ModelAccuracy of Validation SetAccuracy of Test SetBest Validation Epoch
ResNet18 without pre-training0.9660.95315
ResNet18 with pre-training0.9700.96511
CBAM-ResNet18 with pre-training0.9890.9925
Table 3. Result of global IQA experiment.
Table 3. Result of global IQA experiment.
DatasetPLCCSRCCAccuracyRMSE
Negative (A.3)0.89640.87650.84020.4615
Table 4. Result of segmentation experiment.
Table 4. Result of segmentation experiment.
ModelPretrainedVal-DiceTest-Dice
VGG16 ImageNet0.8346 ± 0.08390.8633 ± 0.0936
PSPNet ImageNet0.8700 ± 0.11740.8789 ± 0.1024
DeepLab V3+ ImageNet0.8968 ± 0.05600.8824 ± 0.0865
U-Res ImageNet0.9056 ± 0.16760.8932 ± 1.1895
ViTImageNet0.9162 ± 0.09640.9065 ± 0.6352
U-Res-CBAMImageNet0.9272 ± 0.11550.9126 ± 0.0971
Table 5. Result of local SF-IQA experiment.
Table 5. Result of local SF-IQA experiment.
ModelValPLCCValSRCCValAccuracyValRMSETestPLCCTestSRCCTestAccuracyTestRMSE
Vgg16 0.55870.52190.62320.71610.63810.60110.62500.6657
PSPNet 0.65460.63070.62890.60280.68300.70550.63640.6307
Deeplab v3+0.65000.65160.66010.65840.72540.71180.69320.5839
U-Res 0.77350.78000.75070.51600.80310.79450.75000.5000
U-Res-CBAM0.83070.84300.81300.45160.84180.84620.80680.4395
Ground-truth0.92550.93940.90650.30580.89960.92510.86360.3693
Table 6. Result of global–local integrated IQA and global only IQA.
Table 6. Result of global–local integrated IQA and global only IQA.
DatasetModelPLCCSRCCAccuracyRMSE95% CI of Acc
Positive (B.3)SR-IQA0.84180.84620.80680.4395(0.7243, 0.8893)
Positive (B.3)BCNN0.66060.70620.71590.6946(0.6217, 0.8101)
Mixed (A.3 + B.3)Global-local0.88510.88340.82880.4541(0.7501, 0.9075)
Mixed (A.3 + B.3)BCNN0.83060.83380.80160.5649(0.7183, 0.8849)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Song, Y.; Zhao, B.; Zhong, Z.; Yao, L.; Lv, F.; Li, B.; Hu, Y. A Soft-Reference Breast Ultrasound Image Quality Assessment Method That Considers the Local Lesion Area. Bioengineering 2023, 10, 940. https://doi.org/10.3390/bioengineering10080940

AMA Style

Wang Z, Song Y, Zhao B, Zhong Z, Yao L, Lv F, Li B, Hu Y. A Soft-Reference Breast Ultrasound Image Quality Assessment Method That Considers the Local Lesion Area. Bioengineering. 2023; 10(8):940. https://doi.org/10.3390/bioengineering10080940

Chicago/Turabian Style

Wang, Ziwen, Yuxin Song, Baoliang Zhao, Zhaoming Zhong, Liang Yao, Faqin Lv, Bing Li, and Ying Hu. 2023. "A Soft-Reference Breast Ultrasound Image Quality Assessment Method That Considers the Local Lesion Area" Bioengineering 10, no. 8: 940. https://doi.org/10.3390/bioengineering10080940

APA Style

Wang, Z., Song, Y., Zhao, B., Zhong, Z., Yao, L., Lv, F., Li, B., & Hu, Y. (2023). A Soft-Reference Breast Ultrasound Image Quality Assessment Method That Considers the Local Lesion Area. Bioengineering, 10(8), 940. https://doi.org/10.3390/bioengineering10080940

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop