Next Article in Journal
The Relationship Between FIB-4 Score and Dynapenia in Older Adults
Previous Article in Journal
Applying Intraoperative Portal Venography in Liver Transplantation Vascular Surgery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Computer-Assisted Diagnosis of Pleural Effusion in Chest X-Rays via Deep Learning

1
Program on Semiconductor Manufacturing Technology Academy of Innovative Semiconductor and Sustainable Manufacturing, National Cheng Kung University, Tainan City 701401, Taiwan
2
Department of Medicine, College of Medicine, Chang Gung University, Taoyuan City 333423, Taiwan
3
Division of Thoracic Oncology, Department of Respiratory and Critical Care Medicine, Chang Gung Memorial Hospital, Chiayi City 613016, Taiwan
4
Department of Respiratory Care, Chang Gung University of Science and Technology, Chiayi County 61363, Taiwan
5
Department of Medical Education, Chang Gung Memorial Hospital Linkou, Taoyuan City 333423, Taiwan
6
Department of Electrical Engineering, Ming Chi University of Technology, New Taipei City 243303, Taiwan
7
Department of Electronic Engineering, Feng Chia University, Taichung City 40724, Taiwan
8
Department of Information Management, Chung Yuan Christian University, Taoyuan City 320317, Taiwan
9
Department of Electronic Engineering, National Cheng Kung University, Tainan City 701401, Taiwan
10
Ateneo Laboratory for Intelligent Visual Environments, Department of Information Systems and Computer Science, Ateneo de Manila University, Quezon City 1108, Philippines
11
Department of Automatic Control Engineering, Feng Chia University, Taichung City 40724, Taiwan
12
Department of Aeronautical Engineering, Chaoyang University of Technology, Taichung City 41349, Taiwan
*
Authors to whom correspondence should be addressed.
Diagnostics 2025, 15(18), 2322; https://doi.org/10.3390/diagnostics15182322
Submission received: 12 August 2025 / Revised: 5 September 2025 / Accepted: 11 September 2025 / Published: 13 September 2025
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

Background/Objectives: Pleural effusion is a common pulmonary condition that, if left untreated, may lead to respiratory distress and severe complications. Chest X-ray (CXR) imaging is routinely used by physicians to identify signs of pleural effusion. However, manually examining large volumes of CXR images on a daily basis can require substantial time and effort. To address this issue, this study proposes an automated pleural effusion detection system for CXR images. Methods: The proposed system integrates image cropping, image enhancement, and the EfficientNet-B0 deep learning model to assist in detecting pleural effusion, a task that is often challenging due to subtle symptom presentation. Image cropping was applied to extract the region from the heart to the costophrenic angle as the target area. Subsequently, image enhancement techniques were employed to emphasize pleural effusion features, thereby improving the model’s learning efficiency. Finally, EfficientNet-B0 was used to train and classify pleural effusion cases based on processed images. Results: In the experimental results, the proposed image enhancement approach improved the model’s recognition accuracy by approximately 4.33% compared with the non-enhanced method, confirming that enhancement effectively supports subsequent model learning. Ultimately, the proposed system achieved an accuracy of 93.27%, representing a substantial improvement of 21.30% over the 77.00% reported in previous studies, highlighting its significant advancement in pleural effusion detection. Conclusions: This system can serve as an assistive diagnostic tool for physicians, providing standardized detection results, reducing the workload associated with manual interpretation, and improving the overall efficiency of pulmonary care.

1. Introduction

According to the chronic obstructive pulmonary disease (COPD) statistics released by the World Health Organization (WHO) in 2023 [1], approximately 390 million people worldwide are affected by COPD, and this number continues to rise each year. The disease ranks as the third leading cause of death globally and is particularly prevalent in low- and middle-income countries [2]. This situation underscores the fact that many patients face challenges in accessing adequate medical services for the prevention and treatment of lung diseases. Promoting lung health is therefore crucial for improving overall public health, as maintaining healthy lung function not only enhances patients’ quality of life but also reduces the burden on healthcare systems [3]. In this context, raising awareness of the importance of early diagnosis and treatment of lung diseases is of particular significance.
Medical imaging technologies have played a crucial role in achieving this goal, with chest X-rays (CXR) serving as one of the most widely used diagnostic tools in pulmonary medicine [4]. Among the conditions detectable through CXR, pleural effusion occurs when an abnormal amount of fluid accumulates in the pleural cavity. Under normal physiological conditions, the pleural cavity contains only a small amount of fluid that facilitates smooth movement between the lungs and the chest wall [5]. Excessive fluid accumulation can result in pleural effusion, which is often associated with severe complications and is positively correlated with various lung diseases. It frequently occurs alongside pathological conditions such as pulmonary infections or tumor-induced compression, contributing to critical complications including inflammation [6,7]. CXR imaging enables clinicians to identify fluid accumulation in the lungs, as shown in Figure 1, making accurate detection of pleural effusion a vital step in diagnosis. In Figure 1a, the red box highlights the costophrenic angle, which represents a key region for identifying pleural effusion.
Building upon the importance of CXR imaging in diagnosing pleural effusion, recent advancements in automated medical imaging technologies have demonstrated strong potential to improve both diagnostic efficiency and accuracy in clinical practice [8]. The accelerated progress of artificial intelligence (AI) has introduced new opportunities for enhancing diagnostic precision, accelerating medical image analysis, and improving treatment efficiency [9,10]. Given the high number of CXRs that radiologists must interpret daily, recent research has explored the integration of deep learning techniques for detecting thoracic diseases, including pleural effusion [11]. Convolutional neural networks (CNNs) have achieved notable results in this area, with applications ranging from image segmentation and enhancement for lung disease detection to object detection models that have achieved COVID-19 detection accuracies exceeding 89% [12,13]. Some techniques have already been proposed for the automated detection of pleural effusion in CXR images, including ResNet-18 [5], support vector machines [14], and DenseNet121 [15], which reported accuracies of 77.00%, 87.70%, and 87.18%, respectively. In addition, S. Serte and A. Serener [16] introduced a CNN-based approach for pleural effusion detection in CXR images that achieved an accuracy as high as 91.00%. Beyond single CNN models, ensemble approaches combining multiple CNNs have also been explored to further improve recognition accuracy [17], incorporating architectures such as AlexNet, MobileNet-V2, VGG, ResNet-18, and ResNet-50. However, despite the growing adoption of deep learning in medical imaging, relatively few studies have specifically targeted the automated detection of pleural effusion in CXR images. This gap underscores the need for focused research to develop accurate, reliable, and automated diagnostic tools.
To address this need, the present study proposes a CXR-based analysis method designed to provide radiologists with standardized diagnostic information, thereby automating clinical workflows and reducing inter-observer variability in image interpretation [18,19]. The proposed system integrates an image cropping algorithm, image enhancement techniques, and the EfficientNet-B0 architecture [20] to achieve automated detection of pleural effusion. A summary of the major contributions is provided below:
  • This study employs image segmentation to pre-process lung X-rays, retaining only the lower half of the lungs, from below the heart to the costophrenic angle, thereby enabling the model to focus more effectively on the target region.
  • An effective image enhancement approach is introduced, which improves model performance by 4.33% compared with results obtained without enhancement.
  • The proposed system integrates image cropping, image enhancement, and CNN classification, achieving an accuracy of 927%, representing a substantial improvement of 21.30% over existing studies.

2. Methods

This study proposed a system for detecting pleural effusion in CXR images by integrating various image processing methods with artificial intelligence. To achieve accurate detection, the workflow was divided into four main stages, as shown in Figure 2. First, raw CXR images were input into the system. Image preprocessing was then performed to remove peripheral noise and irrelevant background information. Next, image cropping was applied to retain only the target regions, thereby reducing the influence of excessive non-essential information on model training efficiency. Image enhancement methods were subsequently used to highlight features associated with pleural effusion. Finally, a classical CNN was employed for model training and evaluation. This study focused on developing a computer-aided diagnostic system to support pulmonologists in clinical decision-making by providing more objective and standardized recognition results, ultimately improving patient care efficiency.

2.1. Image Preprocessing for CXR Image

However, the costophrenic angle is a key anatomical landmark for examining pleural effusion. Under normal conditions, the structure of the costophrenic angle is clearly visible on CXR images. When pleural effusion occurs, fluid accumulates in these angles, causing them to appear blurred or obscured on CXR images. Additionally, since pleural effusion is influenced by gravity and tends to accumulate in the lower parts of the lungs, this study focused on the lower half of the lung, specifically the region from the heart to the costophrenic angle, as the target area. To further improve the accuracy of region extraction, an image preprocessing method was introduced to enable precise segmentation of the costophrenic angle.

2.1.1. Standardization Negative

To enhance the clarity of lung boundaries, the original CXR images were first converted to grayscale before applying further preprocessing, thereby reducing computational complexity. Each pixel was converted based on the weighted average of its red, green, and blue (RGB) channel values. The resulting grayscale image is shown in Figure 3a.
Subsequently, a negative transformation was applied to the grayscale image, as defined in Equation (1). By inverting the grayscale values, darker regions became brighter, while brighter regions became darker, increasing the contrast between the lungs and surrounding tissues. This process effectively enhanced the lung contours, making the lung regions more distinguishable, as shown in Figure 3b.
I n e g a t i v e = 1 I g r a y

2.1.2. AHE

After grayscale conversion and negative transformation, the lung regions became more visually prominent. To further enhance image contrast and improve the visibility of fine details, adaptive histogram equalization (AHE) was applied to the preprocessed images, as shown in Figure 3c. AHE enhances local contrast by redistributing pixel intensities within small regions of the image, making subtle structures such as lung contours more clearly visible. Compared to conventional histogram equalization, AHE is more effective in handling images with uneven brightness, making it particularly suitable for medical imaging applications.

2.1.3. Binarization

To further simplify the image for subsequent segmentation, binarization was applied to convert the grayscale image into a binary format containing only black and white pixels, thereby reducing its complexity. A threshold value of 0.5 was selected for the binarization process. This process enhanced the contrast between the lung regions and the surrounding background, facilitating more efficient and accurate analysis in the following computational steps. After binarization, the lung area could be delineated more precisely. However, to further refine the lung boundaries and eliminate noise, morphological operations were employed to improve the structural integrity of the segmented regions. Morphological processing is a widely used technique in image analysis, particularly effective in enhancing specific image features. Among various operations, dilation and erosion are the most fundamental, with their definitions provided in Equations (2) and (3), respectively. The result of the image preprocessing step is presented in Figure 3d.
A B = ( A B ) B
A B = ( A B ) B

2.2. Image Segmentation

After binarization, the pixel values of the image were summed column by column to generate a vertical projection profile. This profile provides an intuitive view of pixel intensity distribution across the image and serves as the basis for determining suitable cropping boundaries. In this study, the vertical projection was used to identify the left and right cropping positions. Specifically, within the leftmost one-fifth of the image, the position with the lowest accumulated pixel value was selected as the candidate for the left boundary. Similarly, the rightmost one-fifth was analyzed to determine the right boundary based on the minimum pixel sum. Once both the left and right boundaries were identified, they were mapped back to the original CXR image to perform precise cropping, as shown in Figure 4a,b.
To further separate the lung region from surrounding structures, horizontal cropping was performed following the vertical cropping step. The horizontal cropping method followed a similar procedure to the vertical approach, in which pixel values were summed row by row to generate the cumulative vertical projection from top to bottom. To determine the upper boundary, the first row within the central one-third of the image whose total pixel value exceeded a predefined threshold was selected. This ensured accurate preservation of the upper margin of the lungs. The lower boundary was determined by identifying the row with the minimum accumulated pixel value in the lower half of the image, thereby preserving the complete structure of the costophrenic angle. Once the upper and lower boundaries were confirmed, they were mapped onto the CXR image that had already been cropped along the left and right boundaries, and vertical cropping was performed accordingly, as shown in Figure 4c,d. This method retains the main anatomical structures of the lungs while effectively excluding irrelevant peripheral regions, ensuring the integrity of the lung area and the removal of non-lung regions.
Additionally, to highlight the key features of pleural effusion, the lung region was further horizontally divided at the midpoint after completing both horizontal and vertical cropping, with only the lower half of the lung, from the heart to the costophrenic angle, being retained. The final cropped images were uniformly resized to 227×227 pixels to standardize the input dimensions for model training and ensure compatibility with the subsequent model requirements. The results of image segmentation and normalization are shown in Figure 4e.

2.3. Image Enhancement

In CXR images, pleural effusion typically occurred in the lower regions of the lungs and appeared as white or grayish-white areas, whereas healthy lung tissue was represented by relatively darker pixels. The visibility of pleural effusion on CXR images was influenced by various factors, including the extent, density, and location of the fluid. Therefore, accurately identifying pleural effusion was challenging. To address this issue, this study developed a feature enhancement algorithm for pleural effusion. The algorithm was applied to the cropped CXR images, representing the region of interest (ROI), and combined histogram stretching with Sobel gradient edge detection [21,22]. Histogram stretching improved image contrast, making the costophrenic angle more discernible, while Sobel edge detection emphasized the boundaries between tissues of different densities, thereby enhancing the clarity of pleural effusion features. In this way, both the visual characteristics and edge structures of pleural effusion in CXR images were effectively enhanced.

2.3.1. Histogram Stretching

Enhanced imaging was necessary to better distinguish pleural effusion symptoms from surrounding lung contours. In CXR images, pleural effusion typically appeared as gray and blurred regions, while normal lung tissue remained relatively dark. This made accurate detection difficult, particularly in early stages. To improve image contrast, this study tested three enhancement methods: logarithmic transformation [23], Contrast Limited Adaptive Histogram Equalization (CLAHE) [24], and histogram stretching [25], with the aim of highlighting the costophrenic angle. Among these methods, histogram stretching demonstrated the most effective performance; therefore, this study employed histogram stretching as a computationally efficient enhancement technique, as shown in Figure 5. Histogram stretching effectively increased the brightness contrast between the symptomatic areas and adjacent normal tissues, making the effusion more visually distinguishable. This enhancement supported improved feature extraction by the CNN, thereby improving classification and detection accuracy.
Histogram stretching worked by identifying the minimum and maximum pixel intensity values in the input image and linearly mapping them to the full dynamic range. This transformation amplified subtle differences in pixel intensity, making previously indistinct features more pronounced. The detailed algorithmic steps are presented in Algorithm 1.
Algorithm 1. Histogram Stretching.
Input  I i : filtering input image.
    I o : Scaling constant.
    ( K w , K h ) : Dimensions of the input image.
Output  I o   : Transformed output image.
     I o x , y = ( I i x , y m i n ) × 255 / ( max m i n )
Hint: x  0 K w 1 ,   n 0 K h 1
    min, max: minimum and maximum pixel values in input image

2.3.2. Sobel Gradient Edge Detection

Since histogram stretching enhanced the visibility of the costophrenic angle but was insufficient to highlight the features of pleural effusion, this study further employed edge detection techniques to enhance pleural effusion characteristics. The performance of several methods, including Sobel [26,27], Sobel gradient, sharpening [28], and Canny edge detection [29], was evaluated in terms of their impact on subsequent model classification. Among these, the Sobel gradient demonstrated the best performance; therefore, it was selected in this study to further enhance the features of pleural effusion. Specifically, the Sobel gradient edge detection algorithm was adopted to identify edge contours within the CXR images. The Sobel operator computed the image intensity gradient at each pixel, indicating the direction of the greatest rate of change from light to dark, along with the magnitude of that change. It utilized two 3 × 3 convolution kernels, one designed to detect horizontal changes and the other for vertical changes. These kernels were convolved with the original image to approximate the gradient in each direction, which were then combined to compute the overall gradient magnitude. This method was particularly effective in medical imaging, as it emphasized the boundaries between different tissue regions while maintaining robustness against noise. The detailed computational steps of the Sobel gradient operator are provided in Algorithm 2.
Algorithm 2. Sobel Gradient Edge Detection.
1. 
Apply Gaussian filter to reduce noise (optional pre-processing step).
G x , y = 1 2 π σ 2 e x p ( x 2 + y 2 ) 2 σ 2
2. 
Apply Sobel operators to compute horizontal and vertical gradients.
G x = 1 ,   0 ,   1 ,   2 ,   0 ,   2 ,   1 ,   0 ,   1 × I x , y
G y = [ [ 1 , 2 , 1 ] ,   [ 0 ,   0 ,   0 ] ,   [ 1 ,   2 ,   1 ] ] × I ( x , y )
3. 
Calculate gradient magnitude.
| G | = ( G x 2 + G y 2 )
4. 
Calculate gradient direction (optional).
θ ( x , y ) = t a n 1 ( G y / G x )
5. 
Apply threshold to create binary edge map (optional).
If |G| > threshold, pixel is an edge
Otherwise, pixel is not an edge
6. 
Combine with original image (optional).
Result = α × O r i g i n a l + β × G r a d i e n t _ m a p
After applying the Sobel gradient edge detection, the resulting gradient map enhanced the visibility of the costophrenic angle. In normal lung tissue, the contour of the costophrenic angle appeared sharp and exhibited a steep gradient. However, in the presence of pleural effusion, this region became blurred. This distinct difference in gradient patterns enabled the model to better capture the characteristics of pleural effusion, thereby improving diagnostic accuracy. The results of image enhancement are shown in Figure 6. In the enhanced healthy lung images, the curvature from the pleura to the costophrenic angle follows a normal arc, as shown in Figure 6a. In contrast, for images containing pleural effusion, this expected curvature is absent due to fluid accumulation, as shown in Figure 6b.

2.4. CNN Training and Validation

With the rapid advancement of machine learning and artificial intelligence technologies, CNNs have been increasingly applied to image processing tasks and have demonstrated significant success in various image classification problems. In terms of model selection, this study evaluated several CNN architectures, including MobileNet-v3 [30], SqueezeNet [31], DarkNet19 [32], AlexNet [33], and EfficientNet-B0. Model performance was assessed using 5-fold cross-validation to ensure robustness and stability. The hardware and software platform used in this study is shown in Table 1. Among these, EfficientNet-B0 achieved the best performance and was therefore chosen as the backbone model. By leveraging clinically acquired CXR images, the proposed system was designed to be well-suited for practical implementation in real-world clinical settings. Among them, EfficientNet-B0 is a highly efficient CNN architecture that balances network depth, width, and input resolution to achieve optimal performance [34]. It employs a compound scaling method that simultaneously scales these dimensions, thereby improving accuracy while reducing computational cost. EfficientNet-B0 has shown strong performance in both general image classification and medical image analysis, effectively enhancing disease recognition accuracy and reducing the risk of overfitting.
During CNN training, hyperparameters played a crucial role in influencing the learning behavior and overall performance of the model. The learning rate, batch size, and epochs are among the most commonly adjusted hyperparameters. The learning rate controlled the step size for weight updates during training. Specifically, a higher learning rate accelerated model convergence but could miss the optimal solution, whereas a lower learning rate ensured more stable updates but required significantly more training time. In addition, the batch size defined how many samples were used to update the model parameters in each iteration of training. Larger batch sizes could accelerate training and provide smoother gradient estimates but required more memory. On the other hand, smaller batch sizes introduced more variance in updates, which could help avoid local minimum but might also lead to instability during training. Moreover, the number of epochs referred to how many times the model was trained on the entire training dataset. Although increasing the number of epochs allowed the model to learn more effectively from the data, it also increased the risk of overfitting. Therefore, selecting appropriate hyperparameters was essential to ensure effective model training. The specific hyperparameter settings used in this study are summarized in Table 2.

3. Results

The study procedures complied with ethical standards and were approved by the Institutional Review Board (IRB) of Chiayi Chang Gung Memorial Hospital (IRB No. 202301914B0). All CXR images used in this study were collected from adult patients (aged 18 and above), and the Chang Gung Medical Foundation Institutional Review Board approves the waiver of the participants’ consent. A total of 922 CXR images were included, comprising 461 images of normal lungs and 461 images diagnosed with pleural effusion. Among them, 80% (738 images) were used for model training, while the remaining 20% (184 images) were reserved for validation. The distribution of the training data is summarized in Table 3.
This section presents the experimental results of the proposed CNN approach for detecting pleural effusion in CXR images. The evaluation includes four main components: the performance metrics used in this study, a comparison of model performance with single-lung and whole-lung as inputs, an assessment of various image enhancement methods, and a comparative analysis of classification results across multiple CNN architectures. The following subsections describe these experiments in detail, highlighting the key findings and their implications for clinical application.

3.1. Performance Metrics

The primary objective was to reduce the workload of radiologists while minimizing the risk of misdiagnosis. Accuracy, precision, recall, and F1 score were adopted as evaluation metrics for the model’s performance, as they are commonly used in classification tasks. As shown in Table 4, the confusion matrix served as the basis for calculating the evaluation metrics defined in Equations (4)–(7).
Accuracy = TP + TN TP + TN + FP + FN
Precision = TP TP + FP
Recall = TP TP + FN
F 1   s core = 2 × Precision   ×   Recall Precision + Recall

3.2. Comparison of Single-Lung and Whole-Lung Inputs

In this study, the performance of single-lung and whole-lung inputs was compared. The comparison results of single-lung and whole-lung inputs are shown in Table 5, where the single-lung input was further divided into left and right lung cases. The left lung achieved an accuracy of 66.30%, whereas the right lung achieved a higher accuracy of 73.37%. This indicated that the right lung outperformed the left lung, likely due to the influence of the heart’s position, which interfered with the model’s ability to detect left-sided pleural effusion.
Furthermore, the heart’s position made the imaging characteristics of the left lung more complex, making it more difficult to accurately capture symptomatic information. As a result, the average single-lung accuracy was 69.84%. In contrast, the whole-lung input achieved an accuracy of 84.24%, showing that pleural effusion features were more clearly represented in the whole-lung images. Based on these comparative results, the whole-lung image was ultimately selected as the CNN input for training and validation.

3.3. Evaluation of Different Image Enhancement Method

To effectively improve model accuracy, this study applied multiple image enhancement techniques to the cropped CXR images. The enhancement process was divided into two stages, targeting the costophrenic angle and pleural effusion features, respectively.
For the enhancement of costophrenic angle features, three methods were compared: logarithmic transformation, CLAHE, and histogram stretching, along with the unenhanced images. The training results for each method are summarized in Table 6. Histogram stretching achieved the best overall performance, with accuracy at 89.67%, precision at 90.06%, recall at 89.67%, and F1 score at 89.65%. The original images without enhancement achieved an accuracy of 87.50%. CLAHE yielded accuracy comparable to the original method, while logarithmic transformation performed the worst, with all evaluation metrics significantly lower than the other three approaches. Therefore, histogram stretching was selected as the enhancement method for costophrenic angle features in this stage.
Following the selection of histogram stretching for enhancing the costophrenic angle features, this study further evaluated multiple edge detection techniques to enhance the features of pleural effusion. The experiments were conducted on images that had been cropped and processed with histogram stretching. The evaluated methods included Sobel, Sobel gradient, sharpening, and Canny edge detection, and the results are presented in Table 7. As shown in Table 7, the Sobel gradient method achieved the best performance, with these four metrics all reaching 91.85%. Sharpening and Sobel yielded accuracies of 88.04% and 86.96%, respectively. In contrast, the Canny edge detection method had the lowest performance among the four methods, with all metrics at 83.15%. Based on these results, histogram stretching combined with the Sobel gradient method was ultimately selected as the image enhancement approach to improve the accuracy and performance for the model training.

3.4. The Classification Results of CNN

To further verify the stability of the proposed pleural effusion detection model, 5-fold cross-validation was conducted to evaluate multiple deep learning architectures. Five different CNN architectures were tested, including MobileNet-v3, SqueezeNet, DarkNet19, AlexNet, and EfficientNet-B0. As shown in Table 8, EfficientNet-B0 attained the best accuracy of 93.27%, closely followed by AlexNet at 92.74%. SqueezeNet and DarkNet19 recorded accuracy of 90.02% and 89.49%, respectively. In contrast, MobileNet-v3 yielded the lowest performance, with an accuracy of only 77.33%. Based on these findings, EfficientNet-B0 was selected as the final model for pleural effusion recognition in this study. Ultimately, the proposed system achieved an accuracy of 93.27% ± 0.02, a precision of 92.95% ± 0.03, a recall of 92.81% ± 0.02, and an F1 score of 92.79% ± 0.02. To further verify the clinical applicability of the model, the National Institutes of Health CXR Dataset [35] containing CXR images with pleural effusion was also used for testing. The results demonstrated an accuracy of 93.50 ± 0.03, a precision of 93.00 ± 0.02, and a recall of 93.00 ± 0.02, confirming the reliability of the system and its potential for clinical implementation.
To evaluate the contribution of the proposed system, a comparison was conducted with previous studies, as summarized in Table 9. The results presented a direct comparison between earlier pleural effusion detection methods and the approach developed in this study. The proposed method achieved the highest accuracy at 93.27%, representing an improvement of approximately 21.30% over Method [5], which achieved 77.00%. In terms of precision, Method [17] obtained the highest value at 99.00%, exceeding the proposed method by 6.05%. For recall, Method [16] and Method [14] achieved 100.00% and 94.70%, respectively, while the proposed method reached 92.81%, indicating potential for further improvement in this metric. However, in terms of the F1 score, which balances precision and recall, the proposed method outperformed all other approaches, demonstrating higher overall reliability. These notable improvements over prior methods indicated that the proposed system was capable of effectively capturing subtle features of pleural effusion in CXR images, thereby offering a more reliable diagnostic tool for clinical applications.

4. Discussion

Building on the comparative results presented in the previous section, this study specifically addressed the diagnostic challenges radiologists face in detecting pleural effusion, particularly in its early stages when the symptoms are subtle and often indistinguishable on standard CXR images. Analysis of our experimental results revealed that, while the detection accuracy for the right lung was comparable to that of whole-lung imaging, the accuracy for the left lung was adversely affected by its anatomical overlap with the heart. This observation motivated our focus on isolating the costophrenic angle to enhance detection accuracy.
To optimize feature extraction, a series of targeted image processing techniques were applied, including selective cropping and enhancement of the costophrenic angle. Among the evaluated enhancement methods, the combination of histogram stretching and Sobel gradient edge detection achieved the most significant improvement, with accuracy, precision, recall, and F1 score all exceeding 91.85%, outperforming other edge detection techniques. Regarding classification performance, EfficientNet-B0 demonstrated the highest overall effectiveness, achieving an accuracy of 93.27%, which was substantially higher than other deep learning models tested. Compared with previous studies, the proposed method achieved a 93.27% accuracy, representing an improvement of 21.30%, and obtained the highest model reliability as reflected in the top-performing F1 score.
These improvements can be attributed to the integrated approach of costophrenic angle cropping, advanced image enhancement, and edge detection, which collectively amplified pleural effusion features, thereby facilitating more efficient and accurate model training. In contrast, methods [5,15,16,17] relied solely on CNN-based classification without incorporating image preprocessing or cropping methods. However, it should be noted that the dataset used in our study for training and validation was collected from a Chiayi Chang-Gung Memorial Hospital during the study period, which differs from the datasets employed in methods [5,15,16,17]. This discrepancy may limit the fairness of direct performance comparisons. Future research will focus on further improving the model’s accuracy and robustness. Compared with previous studies, the proposed system still has room for improvement in recall performance, which may be influenced by hyperparameter settings during training or the limited size of the dataset. To address this, future work will combine multiple publicly available datasets for model training and validation. This approach is expected not only to expand the dataset for training but also to enhance the model’s applicability in clinical practice. In addition, future research plans to incorporate Generative Adversarial Networks (GANs) for image augmentation, comparing their performance and computational cost with traditional preprocessing methods. This approach is expected to reduce decision-making time while providing more comprehensive support for medical diagnosis and contributing to improved pulmonary healthcare services. Overall, the findings of this study demonstrate that a carefully designed combination of anatomical region selection, targeted image enhancement, and deep learning classification can substantially improve the detection of pleural effusion in CXR images, offering a clinically valuable and computationally efficient tool for early diagnosis and treatment planning. In the future, this system could be integrated with existing Picture Archiving and Communication Systems (PACS) to provide physicians and patients with preliminary identification results of pleural effusion, thereby enhancing clinical decision-making and fostering better communication between physicians and patients.

5. Conclusions

This study proposed an automated pleural effusion detection system that achieved significant advancements in diagnosis through optimized image preprocessing and deep learning models. The system first performed precise cropping of CXR images to preserve the critical costophrenic angles, followed by image enhancement using histogram stretching and Sobel gradient edge detection. The proposed preprocessing workflow proved more effective than other approaches, increasing the accuracy from 87.50% without enhancement to 91.85%, thereby confirming that image enhancement can effectively improve the visibility of key features and provide a solid foundation for the subsequent classification stage.
In the classification step, 5-fold cross-validation was employed to evaluate the performance of multiple CNN architectures. EfficientNet-B0 was ultimately selected as the classification model, achieving an accuracy of 93.27%. Compared with previous methods, the proposed system improved accuracy by approximately 21.30%, with both precision and recall exceeding 92.00%, demonstrating reliable detection capabilities for both positive and negative pleural effusion cases.
Overall, by integrating precise cropping with efficient image enhancement, the proposed system substantially improved the accuracy and robustness of pleural effusion detection, showing strong potential for clinical application. Future study will focus on expanding the dataset and further enhancing image preprocessing techniques to continually improve the model’s practicality and reliability in real-world clinical environments.

Author Contributions

Conceptualization, T.-K.C. and K.-C.L.; Data curation, Y.-C.L. and S.-H.T.; Formal analysis, S.-W.C. and C.-C.C.; Funding acquisition, T.-K.C. and T.-Y.C.; Investigation, Y.-C.L. and S.-H.T.; Methodology, Y.-Y.H. and T.-Y.C.; Project administration, T.-Y.C.; Resources, Y.-C.L. and S.-H.T.; Software, S.-W.C. and C.-C.C.; Supervision, T.-K.C. and W.-C.T.; Validation, S.-W.C. and C.-C.C.; Visualization, K.-C.L. and W.-C.T.; Writing—original draft, Y.-Y.H., K.-C.L. and P.A.R.A.; Writing—review and editing, Y.-Y.H. and P.A.R.A. All authors will be updated at each stage of manuscript processing, including submission, revision, and revision reminder, via emails from our system or the assigned Assistant Editor. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported, in part, by the National Science and Technology Council, Taiwan, under grant numbers 112-2410-H-197-002-MY2, 113-2314-B-182A-140, 113-2221-E-131-026, 114-2221-E-035-032, 114-2221-E-131-009, and 114-2314-B-182A-051. And This work was supported by the Research Support of the Feng Chia University Research Program, grant no. 24H00810.

Institutional Review Board Statement

Chang Gung Medical Foundation Institutional Review Board; IRB number: 202301914B0; Date of Approval: 20 December 2023; Protocol Title: Application of Convolutional Neural Network for Disease Identification from Chest Radiograph; Executing Institution: Chang-Gung Medical Foundation Chiayi Chang-Gung Memorial Hospital of Chiayi; Duration of Approval: From 1 January 2024 to 31 December 2024; The IRB reviewed and determined that it is expedited review according to Case research or cases treated or diagnosed by clinical routines. However, this does not include HIV-positive cases.

Informed Consent Statement

The Chang Gung Medical Foundation Institutional Review Board approves the waiver of the participants’ consent. The research does not adversely affect the rights and welfare of the subjects. The study uses de-identified or non-traceable data, records, documents, information, or specimens obtained from a legally established biological database, ensuring that individual identities cannot be identified.

Data Availability Statement

The datasets presented in this article are not readily available because they are part of an ongoing study and will be made available only after the completion of data collection and analysis. Requests to access the datasets should be directed to the corresponding authors at simon-chi@mail.mcut.edu.tw or tsungychen@fcu.edu.tw.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tripathy, R.K.; Dash, S.; Rath, A.; Panda, G.; Pachori, R.B. Automated Detection of Pulmonary Diseases From Lung Sound Signals Using Fixed-Boundary-Based Empirical Wavelet Transform. IEEE Sens. Lett. 2022, 6, 7001504. [Google Scholar] [CrossRef]
  2. Chen, Z.; Zhou, Y.P.; Liu, X.; Jiang, X.; Wu, T.; Ghista, D.; Xu, X.Q.; Zhang, H.; Jing, Z.C. A Personalized Pulmonary Circulation Model to Non-Invasively Calculate Fractional Flow Reserve for Artery Stenosis Detection. IEEE Trans. Biomed. Eng. 2022, 69, 1435–1448. [Google Scholar] [CrossRef] [PubMed]
  3. Roy, A.; Thakur, A.; Satija, U. VGAResNet: A Unified Visibility Graph Adjacency Matrix-Based Residual Network for Chronic Obstructive Pulmonary Disease Detection Using Lung Sounds. IEEE Sens. Lett. 2023, 7, 7006604. [Google Scholar] [CrossRef]
  4. Shah, P.M.; Ullah, F.; Shah, D.; Gani, A.; Maple, C.; Wang, Y.; Abrar, M.; Islam, S.U. Deep GRU-CNN Model for COVID-19 Detection From Chest X-Rays Data. IEEE Access 2022, 10, 35094–35105. [Google Scholar] [CrossRef] [PubMed]
  5. Serte, S.; Serener, A. Early pleural effusion detection from respiratory diseases including COVID-19 via deep learning. In Proceedings of the 2020 Medical Technologies Congress (TIPTEKNO), Antalya, Turkey, 19–20 November 2020; pp. 1–4. [Google Scholar] [CrossRef]
  6. Hu, J.; Zhang, C.; Zhou, K.; Gao, S. Chest X-Ray Diagnostic Quality Assessment: How Much Is Pixel-Wise Supervision Needed? IEEE Trans. Med. Imaging 2022, 41, 1711–1723. [Google Scholar] [CrossRef]
  7. Wu, Y.; Kong, Q.; Zhang, L.; Castiglione, A.; Nappi, M.; Wan, S. CDT-CAD: Context-Aware Deformable Transformers for End-to-End Chest Abnormality Detection on X-Ray Images. IEEE/ACM Trans. Comput. Biol. Bioinform. 2024, 21, 823–834. [Google Scholar] [CrossRef]
  8. Nahiduzzaman, M.; Goni, M.O.F.; Anower, M.S.; Islam, M.R.; Ahsan, M.; Haider, J.; Gurusamy, S.; Hassan, R.; Islam, M.R. A Novel Method for Multivariant Pneumonia Classification Based on Hybrid CNN-PCA Based Feature Extraction Using Extreme Learning Machine With CXR Images. IEEE Access 2021, 9, 147512–147526. [Google Scholar] [CrossRef]
  9. Paranjape, K.; Schinkel, M.; Nanayakkara, P. Short Keynote Paper: Mainstreaming Personalized Healthcare–Transforming Healthcare Through New Era of Artificial Intelligence. IEEE J. Biomed. Health Inform. 2020, 24, 1860–1863. [Google Scholar] [CrossRef] [PubMed]
  10. Ahmed, I.; Jeon, G.; Piccialli, F. From Artificial Intelligence to Explainable Artificial Intelligence in Industry 4.0: A Survey on What, How, and Where. IEEE Trans. Ind. Inform. 2022, 18, 5031–5042. [Google Scholar] [CrossRef]
  11. Sogancioglu, E.; Van Ginneken, B.; Behrendt, F.; Bengs, M.; Schlaefer, A.; Radu, M.; Xu, D.; Sheng, K.; Scalzo, F.; Marcus, E.; et al. Nodule Detection and Generation on Chest X-Rays: NODE21 Challenge. IEEE Trans. Med. Imaging 2024, 43, 2839–2853. [Google Scholar] [CrossRef]
  12. Ji, H.; Li, J.; Wang, L.; Fan, L.; Zhang, Y.; Wang, W. COVID-19 Detection in CXR Image Using High Frequency Emphasis Filtering Based Convolutional Neural Network. In Proceedings of the 2022 IEEE 11th Data Driven Control and Learning Systems Conference (DDCLS), Chengdu, China, 3–5 August 2022; pp. 929–934. [Google Scholar] [CrossRef]
  13. Bhadra, R.; Kar, S. COVID Detection from CXR Scans using Deep Multi-layered CNN. In Proceedings of the 2020 IEEE Bombay Section Signature Conference (IBSSC), Mumbai, India, 4–6 December 2020; pp. 214–218. [Google Scholar] [CrossRef]
  14. Liu, Y.; Liang, Z.; Yang, J.; Yuan, S.; Wang, S.; Huang, W.; Wu, A. Diagnostic and comparative performance for the prediction of tuberculous pleural effusion using machine learning algorithms. Int. J. Med. Inform. 2024, 182, 105320. [Google Scholar] [CrossRef]
  15. Thapa, M.; Kaur, R. An Explainable Deep—Learning based Multi-Label Image Classification for Chest X-Rays. Procedia Comput. Sci. 2025, 258, 2425–2434. [Google Scholar] [CrossRef]
  16. Furriel, B.C.; Paulo, A.J.; Ribeiro, G.A.; Santos, P.V.; Reis, E.P.; Mendes, G.S.; Paiva, J.Q.; Loureiro, R.M.; Rittner, L.; Reis, M.R. Convolutional Neural Network for Pleural Effusion Classification. In Proceedings of the 2023 IEEE Latin American Conference on Computational Intelligence (LA-CCI), Recife-Pe, Brazil, 29 October–1 November 2023; pp. 1–6. [Google Scholar] [CrossRef]
  17. Serte, S.; Serener, A. Classification of COVID-19 and pleural effusion on chest radiographs using CNN fusion. In Proceedings of the 2021 International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Kocaeli, Turkey, 25–27 August 2021; pp. 1–6. [Google Scholar] [CrossRef]
  18. Chowdary, G.J.; Kanhangad, V. A Dual-Branch Network for Diagnosis of Thorax Diseases From Chest X-Rays. IEEE J. Biomed. Health Inform. 2022, 26, 6081–6092. [Google Scholar] [CrossRef] [PubMed]
  19. Taphorn, K.; Mechlem, K.; Sellerer, T.; De Marco, F.; Viermetz, M.; Pfeiffer, F.; Pfeiffer, D.; Herzen, J. Direct Differentiation of Pathological Changes in the Human Lung Parenchyma With Grating-Based Spectral X-ray Dark-Field Radiography. IEEE Trans. Med. Imaging 2021, 40, 1568–1578. [Google Scholar] [CrossRef] [PubMed]
  20. Ijaz, A.; Akbar, S.; AlGhofaily, B.; Hassan, S.A.; Saba, T. Deep Learning for Pneumonia Diagnosis Using CXR Images. In Proceedings of the 2023 Sixth International Conference of Women in Data Science at Prince Sultan University (WiDS PSU), Riyadh, Saudi Arabia, 14–15 March 2023; pp. 53–58. [Google Scholar] [CrossRef]
  21. Langarizadeh, M.; Mahmud, R.; Ramli, A.R.; Napis, S.; Beikzadeh, M.R.; Rahman, W.E.Z.W.A. Improvement of digital mammogram images using histogram equalization, histogram stretching and median filter. J. Med. Eng. Technol. 2011, 35, 103–108. [Google Scholar] [CrossRef]
  22. AS, R.A.; Gopalan, S. Comparative Analysis of Eight Direction Sobel Edge Detection Algorithm for Brain Tumor MRI Images. Procedia Comput. Sci. 2022, 201, 487–494. [Google Scholar] [CrossRef]
  23. Sedgwick, P. Log transformation of data. BMJ 2012, 345, e6727. [Google Scholar] [CrossRef]
  24. Sahu, S.; Singh, A.K.; Ghrera, S.P.; Elhoseny, M. An approach for de-noising and contrast enhancement of retinal fundus image using CLAHE. Opt. Laser Technol. 2019, 110, 87–98. [Google Scholar] [CrossRef]
  25. Li, X.; Hu, H.; Zhao, L.; Wang, H.; Yu, Y.; Wu, L.; Liu, T. Polarimetric image recovery method combining histogram stretching for underwater imaging. Sci. Rep. 2018, 8, 12430. [Google Scholar] [CrossRef]
  26. Gao, W.; Zhang, X.; Yang, L.; Liu, H. An improved Sobel edge detection. In Proceedings of the 2010 3rd International Conference on Computer Science and Information Technology, Chengdu, China, 9–11 July 2010; pp. 67–71. [Google Scholar] [CrossRef]
  27. Sun, J.; Lv, Y.; Tang, C.; Sima, H.; Wu, X. Face Recognition Based on Local Gradient Number Pattern and Fuzzy Convex-Concave Partition. IEEE Access 2020, 8, 35777–35791. [Google Scholar] [CrossRef]
  28. Ding, F.; Zhu, G.; Yang, J.; Xie, J.; Shi, Y.-Q. Edge Perpendicular Binary Coding for USM Sharpening Detection. IEEE Signal Process. Lett. 2015, 22, 327–331. [Google Scholar] [CrossRef]
  29. Bao, P.; Zhang, L.; Wu, X. Canny edge detection enhancement by scale multiplication. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1485–1490. [Google Scholar] [CrossRef]
  30. Huang, J.; Mei, L.; Long, M.; Liu, Y.; Sun, W.; Li, X.; Shen, H.; Zhou, F.; Ruan, X.; Wang, D.; et al. BM-Net: CNN-Based MobileNet-V3 and Bilinear Structure for Breast Cancer Detection in Whole Slide Images. Bioengineering 2022, 9, 261. [Google Scholar] [CrossRef] [PubMed]
  31. Tsivgoulis, M.; Papastergiou, T.; Megalooikonomou, V. An improved SqueezeNet model for the diagnosis of lung cancer in CT scans. Mach. Learn. Appl. 2022, 10, 100399. [Google Scholar] [CrossRef]
  32. Baygin, M.; Yaman, O.; Barua, P.D.; Dogan, S.; Tuncer, T.; Acharya, U.R. Exemplar Darknet19 feature generation technique for automated kidney stone detection with coronal CT images. Artif. Intell. Med. 2022, 127, 102274. [Google Scholar] [CrossRef] [PubMed]
  33. Shanthi, T.; Sabeenian, R.S. Modified Alexnet architecture for classification of diabetic retinopathy images. Comput. Electr. Eng. 2019, 76, 56–64. [Google Scholar] [CrossRef]
  34. Goutham, V.; Sameerunnisa, A.; Babu, S.; Prakash, T.B. Brain Tumor Classification using EfficientNet-B0 Model. In Proceedings of the 2022 2nd International Conference on Advance Computing and Innovative Technologies in Engineering (ICACITE), Greater Noida, India, 28–29 April 2022; pp. 2503–2509. [Google Scholar] [CrossRef]
  35. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-ray8: Hospital-scale Chest X-ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3462–3471. [Google Scholar] [CrossRef]
Figure 1. Pleural effusion symptoms on CXR image. (a) Absence of pleural effusion symptoms. (b) Presence of pleural effusion symptoms.
Figure 1. Pleural effusion symptoms on CXR image. (a) Absence of pleural effusion symptoms. (b) Presence of pleural effusion symptoms.
Diagnostics 15 02322 g001
Figure 2. The flowchart of the proposed system for automated detection of pleural effusion in CXR images.
Figure 2. The flowchart of the proposed system for automated detection of pleural effusion in CXR images.
Diagnostics 15 02322 g002
Figure 3. The CXR image preprocessing result. (a) Gray scale. (b) Negative. (c) AHE. (d) Binarization.
Figure 3. The CXR image preprocessing result. (a) Gray scale. (b) Negative. (c) AHE. (d) Binarization.
Diagnostics 15 02322 g003
Figure 4. The results of the CXR image segmentation. (a) Vertical cropping image. (b) Vertical cumulative pixelmap. (c) Horizontal cropping image. (d) Horizontal cumulative pixelmap. (e) Final segmented and normalized image.
Figure 4. The results of the CXR image segmentation. (a) Vertical cropping image. (b) Vertical cumulative pixelmap. (c) Horizontal cropping image. (d) Horizontal cumulative pixelmap. (e) Final segmented and normalized image.
Diagnostics 15 02322 g004aDiagnostics 15 02322 g004b
Figure 5. Results after applying histogram stretching. (a) Normal lung (b) Pleural effusion.
Figure 5. Results after applying histogram stretching. (a) Normal lung (b) Pleural effusion.
Diagnostics 15 02322 g005
Figure 6. Results after applying image enhancement method. (a) Normal lung (b) Pleural effusion.
Figure 6. Results after applying image enhancement method. (a) Normal lung (b) Pleural effusion.
Diagnostics 15 02322 g006
Table 1. The hardware and software platform used in this study.
Table 1. The hardware and software platform used in this study.
Hardware PlatformVersion
CPUIntel® Core™ i5-12400
GPUNVIDIA GeForce RTX™ 3060 Ti
Software platformVersion
Matlab2023a
Table 2. The hyperparameters used in CNN model training.
Table 2. The hyperparameters used in CNN model training.
HyperparameterValue
Learning Rate0.00001
Batch Size32
Epochs20
Table 3. The distribution of the training and validation database.
Table 3. The distribution of the training and validation database.
DatabaseTrainingValidationTotal
Normal Lung36992461
Pleural Effusion36992461
Table 4. An example of the confusion matrix.
Table 4. An example of the confusion matrix.
Ground Truth Value
TrueFalse
Predicted ValueTrue T p
(True Positive)
F p
(False Positive)
False F n
(False Negative)
T n
(True Negative)
Table 5. Performance comparison of using CNN to train single-lung and whole-lung pleural effusion classification.
Table 5. Performance comparison of using CNN to train single-lung and whole-lung pleural effusion classification.
Left LungRight LungAverage LungsWhole Lungs
Accuracy66.30%73.37%69.84%84.24%
Precision79.87%82.62%81.25%86.97%
Recall66.30%73.37%69.84%84.24%
F1 Score61.99%71.34%66.67%83.94%
Table 6. Comparison between the case without image enhancement and various image enhancement methods.
Table 6. Comparison between the case without image enhancement and various image enhancement methods.
OriginalLog TransformCLAHEHistogram Stretch
Accuracy87.50%64.13%87.50%89.67%
Precision88.52%77.38%88.04%90.06%
Recall87.50%64.13%87.50%89.67%
F1 Score87.42%59.19%87.46%89.65%
Table 7. The comparison between various edge detection.
Table 7. The comparison between various edge detection.
Histogram StretchHistogram Stretch + CannyHistogram Stretch + SharpenHistogram Stretch + Sobel Gradient
Accuracy89.67%83.15%88.04%91.85%
Precision90.06%83.16%88.50%91.86%
Recall89.67%83.15%88.04%91.85%
F1 Score89.65%83.15%88.01%91.85%
Table 8. The result with 5-fold cross-validation results of different CNN architectures.
Table 8. The result with 5-fold cross-validation results of different CNN architectures.
Mobilenet_v3SqueezenetDarknet19AlexnetEfficientnet_b0
Accuracy77.53%90.02%89.49%92.74%93.27%
Table 9. Comparison of Deep Learning Methods for Pleural Effusion Detection.
Table 9. Comparison of Deep Learning Methods for Pleural Effusion Detection.
Method
in [17]
Method
in [5]
Method
in [16]
Method
in [14]
Method
in [15]
This Work
Accuracy (%)83.0077.0091.0087.7087.1893.27
Precision (%)99.0082.2584.2185.30N/A92.95
Recall (%)75.0091.50100.0094.70N/A92.81
F1 score (%)75.0082.0091.4386.00N/A92.79
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Y.-Y.; Lin, Y.-C.; Tsai, S.-H.; Chi, T.-K.; Chen, T.-Y.; Chung, S.-W.; Li, K.-C.; Tu, W.-C.; R. Abu, P.A.; Chen, C.-C. Automated Computer-Assisted Diagnosis of Pleural Effusion in Chest X-Rays via Deep Learning. Diagnostics 2025, 15, 2322. https://doi.org/10.3390/diagnostics15182322

AMA Style

Huang Y-Y, Lin Y-C, Tsai S-H, Chi T-K, Chen T-Y, Chung S-W, Li K-C, Tu W-C, R. Abu PA, Chen C-C. Automated Computer-Assisted Diagnosis of Pleural Effusion in Chest X-Rays via Deep Learning. Diagnostics. 2025; 15(18):2322. https://doi.org/10.3390/diagnostics15182322

Chicago/Turabian Style

Huang, Ya-Yun, Yu-Ching Lin, Sung-Hsin Tsai, Tsun-Kuang Chi, Tsung-Yi Chen, Shih-Wei Chung, Kuo-Chen Li, Wei-Chen Tu, Patricia Angela R. Abu, and Chih-Cheng Chen. 2025. "Automated Computer-Assisted Diagnosis of Pleural Effusion in Chest X-Rays via Deep Learning" Diagnostics 15, no. 18: 2322. https://doi.org/10.3390/diagnostics15182322

APA Style

Huang, Y.-Y., Lin, Y.-C., Tsai, S.-H., Chi, T.-K., Chen, T.-Y., Chung, S.-W., Li, K.-C., Tu, W.-C., R. Abu, P. A., & Chen, C.-C. (2025). Automated Computer-Assisted Diagnosis of Pleural Effusion in Chest X-Rays via Deep Learning. Diagnostics, 15(18), 2322. https://doi.org/10.3390/diagnostics15182322

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop