Next Article in Journal
Visualization of Rainfall Using Participatory Mobile Sensing for Crop Cultivation Support
Previous Article in Journal
Development of a FATEK PLC Simulator for Industrial Processes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Determination of Anteroposterior and Posteroanterior Imaging Positions on Chest X-Ray Images Using Deep Learning †

by
Fatih Gökçimen
1,*,
Alpaslan Burak İnner
1 and
Özgür Çakır
2
1
Department of Computer Engineering, Artificial Intelligence and Simulation Systems R&D Laboratory, Kocaeli University, 41380 İzmit, Türkiye
2
Department of Radiology, Faculty of Medicine, Kocaeli University, 41000 Kocaeli, Türkiye
*
Author to whom correspondence should be addressed.
Presented at the International Conference on Electronics, Engineering Physics and Earth Science (EEPES 2025), Alexandroupolis, Greece, 18–20 June 2025.
Eng. Proc. 2025, 104(1), 58; https://doi.org/10.3390/engproc2025104058
Published: 28 August 2025

Abstract

This study proposes a deep learning framework to classify anteroposterior (AP) and posteroanterior (PA) chest X-ray projections automatically. Multiple convolutional neural networks (CNNs), including ResNet18, ResNet34, ResNet50, DenseNet121, EfficientNetV2-S, and ConvNeXt-Tiny, were utilized. The NIH Chest X-ray Dataset, with 112,120 images, was used with strict patient-wise splitting to prevent data leakage. ResNet34 achieved the highest performance: 99.65% accuracy, 0.9956 F1 score, and 0.9994 ROC-AUC. Grad-CAM visualized model decisions, and expert-reviewed misclassified samples were removed to enhance dataset quality. This methodology highlights the importance of robust preprocessing, model interpretability, and clinical applicability in radiographic view classification tasks.

1. Introduction

In recent years, deep learning-based approaches have made significant progress in the field of medicine and have become widely used, especially in radiological image analysis. More than 1 billion X-ray imaging protocols are performed worldwide each year, the majority of which are chest X-rays [1]. The correct classification of image acquisition positions is of great importance for the effective use of chest X-rays in diagnostic processes. Differences between anteroposterior (AP) and posteroanterior (PA) projections can lead to differences in the size and shape of the heart and mediastinal structures, which can increase the risk of misdiagnosis and evaluation [2]. Therefore, the automatic classification of AP and PA images and detection of mislabeled data play a critical role in increasing the accuracy of clinical decision support systems [1,3,4].
Various studies have been conducted on PA/AP classification in the literature [1,3]. For example, AP and PA images were classified in a deep learning model developed by Kim et al. (2019) using the NIH ChestX-ray14 dataset [3]. However, it was not clearly stated in this study whether patient-based data separation was performed. If different images belonging to the same patient were distributed to the training and test sets, the model might have learned patient-based features and provided an artificial success increase. This may reduce the generalization performance of the model in the real world.
In this study, patient-based data separation was carefully performed to eliminate deficiencies in the literature and increase the clinical applicability of the model. Data leakage was prevented by ensuring that images belonging to the same patient were assigned to only one dataset (training, validation, or testing). Thus, the model was evaluated on patients it had never encountered before and was tested in a way that was more suitable for real clinical scenarios.
In addition, age is known to be an important factor when determining the AP and PA imaging positions. In pediatric patients (especially those under 5 years of age), the AP imaging technique is used more frequently because young children are usually unable to stand and may have difficulty following breath-holding instructions. In this study, unlike previous studies, pediatric patients (under 18 years of age) in the dataset were analyzed without separation. The generalization ability of the model was increased by ensuring consistent performance across different age groups.
To analyze the performance of the model, the images with the highest loss values were determined in the validation dataset. These images were evaluated by an expert radiologist, and examples containing mislabeling, acquisition errors, or technical problems were removed from the dataset. In addition, images that the model had difficulty with but were medically meaningful were included in the training process. This process was performed to increase the generalization ability of the model and improve its clinical applicability.
Our proposed model differs from previous studies in terms of not only patient-based data separation but also performance evaluation methods. The success of our model, accuracy, error rate, F1 score, Cohen’s kappa, recall, specificity, precision, Brier score, balanced accuracy, and ROC-AUC were analyzed comprehensively using various metrics. This detailed evaluation demonstrates the reliability and clinical applicability of our model for PA/AP classification more strongly than in previous studies. The main objective of this study was to develop a reliable deep learning model that can accurately classify the shooting positions in chest X-rays and detect mislabeled or incorrectly captured images. With the proposed methodology, data quality is increased, the generalization ability of the model is improved, and a more reliable contribution is made to clinical decision support systems.

2. Method

2.1. Dataset

In this study, the NIH Chest X-ray Dataset [4] was used for AP and PA position detection. This dataset contained 112,120 frontal chest X-ray images from 30,805 unique patients. Each image had labels associated with 14 thoracic diseases.
All images in the dataset had 1024 × 1024 pixels, grayscale, and 8-bit color depth. In addition, the resolution of the images was specified to be 96 dpi.
This dataset was publicly shared and did not require special permission for use. In this study, data were obtained by observing patient privacy and ethical rules [5].
The number of patients in this dataset and their distributions are listed in Table 1.
When the dataset was examined, it was observed that there was a total of 112,120 images, with 67,310 (60.1%) in the PA position and 44,810 (39.9%) in the AP position. In binary classification problems, class balance is an important factor that affects the learning performance of a model. When the class ratios (60.1–39.9%) in this dataset are considered, it is seen that although it is not completely balanced (homogeneous), it does not contain any serious imbalance. Therefore, metrics such as precision, recall, and F1 score were taken into account when examining the results. When the dataset was examined, it was determined that more than one image belonged to a patient. It was observed that there was more than one PA or AP image belonging to the same patient, as well as images taken in both the PA and AP positions belonging to the same patient. The distribution of the number of images per patient is presented graphically in Figure 1.
The distribution of patient images with the most images is presented graphically in Figure 2. The colors in the bars represent a sequential color gradient to visually distinguish patients and do not correspond to any additional variable. Such repeated images require attention to the principle of patient-based independence during training. Splitting images belonging to the same patient between the training and test datasets is a critical step in reducing the risk of overfitting the model and increasing its generalizability to the real world. In addition, the presence of more than one image belonging to a patient in the dataset can lead to the risk of data leakage. This is especially true if the test dataset is included in the training dataset or if the test data are indirectly learned by the model during the training process. Data leakage can cause the model to overfit, and the results are inconsistent with the real-world performance. In addition, using more than one image belonging to the same patient can cause the model to “see” the test dataset, misleading its true performance. Although there were 112,120 images in the dataset used in this study, the number of patients with a single image was 17,503. The number of images belonging to a patient reached 184, indicating that data leakage should be carefully considered.
Sample radiographs of the AP and PA projections are shown in Figure 3. These images allow for visual comparison of the significant anatomical differences between the two viewing positions.

2.2. Fastai

Fastai is defined as a library that enables fast and easy development of deep learning models [6]. This library provides high-level components for practitioners and flexible and modular low-level components for researchers. Aiming to strike a balance between ease of use, flexibility, and performance, Fastai (version 2.8.2) has an architecture that optimizes deep learning and data processing by leveraging Python’s dynamism (version 3.12.8) and PyTorch’s flexibility (version 2.5.1).

2.3. Classification Studies

In this study, a computer with an i7 10th generation processor, an Nvidia RTX 2060 6 GB graphics card, and 16 GB RAM was used.
As mentioned above, considering the risk of data leakage, images from the same patient were included in only one dataset (training, validation, or test). The dataset was randomly divided into three subsets: 70% training (78,566 images), 15% validation (17,063 images), and 15% testing (16,491 images). In addition, the distribution of AP and PA positions was maintained at 40–60% in each dataset. This helped minimize the imbalance in the dataset so that the model could provide more generalizable and accurate results.
In this study, ResNet18, ResNet34, ResNet50, DenseNet121, EfficientNetV2-S, and Con-vNeXt-Tiny models were applied for classification using the Fastai library. During the training process, the models resized the images from 1024 to 224 pixels. Data augmentation techniques were used during the training and testing stages. However, these techniques were carefully selected based on the structure of radiological images. In particular, methods such as horizontal and vertical flipping were not applied because the images were contrary to body symmetry, and it was not meaningful to flip an X-ray image. Instead, more limited data augmentation techniques, such as slight rotation (maximum 5°), slight zoom (maximum 10%), and slight pan (maximum 10%), were used. Fastai randomly applied these data augmentation methods in each training stage, which allowed the model to encounter a wider variety of images and increased its generalization ability. In addition, considering the class imbalances in the dataset, such careful data augmentation techniques allow for more robust and generalizable models to be obtained.
In this study, the employed deep learning architectures and the resulting performance metrics are presented in Table 2. Multiple convolutional neural network (CNN) models—including ResNet18, ResNet34, ResNet50, DenseNet121, EfficientNetV2-S, and ConvNeXt-Tiny—were evaluated for the task of distinguishing AP and PA chest X-ray projections.
Each model was trained under two distinct conditions. Table 3 summarizes the results obtained with data augmentation techniques such as slight rotation, zoom, and pan, while Table 2 presents the outcomes without any augmentation. Key classification metrics—accuracy, error rate, F1 score, Cohen’s kappa score, and the area under the receiver operating characteristic curve (ROC-AUC)—were used to comprehensively evaluate model performance.
Multiple metrics were used to evaluate the model success. Accuracy indicates the overall success rate of the model, whereas the error rate indicates the percentage of incorrectly predicted examples. However, because the dataset is not fully balanced and may contain some class imbalance, the balanced accuracy score metric, which evaluates the accuracy of both classes by assigning equal weights, was also used.
Additional metrics were also used to evaluate the classification performance more comprehensively. Recall/sensitivity measures the rate of correct prediction of the positive class, while precision shows the accuracy of the model’s positive predictions. The F1 score, which provides a balance between these two metrics, was calculated to determine whether the model exhibited balanced performance in terms of both sensitivity and precision.
In addition, Cohen’s kappa score, which measures how successful the model is compared to random prediction, was used as an important performance indicator in cases of data imbalance. Finally, the ROC-AUC score is a critical measure that shows how well the model can distinguish between positive and negative classes at different decision thresholds. This metric was analyzed to evaluate the overall discrimination power of the model.
The validation loss was examined to evaluate the success of the model using the validation data. A decrease in the validation loss during the training process may indicate that the generalization ability of the model has increased, and an increase after a certain point may indicate signs of overfitting. In this context, while the validation loss in models without data augmentation remained at higher levels, particularly for ResNet50, it was observed that the validation loss decreased in models with data augmentation. This indicates that data augmentation improves the generalization ability of the model and reduces the risk of overfitting.
In addition, the Brier Score Loss metric was used to evaluate the reliability of the probabilistic estimates of the model. This metric is calculated by taking the average of the square of the difference between the probabilities predicted by the model and actual class labels. Lower Brier Score Loss values indicate that the model provides better-calibrated and reliable probability estimates.
In this study, the effects of different deep learning models and data augmentation methods on Brier Score Loss were analyzed. It was observed that the Brier Score Loss values decreased in most of the models with data augmentation, indicating that the model produced more reliable probability estimates. When this result was evaluated together with the ROC-AUC and other performance metrics, it was revealed that the model provided better-calibrated probabilities in the decision-making process.
In the experiments, the performances of used models on the training, validation, and test data were analyzed. The overall success of the models was compared by applying different data augmentation scenarios.
The performance results of the developed models for the test dataset are presented in Table 4. In the evaluation, AP was considered the correct class, whereas misclassified PA was regarded as the incorrect class (False PA).
In the experimental studies, the performance of deep learning architectures used for the AP/PA orientation classification task has been analyzed comprehensively. The validation data results presented in Table 2 and Table 3 and the test data results presented in Table 4 show that all models achieved very high and close success levels above 99% in metrics such as accuracy, ROC-AUC, and F1 score. This confirms that all examined architectures have high competence in the mentioned classification task.
However, when a detailed comparison of the metrics is made, the ResNet34 model stands out as the architecture that exhibits the most balanced and consistent performance. The model achieved an ROC-AUC score of 0.999726, an F1 score of 0.996863, and an accuracy rate of 0.996249 on the test data. One of the most important findings supporting this high success is that the model recorded the lowest error rate among all the compared models with a value of 0.003751 in the data augmentation scenario. This extremely low error rate shows that the model has both high overall classification success and is more resistant to erroneous predictions compared to other models. In addition, the low valid loss of 0.014919 obtained during the training process is another important finding that supports the strength of the model’s generalization ability and its resistance to overfitting.
In addition to this consistent success in numerical metrics, practical advantages were also decisive in choosing the ResNet34 architecture. Having fewer parameters compared to deeper and more complex models such as ResNet50 provides significant efficiency advantages such as faster training times and lower computational cost. The lighter structure of the model reduces the risk of overfitting, while facilitating the applicability and deployment of the model, especially in clinical environments or on limited hardware resources.
In conclusion, although all tested models showed high performance, the ResNet34 model was determined to be the most suitable model to be used in this study due to its consistency in numerical metrics, the lowest error rate, and the optimal balance between the computational efficiency it offers. In this regard, the ResNet34 model was taken as basis in all subsequent analysis and evaluation stages of the project.
The images with the highest loss values in the validation phase of the model and examples that were incorrectly predicted in the test data were examined by an expert radiologist. The images that were truly faulty or problematic for the model were removed from the dataset; however, the images that the model had difficulty with but were medically meaningful and correct were included in the training process. The sample images extracted from the dataset are shown in Figure 4.
Figure 4a–i present sample images that were misclassified by the model and subsequently removed from the dataset. Below each image, the predicted class label, the true class label, the loss value, and the prediction probability are shown. For instance, in Figure 4a, the model predicted the image as “PA”, whereas the true label is “AP”, with a loss value of 7.18 and a prediction probability of 1.00. These images were analyzed to better understand the challenges in the model’s classification performance and the potential causes of misclassification.
In this study, the binary cross-entropy loss function provided by the Fastai library was employed for binary classification tasks. The loss values were calculated using the following formula:
L y , y ^ = 1 N i = 1 N [ y i l o g ( σ ( y ^ i ) ) + 1 y i l o g ( 1 σ y ^ i ) ]
Here, N denotes the total number of samples, y i ∈ {0, 1} represents the true class label, y ^ i is the logit output produced by the model, σ( y ^ i ) = 1/(1 + e^(− y ^ i )) refers to the sigmoid activation function, and L( y , y ^ ) denotes the average binary cross-entropy loss. This loss function enhances the sensitivity of the model to misclassifications, thereby enabling it to produce more reliable and robust predictions.
After the removal of the misclassified samples from the dataset, the model was retrained, resulting in a noticeable improvement in the performance metrics. Evaluations conducted on the test dataset showed a statistically significant enhancement in the model performance. The training and test results are listed in Table 5 and Table 6, respectively.
To enhance the interpretability of the proposed CNN-based model, Gradient-weighted Class Activation Mapping (Grad-CAM) was employed to visualize class-discriminative regions in selected chest X-ray images.
Figure 5 illustrates two representative cases, each corresponding to an AP and PA projection, respectively. In the AP-predicted image (Figure 5a), the model focuses significantly on the lower right thoracic region and the presence of medical devices (e.g., ECG leads or a port on the shoulder), which are frequently associated with supine positioning and thus AP projection. This localization suggests the model leverages clinically relevant visual cues to distinguish the view.
Conversely, in the PA-predicted image (Figure 5b), the attention is primarily distributed over the central thoracic cavity, particularly the heart and lung hilum areas. This pattern is consistent with standard upright PA positioning, where the anatomical structures are more uniformly exposed and less obstructed by external artifacts.
These Grad-CAM results confirm that the model not only achieves high classification accuracy but also bases its decisions on plausible anatomical and contextual features, thereby supporting its reliability and potential clinical applicability.

3. Conclusions

In this study, the ResNet34 deep learning architecture demonstrated high performance in classifying AP and PA chest X-ray positions. Patient-based data separation and carefully selected data augmentation techniques enhanced the generalization ability of the model, supporting its applicability in clinical settings. The high-performance metrics, 99.65% accuracy and 0.99935 ROC-AUC, indicate that the model can serve as a reliable aid in diagnostic processes.
The exclusion of misclassified images following expert radiologist reviews highlights the critical importance of data quality in model performance. However, further evaluation of the model using external datasets and multicenter studies is recommended to better assess its generalizability. Future studies may also focus on incorporating factors, such as image quality and patient demographics, to further enhance the model’s clinical utility.
This study provides a solid foundation for the integration of automated classification systems into routine radiological practices.

Author Contributions

Conceptualization, F.G. and A.B.İ.; methodology, F.G. and A.B.İ.; software, F.G.; validation, Ö.Ç.; formal analysis, F.G.; investigation, F.G.; resources, Ö.Ç.; data curation, F.G.; writing—original draft preparation, F.G.; writing—review and editing, A.B.İ. and Ö.Ç.; visualization, F.G.; supervision, A.B.İ. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Dataset available on request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hosch, R.; Kroll, L.; Nensa, F.; Koitka, S. Differentiation between Anteroposterior and Posteroanterior Chest X-Ray View Position with Convolutional Neural Networks. RoFo 2021, 193, 168–176. [Google Scholar] [CrossRef] [PubMed]
  2. Raoof, S.; Feigin, D.; Sung, A.; Raoof, S.; Irugulpati, L.; Rosenow, E.C. Interpretation of plain chest roentgenogram. Chest 2012, 141, 545–558. [Google Scholar] [CrossRef] [PubMed]
  3. Kim, T.K.; Yi, P.H.; Wei, J.; Shin, J.W.; Hager, G.; Hui, F.K.; Sair, H.I.; Lin, C.T. Deep Learning Method for Automated Classification of Anteroposterior and Posteroanterior Chest Radiographs. J. Digit. Imaging 2019, 32, 925–930. [Google Scholar] [CrossRef] [PubMed]
  4. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-ray8: Hospital-scale Chest X-ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3462–3471. [Google Scholar] [CrossRef]
  5. Nguyen, H.Q.; Lam, K.; Le, L.T.; Pham, H.H.; Tran, D.Q.; Nguyen, D.B.; Le, D.D.; Pham, C.M.; Tong, H.T.T.; Dinh, D.H.; et al. VinDr-CXR: An open dataset of chest X-rays with radiologist’s annotations. Sci. Data 2022, 9, 429. [Google Scholar] [CrossRef] [PubMed]
  6. Howard, J.; Gugger, S. Fastai: A Layered API for Deep Learning. Information 2020, 11, 108. [Google Scholar] [CrossRef]
Figure 1. Distribution of the number of chest X-ray images per patient (NIH Dataset).
Figure 1. Distribution of the number of chest X-ray images per patient (NIH Dataset).
Engproc 104 00058 g001
Figure 2. Distribution of the number of images for patients with the most images (first 20 patients).
Figure 2. Distribution of the number of images for patients with the most images (first 20 patients).
Engproc 104 00058 g002
Figure 3. Visualization of exemplary radiographs from both view positions (PA and AP). The text markers (e.g., “L”, “PORTABLE”) appearing in some images are standard radiographic orientation indicators automatically included during image acquisition. These markers were not used in the analysis.
Figure 3. Visualization of exemplary radiographs from both view positions (PA and AP). The text markers (e.g., “L”, “PORTABLE”) appearing in some images are standard radiographic orientation indicators automatically included during image acquisition. These markers were not used in the analysis.
Engproc 104 00058 g003
Figure 4. Sample images misclassified by the model, identified as technically flawed by an expert radiologist, and subsequently excluded from the dataset.
Figure 4. Sample images misclassified by the model, identified as technically flawed by an expert radiologist, and subsequently excluded from the dataset.
Engproc 104 00058 g004
Figure 5. Grad-CAM visualizations for (a) AP and (b) PA chest X-ray classification. The heatmap overlay indicates the regions of the image most influential for the model’s prediction, where warmer colors (e.g., red) denote higher importance and cooler colors (e.g., blue) denote lower importance.
Figure 5. Grad-CAM visualizations for (a) AP and (b) PA chest X-ray classification. The heatmap overlay indicates the regions of the image most influential for the model’s prediction, where warmer colors (e.g., red) denote higher importance and cooler colors (e.g., blue) denote lower importance.
Engproc 104 00058 g005
Table 1. Distribution of patients in the dataset used.
Table 1. Distribution of patients in the dataset used.
FeatureValue
Total Number of Images112,120
Number of PA Images (%)67,310 (60.1%)
Number of AP Images (%)44,810 (39.9%)
Number of Male Patient Images63,340
Number of Female Patient Images48,780
Number of Unique Patients30,805
Average Number of Images per Patient3.64
Table 2. AP/PA classification performance of different CNN architectures without data augmentation.
Table 2. AP/PA classification performance of different CNN architectures without data augmentation.
MetricEfficientNetV2-SConvNeXt-TinyDenseNet-121ResNet-18ResNet-34ResNet-50
Train Loss0.0168560.0332270.0244160.0003120.0002390.013660
Valid Loss0.0163910.0215570.0178440.0308120.0236180.023737
Error Rate0.0039850.0051570.0043950.0040440.0034580.004747
Accuracy0.9960150.9948430.9956050.9959560.9965420.995253
F1 Score0.9966680.9956900.9963250.9966200.9971090.996035
Cohen’s Kappa0.9917110.9892710.9908570.9915880.9928080.990121
Recall0.9976460.9969590.9972540.9978420.9980380.997940
Precision0.9956930.9944230.9953980.9954010.9961820.994137
Brier Score Loss0.0039850.0051570.0043950.0040440.0034580.004747
Balanced Accuracy0.9956200.9943300.9952050.9954990.9961800.994602
ROC-AUC0.9995100.9985670.9993840.9991510.9995520.999357
Table 3. AP/PA classification performance of different CNN architectures with data augmentation.
Table 3. AP/PA classification performance of different CNN architectures with data augmentation.
MetricEfficientNetV2-SConvNeXt-TinyDenseNet-121ResNet-18ResNet-34ResNet-50
Train Loss0.0240580.0126950.0090120.0279340.0223220.025057
Valid Loss0.0172300.0205570.0177760.0219100.0149190.022696
Error Rate0.0042200.0048060.0041610.0053330.0037510.005392
Accuracy0.9957800.9951940.9958390.9946670.9962490.994608
F1 Score0.9964730.9959820.9965200.9955420.9968630.995496
Cohen’s Kappa0.9912230.9900040.9913470.9889060.9921990.988781
Recall0.9975480.9969590.9970570.9966650.9975480.997254
Precision0.9954000.9950070.9959830.9944220.9961800.993745
Brier Score Loss0.0042200.0048060.0041610.0053330.0037510.005392
Balanced Accuracy0.9953520.9947670.9955440.9941830.9959350.993967
ROC-AUC0.9994790.9988760.9994290.9993030.9997260.998885
Table 4. Test results.
Table 4. Test results.
ModelAug.TPFPFNTNAccuracyPrecisionRecallF1 ScoreROC-AUC
EfficientNetV2-SNo6569353898490.9955730.9955740.9955730.9955740.999037
Yes6569354198460.9953910.9953920.9953910.9953920.998984
ConvNeXt-TinyNo6564403898490.9952700.9952700.9952700.9952700.998387
Yes6567374198460.9952700.9952710.9952700.9952700.998546
DenseNet121No6568363598520.9956950.9956950.9956950.9956950.999136
Yes6570344298450.9953910.9953930.9953910.9953920.999081
ResNet18No6574303398540.9961800.9954570.9950050.9952310.995985
Yes6564403298550.9956340.9939430.9951490.9945450.995553
ResNet34No6575293298550.9963010.9956090.9951570.9953830.996111
Yes6577273898490.9960580.9959120.9942550.9950830.995761
ResNet50No6573313398540.9961190.9953060.9950050.9951550.995934
Yes6562422998580.9956950.9936400.9956000.9946190.995679
Table 5. Classification results for the AP/PA validation dataset using the ResNet34 model.
Table 5. Classification results for the AP/PA validation dataset using the ResNet34 model.
MetricValues
Train Loss0.009089
Valid Loss0.004667
Error Rate0.001176
Accuracy0.998824
F1 Score0.999018
Cohen’s Kappa0.997554
Recall0.999607
Precision0.998430
Brier Score Loss0.998430
Balanced Accuracy0.998633
ROC-AUC0.999977
Table 6. Test results of the developed model on the test dataset (True AP, False PA accepted).
Table 6. Test results of the developed model on the test dataset (True AP, False PA accepted).
TPFPFNTNAccuracyPrecisionRecallF1 ScoreROC-AUC
6575292998580.996490.995610.995610.995610.99935
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gökçimen, F.; İnner, A.B.; Çakır, Ö. Determination of Anteroposterior and Posteroanterior Imaging Positions on Chest X-Ray Images Using Deep Learning. Eng. Proc. 2025, 104, 58. https://doi.org/10.3390/engproc2025104058

AMA Style

Gökçimen F, İnner AB, Çakır Ö. Determination of Anteroposterior and Posteroanterior Imaging Positions on Chest X-Ray Images Using Deep Learning. Engineering Proceedings. 2025; 104(1):58. https://doi.org/10.3390/engproc2025104058

Chicago/Turabian Style

Gökçimen, Fatih, Alpaslan Burak İnner, and Özgür Çakır. 2025. "Determination of Anteroposterior and Posteroanterior Imaging Positions on Chest X-Ray Images Using Deep Learning" Engineering Proceedings 104, no. 1: 58. https://doi.org/10.3390/engproc2025104058

APA Style

Gökçimen, F., İnner, A. B., & Çakır, Ö. (2025). Determination of Anteroposterior and Posteroanterior Imaging Positions on Chest X-Ray Images Using Deep Learning. Engineering Proceedings, 104(1), 58. https://doi.org/10.3390/engproc2025104058

Article Metrics

Back to TopTop