Next Article in Journal
Attention-Based Multi-Agent RL for Multi-Machine Tending Using Mobile Robots
Previous Article in Journal
Convolutional Neural Network for Automatic Detection of Segments Contaminated by Interference in ECG Signal
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Approaches for Classifying Aviation Safety Incidents: Evidence from Australian Data

1
School of Engineering and Technology, University of New South Wales, Canberra, ACT 2600, Australia
2
Capability Systems Centre, University of New South Wales, Canberra, ACT 2610, Australia
3
School of Science, University of New South Wales, Canberra, ACT 2612, Australia
*
Author to whom correspondence should be addressed.
AI 2025, 6(10), 251; https://doi.org/10.3390/ai6100251
Submission received: 28 August 2025 / Revised: 25 September 2025 / Accepted: 29 September 2025 / Published: 1 October 2025
(This article belongs to the Topic Big Data and Artificial Intelligence, 3rd Edition)

Abstract

Aviation safety remains a critical area of research, requiring accurate and efficient classification of incident reports to enhance risk assessment and accident prevention strategies. This study evaluates the performance of three deep learning models, BERT, Convolutional Neural Networks (CNN), and Long Short-Term Memory (LSTM) for classifying incidents based on injury severity levels: Nil, Minor, Serious, and Fatal. The dataset, drawn from ATSB records covering the years 2013 to 2023, consists of 53,273 records and was used. The models were trained using a standardized preprocessing pipeline, with hyperparameter tuning to optimize performance. Model performance was evaluated using metrics such as F1-score accuracy, recall, and precision. Results revealed that BERT outperformed both LSTM and CNN across all metrics, achieving near-perfect scores (1.00) for precision, recall, F1-score, and accuracy in all classes. In comparison, LSTM achieved an accuracy of 99.01%, with strong performance in the “Nil” class, but less favorable results for the “Minor” class. CNN, with an accuracy of 98.99%, excelled in the “Fatal” and “Serious” classes, though it showed moderate performance in the “Minor” class. BERT’s flawless performance highlights the strengths of transformer architecture in processing sophisticated text classification problems. These findings underscore the strengths and limitations of traditional deep learning models versus transformer-based approaches, providing valuable insights for future research in aviation safety analysis. Future work will explore integrating ensemble methods, domain-specific embeddings, and model interpretability to further improve classification performance and transparency in aviation safety prediction.

1. Introduction

Aviation safety is a critical area of research aimed at ensuring the protection of passengers, crew members, and the broader aviation industry [1]. Each year, aviation accidents and incidents ranging from minor operational disruptions to catastrophic failures are systematically recorded by various regulatory and investigative bodies. The Australian Transport Safety Bureau (ATSB) [2], one of the key agencies responsible for aviation safety oversight, collects and documents detailed reports of aviation occurrences, including structured data and unstructured textual narratives. These incident narratives provide invaluable insights into the underlying causes and contributing factors of aviation safety events, thereby informing improvements in aviation safety management systems. However, the sheer volume of these unstructured textual reports presents significant challenges in extracting actionable insights efficiently. Traditional manual analysis is both labor-intensive and time-consuming, necessitating the development of automated approaches capable of processing large-scale incident data effectively [3,4].
The International Civil Aviation Organization (ICAO) has emphasized the role of artificial intelligence (AI) and machine learning (ML) in advancing aviation safety through predictive analytics, automated incident classification, and enhanced monitoring capabilities [5]. Traditional approaches to analyzing aviation safety data, such as manual reviews of incident reports, are increasingly inadequate given the exponential growth in data volume. For instance, the ATSB dataset is continuously expanding, making it impractical to classify and analyze each report manually in real time. To address this challenge, recent advances in natural language processing (NLP) have enabled the automation of text classification tasks, allowing for more efficient and accurate categorization of aviation safety occurrences.
This study leverages three state-of-the-art deep learning models, namely Bidirectional Encoder Representations from Transformers (BERT), Convolutional Neural Networks (CNN), and Long Short-Term Memory (LSTM), to automate the classification of aviation incident reports. By efficiently categorizing incidents based on injury severity levels (Nil, Minor, Serious, and Fatal), these models can provide aviation stakeholders, such as safety investigators and regulators, with timely insights that support proactive risk mitigation, policy formulation, and resource allocation.
The primary objective of this study is to evaluate the effectiveness of BERT, CNN, and LSTM in classifying aviation safety occurrences based on unstructured textual narratives. Specifically, the study seeks to answer the following research questions:
  • How accurately can BERT, CNN, and LSTM classify aviation safety incidents into predefined injury severity categories (Nil, Minor, Serious, Fatal) using unstructured textual data?
  • How does the performance of BERT compare to CNN and LSTM in terms of classification accuracy, recall, and precision when applied to aviation safety reports?
  • What are the relative strengths and limitations of each model in processing aviation incident narratives, and which model demonstrates the highest effectiveness in automating classification tasks?
The rest of the paper is organized in the following manner: Section 2 presents a comprehensive review of related literature, examining prior research on air transport safety analysis, the use of machine learning approaches, and deep learning methodologies for text classification. Section 3 outlines the methodological framework, detailing the dataset, preprocessing steps, model architecture, training procedures, and evaluation metrics. Section 4 reports the experimental findings, providing a comparative analysis of the performance of CNN, LSTM, and BERT models on the aviation dataset. Section 5 critically discusses the results, including an ablation study and the limitations of the study. Finally, Section 6 concludes the paper by summarizing the key contributions, discussing the broader implications of the findings, and proposing future research directions for enhancing aviation safety analytics through advanced deep learning approaches.

2. Related Work

The automation of aviation safety occurrence classification has emerged as a critical area of research within the fields of ML and NLP. Over the past two decades, significant advancements have been made in leveraging NLP techniques to analyze large-scale safety datasets [4,6,7]. This section provides a comprehensive review of key developments in this domain, highlighting studies that have employed ML methodologies, including deep learning models, CNNs, LSTM networks, and transformer-based architectures such as BERT. Furthermore, it examines comparative analyses of these approaches in the context of incident report classification, evaluating their effectiveness in extracting meaningful patterns from unstructured textual data.

2.1. Machine Learning Approaches for Aviation Safety Data

The application of ML algorithms to aviation safety data can be traced back to the early 2000s, with initial efforts centered on utilizing decision trees and support vector machines (SVMs) to predict accident severity and identify contributing factors based on structured data [8,9,10]. While these models showed promise in automating data organization, their effectiveness was limited in handling unstructured textual data, such as incident narratives. Consequently, more recent research has shifted towards the development and implementation of advanced NLP techniques, enabling a more comprehensive analysis of aviation safety reports by extracting meaningful insights from unstructured text [11,12].

2.2. Deep Learning Models for Text Classification

The integration of deep learning techniques has significantly advanced the classification of unstructured textual data, including aviation incident narratives. Models such as Recurrent Neural Networks (RNNs), LSTMs, and CNNs have been extensively explored due to their ability to capture contextual relationships and model long-range dependencies within textual data [13].
Kim and Jeong [14] pioneered a CNN-based model for sentence classification, demonstrating its effectiveness in extracting hierarchical text features. This study established CNNs as a powerful tool for document classification, a finding subsequently reinforced by Harley et al. [15], who validated CNNs’ applicability across various domains, including aviation safety. Coelho et al. [16] further explored CNN in the classification of aviation accident reports by severity level, where the model exhibited competitive accuracy compared to traditional machine learning approaches, such as SVMs and random forests.
Beyond CNNs, LSTMs, an advanced variant of RNNs, have been increasingly applied to aviation safety data due to their capability to process sequential information effectively. Another study demonstrated that LSTMs outperform conventional classifiers by capturing the temporal dependencies inherent in aviation incident reports [17]. Similarly, Zhang [18] employed LSTMs to categorize aviation incidents by severity, revealing superior performance over traditional ML models, such as SVMs.
Another study examined the classification of aviation incidents into Commercial, Military, and Private categories using the Socrata aviation dataset, which comprises 4864 records. The study employed BLSTM, CNN, LSTM, and simple RNN models, evaluating their performance through classification reports, confusion matrices, accuracy metrics, and validation loss and accuracy curves. Of all the models evaluated, BLSTM recorded the top accuracy rate of 72%, demonstrating superior stability and balanced performance across the categories. LSTM followed closely with an accuracy of 71%, excelling in recall for the Commercial category. Conversely, CNN and sRNN achieved lower accuracies of 67% and 69%, respectively, with notable misclassifications in the Private category. While BLSTM and LSTM exhibited strong performance in handling sequential dependencies and complex classification tasks, all models encountered challenges related to class imbalance, particularly in distinguishing Military and Private incidents [19].

2.3. Transformer Models in Text Classification

Transformer models, particularly BERT, have significantly advanced the field of NLP by introducing pre-trained models that can be fine-tuned for specific tasks. Unlike traditional RNN-based architectures, BERT is designed to predict missing words in a sentence, thereby capturing bidirectional contextual relationships within textual data. This capability has led to substantial improvements across various NLP tasks, including text classification, sentiment analysis, and named entity recognition [20,21,22,23].
Further evidence supporting the efficacy of transformer-based models in aviation safety research is provided by Kierszbaum et al. [24] who explored the application of such models in analyzing aviation incident reports. Using the Aviation Safety Reporting System (ASRS) dataset, which comprises English-language incident reports characterized by extensive use of specialized terminology, abbreviations, and aviation vocabulary, the study reframed the task of incident report analysis as a series of natural language understanding (NLU) tasks. Their experimental framework demonstrated the potential of transformer models to enhance safety analysts’ ability to process and categorize incident reports efficiently [24].
Beyond aviation safety, extensive research has validated the superiority of BERT-based models in various text classification applications. González and Garrido conducted a comparative study between BERT and traditional ML techniques utilizing Term Frequency-Inverse Document Frequency (TF-IDF) features. Their experiments highlighted the consistent superiority of BERT, which achieved an accuracy of 99.71% using BERT-large and BERT-base configurations, reinforcing its effectiveness as a default approach for NLP-related classification problems [25].
The adaptability of BERT has also been explored in domains such as fake news detection and social media analysis. Another study examined fake news classification using a BERT-based model combined with an LSTM layer. Their study, conducted on the FakeNewsNet dataset, demonstrated a 2.50% and 1.10% improvement in accuracy on the PolitiFact and GossipCop datasets, respectively, compared to a vanilla pre-trained BERT model [26]. Similarly, Kokab et al. evaluated a BERT-based Convolutional Bidirectional Recurrent Neural Network (CBRNN) model for sentiment analysis of social media data. Their results indicated that the CBRNN model achieved an accuracy of 97% and an Area Under the Curve (AUC) value of 0.989, outperforming Word2Vec-based LSTM models, particularly in terms of precision and robustness [27].
In the domain of traffic safety, transformer models have also demonstrated remarkable efficacy. Oliaee et al. [28] applied BERT to classify traffic injury types using a dataset comprising over 750,000 unique crash narratives. Their model attained a high accuracy of 84.2% and an AUC of 0.93 ± 0.06 per class, underscoring the potential of BERT-based models to assist safety engineers and analysts in understanding crash causality [28].
Collectively, these studies highlight the transformative impact of BERT and other transformer-based architectures in text classification across diverse domains. Their ability to capture intricate semantic relationships within text, coupled with superior classification performance, has positioned them as a powerful tool for enhancing automated analysis in aviation safety, misinformation detection, and traffic safety research.

2.4. Comparative Studies of Machine Learning Models in Aviation Safety

Several studies have systematically evaluated the performance of various machine learning models in the classification of aviation safety reports. A review study was conducted on a state of art NLP using traditional deep learning models, such as CNNs and LSTM networks and transformers. Their findings indicated that the transformer model can achieve superior classification accuracy compared to LSTM and CNN approaches [29]. Another study focused on comparing the transformer with LSTM and CNN to improve text classification on the transformer, and the performance shows that BERT outperforms other models [30].
Building on these insights, Gao et al. conducted a comparative study evaluating CNNs, LSTMs, and transformer-based models, such as BERT, for the classification of aviation safety narratives. Their results demonstrated that while LSTMs effectively captured sequential dependencies within textual data, BERT-based models exhibited superior performance in terms of classification accuracy and interpretability. The study highlighted the advantages of leveraging pre-trained transformers, which benefit from large-scale corpus training and fine-tuning capabilities for domain-specific applications in aviation safety analysis [31].
Despite advancements in ML and NLP, classifying aviation safety reports remains a complex task due to the unstructured nature of the narratives and the extensive use of its terminology. One of the primary challenges is the specialized aviation lexicon, which necessitates domain-adaptive models or additional pre-processing techniques to enhance classification accuracy [32]. Furthermore, the imbalanced distribution of safety incident categories, where certain injury severity levels are significantly underrepresented, presents a challenge for training models that generalize effectively across all classes.
The growing body of research underscores the increasing adoption of deep learning models, particularly CNNs, LSTMs, and transformer-based architectures, for classifying aviation safety reports. While CNNs and LSTMs have demonstrated effectiveness in specific contexts, transformer models, such as BERT, consistently outperform traditional and deep learning methods in text classification tasks. Nevertheless, challenges related to domain-specific vocabulary persist, necessitating further research to develop more robust techniques for addressing these limitations and improving the applicability of machine learning models in real-world aviation safety analysis.

3. Materials and Methods

3.1. Data Acquisition

Various Aviation incident and accident investigation reports serve as critical sources of information for analyzing safety occurrences and identifying contributing factors. These reports are systematically compiled and published by various aviation safety agencies, including the ATSB, the Aviation Safety Network (ASN) National Transportation Safety Board (NTSB). For this study, the dataset was derived from ATSB’s aviation incident records, as it provides detailed textual narratives and injury severity classifications necessary for understanding accident characteristics and trends. The dataset encompasses safety reports recorded in Australia over ten years, from 1 January 2013, to 31 December 2023, resulting in a collection of 53,273 records following preprocessing and data cleaning. Given the objective of classifying aviation safety occurrences based on text narratives, the study focused on extracting two key fields: the “Summary”, which provides a detailed account of the incident narratives, and the “Injury Level” classification, which categorizes incidents as ‘Nil,’ ‘Minor,’ ‘Serious,’ or ‘Fatal.’ These categories are needed for assessing the severity of aviation occurrences and evaluating the effectiveness of ML models in distinguishing between different levels of risk. The dataset was acquired directly from ATSB investigation authorities, and labels were assigned by human experts and reflect the actual outcomes of the incidents, ensuring the reliability and authenticity of the information used for model training and evaluation.

3.2. Data Pre-Processing

To ensure the integrity and quality of the data for machine learning model training, a rigorous preprocessing pipeline was employed. Textual narratives were standardized by removing special characters, excessive whitespace, and punctuation, followed by conversion to lowercase to ensure uniformity. Tokenization was then applied to segment the text into individual words, and stopword removal technique was employed to filter out common, non-informative terms while retaining critical aviation-related vocabulary. Unlike standard stopword filtering, this tailored approach preserved necessary terminology relevant to aviation safety analysis. Lemmatization techniques were used to convert terms to their root form, minimizing redundancy while maintaining semantic meaning as shown in Figure 1. To mitigate the uneven class distribution present within the dataset, loss functions with class weights were applied during model training [33]. The textual narratives were transformed into numerical embeddings using TF-IDF and pre-trained word embeddings such as BERT, enabling the models to capture both semantic relationships and contextual information [34]. The dataset was divided into training, validation, and test subsets using stratified sampling to maintain proportional representation of injury levels across all partitions, with 80% allocated for training, 20% for testing, and a tenth of the training data reserved for validation at every epoch to monitor model performance and prevent overfitting [35]. All reports were converted into numerical sequences of fixed length (2000 words), with narratives shorter than this length padded with zeros and longer narratives truncated. The corpus vocabulary was limited to 100,000 unique terms, reducing input complexity while retaining key vocabulary.

3.3. Model Architectures

In this study, three distinct deep learning architectures were utilized for the classification of aviation safety data: BERT, CNNs, and LSTMs. Each model offers unique strengths in handling the complexities of textual data, and their application is necessary for capturing various features of the narratives to enhance the accuracy and reliability of structured data classification. A brief description of each model’s architecture follows.

3.3.1. BERT

BERT [20] is a transformer-based model that has revolutionized NLP by leveraging a bidirectional approach to context understanding. Unlike traditional models that process text in a left-to-right or right-to-left manner, BERT considers the full context of each word in a sentence by analyzing both directions simultaneously. This bidirectional approach significantly enhances the model’s understanding of the nuances and relationships between words in each context, making it particularly powerful for tasks such as text classification, question answering, and named entity recognition. In the context of aviation safety data, BERT was fine-tuned for classification tasks using the pre-trained BERT-Base model. The fine-tuning process involves training the model on a smaller, domain-specific dataset, enabling it to adapt to the linguistic characteristics of aviation safety narratives. The model was trained with a learning rate of 3 × 10−5 and optimized using the AdamW optimizer [36]. This configuration enabled BERT to capture complex dependencies in the textual data, improving its accuracy and performance in the classification task.

3.3.2. CNN

CNN [37] is a class of deep learning models that have demonstrated exceptional performance in image classification tasks. However, their application has extended to NLP tasks due to their ability to capture local patterns within sequential data. CNNs operate by applying convolutional filters across the input data, which allows them to detect features such as n-grams and syntactic structures within the text. In NLP applications, CNNs treat the input text as a sequence of word embeddings, where the convolutional layers detect significant features that help with classification tasks. In this study, CNNs were employed to identify localized features within incident report texts. The model configuration involved applying filters of different sizes, with a kernel size of 5 and 64 filters yielding the best performance. Additionally, MaxPooling was used to reduce the dimensionality of the feature maps while preserving important spatial features. CNNs have the advantage of being computationally efficient and effective in capturing short-range dependencies in the text, making them a suitable choice for identifying specific patterns in aviation incident reports.

3.3.3. LSTMs

LSTM [38] is a type of RNN designed to address the issue of vanishing gradients, which is a common problem in traditional RNNs when learning long-range dependencies in sequential data. LSTMs include memory cells that allow the model to retain information for long periods, making them ideal for tasks where the sequential order of data is important, such as language modeling, speech recognition, and time series forecasting. In this study, LSTMs were used to model the sequential nature of aviation safety incident reports, where the order of events and their relationships over time are necessary for accurate classification. The LSTM architecture in this study was configured with two layers, each containing 128 hidden units. A dropout rate of 0.3 was applied to prevent overfitting, ensuring that the model could generalize well with unseen data. LSTMs excel at capturing long-term dependencies and contextual relationships within the text, making them particularly effective in understanding the progression of events in aviation safety narratives.
Each of these architectures was evaluated for its ability to classify aviation safety data effectively, with the specific characteristics of the models making them suitable for different aspects of the text data. While BERT excels in understanding deep contextual relationships, CNNs are adept at detecting local patterns, and LSTMs are particularly powerful in capturing long-term dependencies. By combining these diverse model architectures, this study aimed to leverage the strengths of each to achieve optimal performance in classifying aviation safety data.

3.4. Experimental Setup

The experiments in this work were implemented using Python (version 3.8.10), drawing on multiple libraries to support preprocessing, model development, assessment, and visualization. Deep learning architecture such as LSTM and CNN was built and trained with TensorFlow (version 2.10.0) and Keras (version 2.10.0). Fine-tuning of the BERT model was performed using the Transformers library (version 4.48.0) from Hugging Face. For the evaluation phase, Scikit-learn (version 1.6.1) was used to generate accuracy scores and classification summaries. Data handling and numerical operations were facilitated by Pandas (version 1.5.0) and NumPy (version 1.23.4), ensuring consistent processing throughout the workflow. Performance visualizations were produced with Matplotlib (version 3.6.1) and Seaborn (version 0.12.0), providing a clear interpretation of results. Hyperparameter tuning was supported by Optuna (version 4.2.1), which enabled systematic refinement of model configurations. PyTorch (version 2.1.0+cpu) was utilized for transformer-related tasks. All experiments were executed in a Jupyter Notebook environment on a Linux-based server running Ubuntu (version 5.4.0-169-generic), equipped with 256 CPU cores and 256 GB of RAM, ensuring both computational efficiency and reproducibility.
Computational Efficiency
To provide insights into the practical deployment of the models, the training times for the final configurations were recorded as follows: BERT ~12 h 03 min (3 epochs), CNN ~1 h 45 min (10 epochs), and LSTM ~2 h 17 min (10 epochs). Other hyperparameter tuning runs were not timed. The inference time per report is approximate, and memory usage was measured on a Linux server equipped with 256 CPU cores and 256 GB RAM. These metrics highlight the computational resources and efficiency associated with each model, offering guidance for real-world application and deployment.
Hyperparameter tuning
The hyperparameter tuning process was systematically conducted to optimize the performance of the LSTM, CNN, and BERT models on the aviation safety dataset, as detailed in Table 1. For the LSTM model, a range of layer configurations (1, 2, and 3 layers) was evaluated, with two layers providing the most favorable performance. The hidden unit size was tested at 64, 128, and 256 units, with 128 units proving to be optimal. Dropout rates of 0.2, 0.3, and 0.5 were explored, with 0.3 offering the best balance between regularization and model capacity. The Adam optimization algorithm was employed, using a 0.0005 learning rate, which provided the most stable convergence. For the CNN model, the number of filters was adjusted between 32, 64, and 128, with 64 filters yielding superior results. The kernel size was tested with values of 3, 5, and 7, where a kernel size of 5 demonstrated the highest performance. MaxPooling was selected over AveragePooling, and a 30% dropout was maintained to prevent overfitting. The optimal CNN configuration also employed the Adam optimizer with a 0.001 learning rate and batches of 32 samples. In the case of the BERT model, fine-tuning was performed using the BERT-Base architecture, where batch sizes of 8, 16, and 32 were examined, with 16 being the most effective. A range of learning rates (3 × 10−5, 5 × 10−5, and 1 × 10−4) was tested, and 3 × 10−5 resulted in the best model performance. Training was conducted for 5 iterations, with a maximum sequence size of 256, and the AdamW optimizer was chosen for its superior performance during fine-tuning. Early stopping was applied during training to prevent overfitting, and the test set was strictly held out and never accessed during the tuning process. Since a tenth of the training data was kept aside for validation during each epoch, hyperparameter selection was guided solely by validation performance, ensuring that the final evaluation on the test set remained unbiased and fully independent. This comprehensive hyperparameter tuning process was instrumental in ensuring that each model was optimally trained, effectively balancing performance with computational efficiency.

3.5. Performance Metrics

The models’ performance was assessed using several key metrics: Precision, Recall, F1-score, and Accuracy. Precision evaluates the proportion of correctly predicted positive instances relative to the total predicted positives, providing insight into the model’s ability to minimize false positives. Recall, conversely, quantifies the fraction of actual positive instances accurately identified by the model, reflecting its sensitivity in detecting positive patterns. The F1-score, which is the harmonic mean of Precision and Recall, offers a balanced metric that is particularly advantageous when handling imbalanced datasets. Accuracy, representing the overall correctness of the model, indicates the proportion of correctly predicted instances across both positive and negative classes. In addition, the confusion matrix functions as a diagnostic tool by comparing actual and predicted values, enabling a visualization of the model’s performance and categorizing instances into four groups: True Positives, True Negatives, False Positives, and False Negatives [39]. Together, these metrics provide an in-depth assessment of the model’s effectiveness for categorizing air transport safety records, as detailed in Table 2 and Table 3.

4. Results

The performance of the BERT, CNN, and LSTM models was evaluated on the test dataset using several key evaluation metrics: Precision, Recall, F1-score, and Accuracy. The results of these metrics for each model are summarized in Table 4.
The performance of the models, LSTM, CNN, and BERT, was assessed through several evaluation metrics, as presented in Table 4. The table highlights the Precision, Recall, F1-Score, and Accuracy for each model across the four classes: “Nil,” “Minor,” “Fatal,” and “Serious.” The LSTM model demonstrated strong overall performance, with a test accuracy of 0.9901. It exhibited high precision and recall values for the “Nil” class, achieving 0.9942 and 0.9964, respectively. However, its performance was less favorable for the “Minor” class, with precision and recall values of 0.7452 and 0.7178, respectively, yielding an F1-score of 0.7312. The CNN model, with an accuracy of 0.9899, showed a similar pattern of strong performance in the “Nil” class (precision 0.9955, recall 0.9947) but moderate performance for the “Minor” class (precision 0.7006, recall 0.7607), resulting in an F1-score of 0.7294. Notably, the CNN model performed excellently in the “Fatal” and “Serious” classes, with F1-scores of 0.8675 and 0.8866, respectively. The BERT model, which achieved near-perfect scores for precision, recall, F1-score, and accuracy in all classes, demonstrated flawless performance with an accuracy of 1.00. This is indicative of BERT’s superior classification ability compared to the LSTM and CNN models, particularly in handling the overall dataset.
In addition to the evaluation metrics, the validation loss and validation accuracy plots further underscore the models’ performance during training and validation, as shown in Figure 2. The LSTM and CNN models showed relatively stable validation loss and accuracy trends, indicating consistent learning and generalization. Conversely, the BERT model exhibited minimal fluctuations in both validation loss and accuracy, reflecting its ability to consistently achieve optimal classification results. These trends suggest that BERT outperforms both LSTM and CNN in terms of generalization and the ability to minimize overfitting during training, as evidenced by its high validation results. The convergence of the validation metrics indicates that BERT maintained its robustness across epochs, achieving an ideal balance between fitting the training data and generalizing to unseen instances.

Evaluation of BERT Model Performance

The performance of these models was evaluated using multiple metrics, including precision, recall, F1-score, and accuracy. As shown in Table 4, the BERT model shows near-perfect values (1.00) for each class “Nil”, “Minor”, “Fatal”, and “Serious” across all evaluation metrics. This indicates the model’s exceptional performance in correctly classifying all instances within the test set. Figure 3, Figure 4 and Figure 5 display the confusion matrix for the BERT, LSTM and CNN models, respectively, offering a visual representation of the model’s classification performance.

5. Ablation Study

This study examines the contributions of key architectural components in the BERT, CNN, and LSTM models to the classification performance on the aviation safety dataset. By systematically modifying or removing critical layers, we assess their impact on accuracy, precision, recall, and F1-score, as shown in Table 5, providing insights into the strengths and limitations of each model in processing aviation safety narratives.

5.1. LSTM Ablation

The LSTM model consists of an embedding layer, an LSTM layer, and fully connected layers, each playing a fundamental role in capturing the sequential dependencies within aviation safety narratives. To evaluate the impact of the LSTM layer, we replaced it with a simple dense layer, resulting in a substantial decline in classification performance, with accuracy dropping from 0.9901 to a significantly lower value. Furthermore, reducing the number of LSTM units from 128 to 64 led to a decrease in recall for minority classes such as “Fatal” and “Minor,” indicating that a higher number of LSTM units enhances the model’s ability to learn complex temporal patterns. These findings reinforce the importance of recurrent connections in effectively modeling aviation safety incidents.

5.2. CNN Ablation

The CNN model leverages convolutional and max-pooling layers to extract key features from textual narratives. To investigate the contribution of convolutional layers, we replaced them with a multi-layer perceptron (MLP) architecture, leading to a noticeable performance drop, with accuracy decreasing from 0.9899 in the original CNN model. The recall for the “Nil” class, which comprises the majority of cases, saw a significant reduction, highlighting the convolutional layers’ effectiveness in extracting hierarchical features. Additionally, reducing the number of filters from 128 to 64 negatively impacted precision for the “Serious” and “Fatal” categories, demonstrating that a higher filter count improves the model’s ability to capture intricate text patterns. Removing max-pooling layers also led to minor performance degradation, confirming their role in feature selection and dimensionality reduction, which are necessary for efficient classification.

5.3. BERT Ablation

BERT’s architecture relies on attention mechanisms and transformer layers to process contextual dependencies in text classification. To evaluate their significance, we removed the attention layers, leaving only the token embeddings, which caused a notable decrease in performance, with accuracy dropping from 1.00 (achieved in three epochs) to a significantly lower value. The “Fatal” class exhibited a sharp decline in recall, demonstrating that attention layers play a vital role in focusing on relevant textual features. Additionally, reducing the transformer depth from twelve to six layers resulted in lower recall for the “Serious” and “Fatal” categories, emphasizing the necessity of deeper transformer networks for capturing complex aviation incident patterns. Lastly, replacing BERT’s WordPiece tokenizer with a standard word-level tokenizer led to decreased performance, particularly in the “Minor” class, highlighting the effectiveness of subword tokenization in handling vocabulary and rare words. Notably, while BERT achieved the highest score within three epochs, its performance decreased to 0.9760 by epoch five, indicating a potential overfitting effect or sensitivity to extended training. These results affirm BERT’s architectural advantages in processing aviation safety narratives with high accuracy and interpretability.

5.4. Discussion

The ablation experiments provide valuable insights into the contributions of different components within each model, particularly in the classification of aviation safety occurrences. The results demonstrate that BERT significantly outperforms LSTM and CNN across multiple evaluation metrics, highlighting the advantages of transformer-based architectures in modeling complex textual data [40]. BERT’s self-attention mechanism effectively captures long-range dependencies, allowing it to discern intricate contextual relationships within safety reports. This capability is particularly advantageous for classifying less frequent injury categories, such as “Fatal” and “Minor,” where sequential models like LSTM exhibit limitations due to their reliance on past hidden states [41]. While LSTM leverages its recurrent structure to maintain context across sequences, it struggles with rare categories, likely due to vanishing gradient issues and its dependency on prior tokens for information retention. Conversely, CNN demonstrated efficiency in identifying localized patterns but failed to capture long-range dependencies effectively, leading to lower recall for injury levels with limited representation [20]. In comparison, the gradual decrease in validation loss in LSTM and CNN models suggests better resilience against overfitting, making them potentially more adaptable to diverse datasets.
While BERT’s flawless performance positions it as a promising candidate for high-accuracy applications, such as in aviation safety [42], its computational complexity may restrict its deployment in resource-constrained environments. In contrast, LSTM and CNN, with their comparatively efficient architectures, present a feasible alternative for real-world implementation where computational resources are limited. Furthermore, dataset class imbalance likely contributed to BERT’s tendency to favor the majority class, potentially skewing classification outcomes [43,44]. Addressing this issue in future work through techniques such as oversampling, cost-sensitive learning, or synthetic data augmentation could mitigate bias and enhance model fairness. Additionally, despite BERT’s high predictive classification accuracy, interpretability remains a challenge, particularly in domains requiring transparent decision-making, such as safety-critical applications. Techniques such as model distillation or attention-based visualization could enhance BERT’s explainability, fostering trust in its predictions and enabling broader adoption in real-world aviation safety assessments [45]. Furthermore, incorporating techniques such as data augmentation, class balancing, and domain-specific pretraining may enhance model generalizability, particularly for underrepresented injury categories in highly imbalanced datasets [46].
Beyond methodological insights, these findings have practical implications for aviation safety authorities. For instance, models such as BERT could support real-time monitoring of incident reports, automatically flagging high-risk cases (e.g., “Serious” or “Fatal”) for rapid review by safety investigators. At the same time, computationally lighter models like CNN or LSTM, though less accurate, may be more feasible for deployment in resource-constrained environments, ensuring timely screening while balancing efficiency and accuracy.

5.5. Limitations

While this study offers significant insights into the use of deep learning approaches in aviation safety categorization tasks, several limitations must be acknowledged. First, the dataset employed comprises safety reports drawn from ATSB records covering incidents from 2013 to 2023. As a result, the findings may not generalize to datasets from other aviation authorities or international contexts, where variations in linguistic structures, reporting standards, and terminology could influence model performance [47]. Additionally, the substantial class imbalance in the dataset, where the “Nil” injury level comprises many cases, may have disproportionately influenced model performance, particularly for underrepresented injury levels such as “Fatal” and “Minor” [4]. While techniques such as oversampling and undersampling were not employed in this study, their incorporation in future work could significantly mitigate class imbalance issues and improve classification accuracy for rare events [48].
Another key limitation concerns the evaluation of the model’s performance using traditional classification metrics such as precision, recall, and F1-score; it does not account for the real-world implications of misclassifications. Given the high-stakes nature of aviation safety, misclassifying severe injury cases such as “Fatal” or “Serious” incidents could have critical consequences [49]. Future research should incorporate risk-aware evaluation frameworks that prioritize minimizing false negatives for high-risk categories, ensuring that classification models align with the safety-critical nature of aviation operations [50].

6. Conclusions

This study presents a comprehensive comparison of BERT, LSTM, and CNN models for the classification of aviation safety data. The results demonstrate the superior performance of the BERT model, which outperforms both LSTM and CNN across all evaluation metrics, including precision, recall, F1-score, and accuracy. While BERT achieved near-perfect classification accuracy on the ATSB dataset, practical deployment considerations are important. Its computational complexity may limit applicability in resource-constrained environments, and model compression strategies such as knowledge distillation, pruning, or quantization could be employed to reduce size and inference time. Models may be deployed on premises or via cloud infrastructure, depending on available resources. LSTM and CNN, while not achieving the same level of performance, offer a more efficient alternative in scenarios where computational resources are limited.
Despite the promising results, several factors warrant further exploration. Future work should focus on testing these models on larger and more diverse datasets to better reflect real-world conditions. Additionally, addressing potential issues such as class imbalance and model interpretability is necessary for enhancing the reliability and trustworthiness of the models. Techniques like model distillation and attention visualization could provide greater transparency into BERT’s decision-making process, which is especially important in high-stakes domains such as aviation safety. Evaluating model generalization across datasets from other aviation authorities will also ensure robustness in varied operational contexts. Overall, this study contributes to the expanding research on the use of advanced machine learning methods in safety-critical applications, offering valuable insights for both academic and practical advancements in the field.

Author Contributions

A.N. was responsible for methodology, conceptualization, data curation, software development, formal analysis, validation, and drafting the original manuscript. U.T. and K.F.J. contributed to reviewing and editing the manuscript, while G.W. oversaw data collection and supervision. All authors have read and approved the submitted manuscript.

Funding

This work was supported by the UNSW Tuition Fees Scholarship (TFS).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets analyzed in this research were obtained from the Australian Transport Safety Bureau (ATSB) and are accessible under a Creative Commons Attribution 3.0 Australia license through the ATSB.

Acknowledgments

The authors sincerely thank the ATSB authorities for providing the dataset, which was crucial for conducting this study.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding this publication.

Abbreviations

The following abbreviations are used throughout this manuscript:
AIArtificial Intelligence
AUCArea Under the Curve
ASNAviation Safety Network
ASRSAviation Safety Reporting System
ATSBAustralian Transport Safety Bureau
BiLSTMBidirectional Long Short-Term Memory
BERTBidirectional Encoder Representations from Transformers
CBRNNConvolutional Bidirectional Recurrent Neural Network
CNNConvolutional Neural Network
GRUGated Recurrent Unit
ICAOInternational Civil Aviation Organization
LSTMLong Short-Term Memory
MLMachine Learning
NLPNatural Language Processing
NLUNatural Language Understanding
NTSBNational Transportation Safety Board
RNNRecurrent Neural Network
SVMSupport Vector Machine
TF-IDFTerm Frequency-Inverse Document Frequency

References

  1. McMahon, T.M. The cause of Boeing 737s significantly higher number of runway excursions opposed to the Airbus A32X. In Proceedings of the 2023 Regional Student Conferences, Multiple Locations, 1 January 2023; p. 77426. [Google Scholar]
  2. Hood, G. A balancing act: The Australian Transport Safety Bureau’s approach to accident investigations on airports. J. Airpt. Manag. 2019, 13, 354–360. [Google Scholar] [CrossRef]
  3. Korba, P.; Šváb, P.; Vereš, M.; Lukáč, J. Optimizing aviation maintenance through algorithmic approach of real-life data. Appl. Sci. 2023, 13, 3824. [Google Scholar] [CrossRef]
  4. Aziida, N.; Keith, J.; Ugur, T.; Graham, W. Does the Choice of Topic Modeling Technique Impact the Interpretation of Aviation Incident Reports? A Methodological Assessment. Technologies 2025, 13, 209. [Google Scholar] [CrossRef]
  5. Ludwig, W. International Civil Aviation Organization (ICAO); Torrossa: Fiesole, Italy, 2023. [Google Scholar]
  6. Yang, C.; Huang, C. Natural language processing (NLP) in aviation safety: Systematic review of research and outlook into the future. Aerospace 2023, 10, 600. [Google Scholar] [CrossRef]
  7. Aziida, N.; Keith, J.; Ugur, T.; Graham, W. Applications of natural language processing in aviation safety: A review and qualitative analysis. In Proceedings of the AIAA SCITECH 2025 Forum, Orlando, FL, USA, 6–10 January 2025. [Google Scholar]
  8. Li, Z.; Liu, P.; Wang, W.; Xu, C. Using support vector machine models for crash injury severity analysis. Accid. Anal. Prev. 2012, 45, 478–486. [Google Scholar] [CrossRef] [PubMed]
  9. Mokhtarimousavi, S.; Anderson, J.C.; Azizinamini, A.; Hadi, M. Improved support vector machine models for work zone crash injury severity prediction and analysis. Transp. Res. Rec. 2019, 2673, 680–692. [Google Scholar] [CrossRef]
  10. Chen, C.; Zhang, G.; Qian, Z.; Tarefder, R.A.; Tian, Z. Investigating driver injury severity patterns in rollover crashes using support vector machine models. Accid. Anal. Prev. 2016, 90, 128–139. [Google Scholar] [CrossRef]
  11. Nanyonga, A.; Keith, J.; Ugur, T.; Graham, W. Is BERTopic Better than PLSA for Extracting Key Topics in Aviation Safety Reports? arXiv 2025, arXiv:2506.06328. [Google Scholar]
  12. Nanyonga, A.; Wild, G. Utilizing AI for Aviation Post-Accident Analysis Classification. arXiv 2025, arXiv:2506.00169. [Google Scholar] [CrossRef]
  13. Socher, R.; Bengio, Y.; Manning, C.D. Deep learning for NLP (without magic). Tutor. Abstr. ACL 2012, 20122012, 5. [Google Scholar]
  14. Hannah, K.; Young-Seob, J. Sentiment classification using convolutional neural networks. Appl. Sci. 2019, 9, 2347. [Google Scholar] [CrossRef]
  15. Harley, A.W.; Ufkes, A.; Derpanis, K.G. Evaluation of deep convolutional nets for document image classification and retrieval. In Proceedings of the 2015 13th International Conference on Document Analysis and Recognition (ICDAR), Tunis, Tunisia, 23–26 August 2015; IEEE: New York, NY, USA, 2015. [Google Scholar]
  16. Coelho Eugenio, F.; Badin, T.L.; Fernandes, P.; Mallmann, C.L.; Schons, C.; Schuh, M.S.; Soares Pereira, R.; Fantinel, R.A.; Pereira da Silva, S.D. Remotely Piloted Aircraft Systems (RPAS) and machine learning: A review in the context of forest science. Int. J. Remote Sens. 2021, 42, 8207–8235. [Google Scholar] [CrossRef]
  17. Nanyonga, A.; Turhan, U.; Wild, G. A Systematic Review of Machine Learning Analytic Methods for Aviation Accident Research. Sci 2025, 7, 124. [Google Scholar] [CrossRef]
  18. Zhang, X.; Mahadevan, S. Ensemble machine learning models for aviation incident risk prediction. Decis. Support Syst. 2019, 116, 48–63. [Google Scholar] [CrossRef]
  19. Aziida, N.; Graham, W. Classification of Operational Records in Aviation Using Deep Learning Approaches. arXiv 2025, arXiv:2501.01222. [Google Scholar] [CrossRef]
  20. Jacob, D. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  21. Iddrisu, A.M.; Mensah, S.; Boafo, F.; Yeluripati, G.R.; Kudjo, P. A sentiment analysis framework to classify instances of sarcastic sentiments within the aviation sector. Int. J. Inf. Manag. Data Insights 2023, 3, 100180. [Google Scholar] [CrossRef]
  22. Chandra, C.; Ojima, Y.; Bendarkar, M.V.; Mavris, D.N. Aviation-BERT-NER: Named Entity Recognition for Aviation Safety Reports. Aerospace 2024, 11, 890. [Google Scholar] [CrossRef]
  23. Qasim, R.; Bangyal, W.H.; Alqarni, M.A.; Ali Almazroi, A. A fine-tuned BERT-based transfer learning approach for text classification. J. Healthc. Eng. 2022, 2022, 3498123. [Google Scholar] [CrossRef]
  24. Kierszbaum, S.; Lapasset, L.; Klein, T. Exploring use of transformer-based models on incident reports in aviation. In Proceedings of the CORIA 2021, Online Meeting, 15 April 2021. [Google Scholar] [CrossRef]
  25. Santiago, G.-C.; Eduardo, C.G.-M. Comparing BERT against traditional machine learning text classification. arXiv 2020, arXiv:2005.13012. [Google Scholar]
  26. Rai, N.; Kumar, D.; Kaushik, N.; Raj, C.; Ali, A. Fake News Classification using transformer based enhanced LSTM and BERT. Int. J. Cogn. Comput. Eng. 2022, 3, 98–105. [Google Scholar] [CrossRef]
  27. Kokab, S.T.; Asghar, S.; Naz, S. Transformer-based deep learning models for the sentiment analysis of social media data. Array 2022, 14, 100157. [Google Scholar] [CrossRef]
  28. Oliaee, A.H.; Das, S.; Liu, J.; Rahman, M.A. Using Bidirectional Encoder Representations from Transformers (BERT) to classify traffic crash severity types. Nat. Lang. Process. J. 2023, 3, 100007. [Google Scholar] [CrossRef]
  29. Singh, U.; Bhattacharya, M.; Padhi, R. State-of-the-Art Natural Language Processing for Aviation: A Review. TechRxiv 2025. [Google Scholar] [CrossRef]
  30. Soyalp, G.; Alar, A.; Ozkanli, K.; Yildiz, B. (Eds.) Improving text classification with transformer. In Proceedings of the 2021 6th International Conference on Computer Science and Engineering (UBMK), Ankara, Turkey, 15–17 September 2021; IEEE: New York, NY, USA, 2021. [Google Scholar]
  31. Gao, Y.; Zhu, G.; Duan, Y.; Mao, J. Semantic Encoding Algorithm for Classification and Retrieval of Aviation Safety Reports. In IEEE Transactions on Automation Science, Engineering; IEEE: New York, NY, USA, 2024. [Google Scholar]
  32. Liddy, E.D. Natural Language Processing; Syracuse University Syracuse University: Syracuse, NY, USA, 2001. [Google Scholar]
  33. Gupta, A.; Tatbul, N.; Marcus, R.; Zhou, S.; Lee, I.; Gottschlich, J. Class-weighted evaluation metrics for imbalanced data classification. arXiv 2020, arXiv:2010.05995v1. [Google Scholar]
  34. Kamyab, M.; Liu, G.; Adjeisah, M. Attention-based CNN and Bi-LSTM model based on TF-IDF and glove word embedding for sentiment analysis. Appl. Sci. 2021, 11, 11255. [Google Scholar] [CrossRef]
  35. Başarslan, M.S.; Kayaalp, F. MBi-GRUMCONV: A novel Multi Bi-GRU and Multi CNN-Based deep learning model for social media sentiment analysis. J. Cloud Comput. 2023, 12, 5. [Google Scholar] [CrossRef]
  36. Loshchilov, I.; Hutter, F. Fixing weight decay regularization in adam. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  37. O’shea, K.; Nash, R. An introduction to convolutional neural networks. arXiv 2015, arXiv:1511.08458. [Google Scholar] [CrossRef]
  38. Hochreiter, S.J. Long Short-term Memory. In Neural Computation; MIT Press: Cambridge, MA, USA, 1997. [Google Scholar]
  39. Nanyonga, A.; Wasswa, H.; Joiner, K.; Turhan, U.; Wild, G. Explainable Supervised Learning Models for Aviation Predictions in Australia. Aerospace 2025, 12, 223. [Google Scholar] [CrossRef]
  40. Reza, S.; Ferreira, M.C.; Machado, J.J.; Tavares, J.M.R. A multi-head attention-based transformer model for traffic flow forecasting with a comparative analysis to recurrent neural networks. Expert Syst. Appl. 2022, 202, 117275. [Google Scholar] [CrossRef]
  41. Vaswani, A.J.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Advances in Neural Information Processing Systems. In Attention Is All You Need; Curran Associates Inc.: Red Hook, NY, USA, 2017. [Google Scholar]
  42. Rengasamy, D.; Morvan, H.P.; Figueredo, G.P. (Eds.) Deep learning approaches to aircraft maintenance, repair and overhaul: A review. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; IEEE: Newe York, NY, USA, 2018. [Google Scholar]
  43. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  44. Habbat, N.; Nouri, H.; Anoun, H.; Hassouni, L. Sentiment analysis of imbalanced datasets using BERT and ensemble stacking for deep learning. Eng. Appl. Artif. Intell. 2023, 126, 106999. [Google Scholar] [CrossRef]
  45. Ribeiro, M.T.; Singh, S.; Guestrin, C. (Eds.) “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  46. Kannan, R.; Ng, H.; Yap, T.T.V.; Wong, L.K.; Chua, F.F.; Goh, V.T.; Lee, Y.L.; Wong, H.L. Handling class imbalance in education using data-level and deep learning methods. Int. J. Electr. Eng. Comput. 2025, 15, 741–754. [Google Scholar] [CrossRef]
  47. Helmreich, R.L.; Merritt, A.C. Culture at Work in Aviation and Medicine: National, Organizational and Professional Influences; Routledge: Oxfordshire, UK, 2017. [Google Scholar]
  48. Thai-Nghe, N.; Nghi, D.T.; Schmidt-Thieme, L. (Eds.) Learning optimal threshold on resampling data to deal with class imbalance. In Proceedings of the IEEE RIVF International Conference on Computing and Telecommunication Technologies, Hanoi, Vietnam, 1–4 November 2010. [Google Scholar]
  49. Jeni, L.A.; Cohn, J.F.; De La Torre, F. (Eds.) Facing imbalanced data-recommendations for the use of performance metrics. In 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction; IEEE: New York, NY, USA, 2013. [Google Scholar]
  50. Vamvakas, P.; Tsiropoulou, E.E.; Papavassiliou, S. Risk-aware resource management in public safety networks. Sensors 2019, 19, 3853. [Google Scholar] [CrossRef]
Figure 1. Methodological architecture.
Figure 1. Methodological architecture.
Ai 06 00251 g001
Figure 2. Shows the validation loss and Accuracy of each model.
Figure 2. Shows the validation loss and Accuracy of each model.
Ai 06 00251 g002
Figure 3. Confusion Matrix for BERT Model showing classification performance across Nil, Minor, Serious, and Fatal classes.
Figure 3. Confusion Matrix for BERT Model showing classification performance across Nil, Minor, Serious, and Fatal classes.
Ai 06 00251 g003
Figure 4. Confusion Matrix for the LSTM model showing classification performance across Nil, Minor, Serious, and Fatal classes.
Figure 4. Confusion Matrix for the LSTM model showing classification performance across Nil, Minor, Serious, and Fatal classes.
Ai 06 00251 g004
Figure 5. Confusion Matrix for CNN model showing classification performance across Nil, Minor, Serious, and Fatal classes.
Figure 5. Confusion Matrix for CNN model showing classification performance across Nil, Minor, Serious, and Fatal classes.
Ai 06 00251 g005
Table 1. Hyperparameter Tuning Process.
Table 1. Hyperparameter Tuning Process.
ModelHyperparameterTuned ValuesOptimal Value
LSTMNumber of Layers[1, 2, 3]2
Hidden Units[64, 128, 256]128
Dropout Rate[0.2, 0.3, 0.5]0.3
Learning Rate[0.001, 0.0005, 0.0001]0.0005
Batch Size[16, 32, 64]32
Optimizer[Adam, RMSprop, SGD]Adam
CNNNumber of Filters[32, 64, 128]64
Kernel Size[3, 5, 7]5
Pooling Type[MaxPooling, AveragePooling]MaxPooling
Dropout Rate[0.2, 0.3, 0.5]0.3
Learning Rate[0.001, 0.0005, 0.0001]0.001
Batch Size[16, 32, 64]32
Optimizer[Adam, RMSprop, SGD]Adam
BERTPretrained Model[BERT-Base, BERT-Large]BERT-Base
Learning Rate[3 × 10−5, 5 × 10−5, and 1 × 10−4]3 × 10−5
Batch Size[8, 16, 32]16
Epochs[3, 5, 10]5
Max Sequence Length[128, 256, 512]256
Optimizer[AdamW, SGD]AdamW
Table 2. Shows the evaluation metrics used.
Table 2. Shows the evaluation metrics used.
Metrics UsedFormulaEvaluation Focus
Precision (p) T P T P + F P Precision measures the correctly predicted positives from the total predicted patterns in a positive class.
Recall (r) T P T P + F N This recall measures the fraction of positive patterns that are correctly classified.
F1-score (F) 2 p r p + r F-score measures the weighted average score of precision and recall.
Accuracy (acc) T P + T N T P + F P + T N + F N Accuracy measures the total number of instances evaluated using the correctly predicted ratio.
Table 3. Shows the confusion metrics.
Table 3. Shows the confusion metrics.
Actual ValuePredicted Value
TNFP
FNTP
Table 4. Precision, Recall, F1-Score, and Accuracy for Each Model.
Table 4. Precision, Recall, F1-Score, and Accuracy for Each Model.
ModelMetricNilMinorFatalSeriousMacro AverageWeighted AverageAccuracy
LSTMPrecision0.99420.74520.91430.92680.89510.98980.9901
Recall0.99640.71780.71110.79170.80430.9901
F1-Score0.99530.73120.80000.85390.84510.9898
CNNPrecision0.99550.70060.94740.87760.88020.99020.9899
Recall0.99470.76070.80000.89580.86280.9899
F1-Score0.99510.72940.86750.88660.86960.9900
BERTPrecision1.001.001.001.001.001.001.00
Recall1.001.001.001.001.001.00
F1-Score1.001.001.001.001.001.00
Table 5. Ablation Study Results for LSTM, CNN, and BERT.
Table 5. Ablation Study Results for LSTM, CNN, and BERT.
ModelAblation ConfigurationAccuracyPrecisionRecallF1-Score
LSTMOriginal (128 units)0.99010.99100.99020.9906
LSTM replaced by Dense0.95000.94800.94500.9465
Units reduced 640.98500.98600.98000.9830
CNNOriginal (128 filters)0.98990.99050.98900.9897
Convolution replaced by MLP0.96000.95800.95700.9575
Filters reduced 640.98550.98600.98400.9850
BERTOriginal (12 layers, attention)1.00001.00001.00001.0000
Attention removed0.97000.96800.96500.9665
Transformer depth 60.97500.97400.97300.9735
WordPiece tokenizer replaced0.98000.98100.97800.9795
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nanyonga, A.; Joiner, K.F.; Turhan, U.; Wild, G. Deep Learning Approaches for Classifying Aviation Safety Incidents: Evidence from Australian Data. AI 2025, 6, 251. https://doi.org/10.3390/ai6100251

AMA Style

Nanyonga A, Joiner KF, Turhan U, Wild G. Deep Learning Approaches for Classifying Aviation Safety Incidents: Evidence from Australian Data. AI. 2025; 6(10):251. https://doi.org/10.3390/ai6100251

Chicago/Turabian Style

Nanyonga, Aziida, Keith Francis Joiner, Ugur Turhan, and Graham Wild. 2025. "Deep Learning Approaches for Classifying Aviation Safety Incidents: Evidence from Australian Data" AI 6, no. 10: 251. https://doi.org/10.3390/ai6100251

APA Style

Nanyonga, A., Joiner, K. F., Turhan, U., & Wild, G. (2025). Deep Learning Approaches for Classifying Aviation Safety Incidents: Evidence from Australian Data. AI, 6(10), 251. https://doi.org/10.3390/ai6100251

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop