Next Article in Journal
Design and Experimental Study of Turbine Blade Cooling Structure for an Engine
Previous Article in Journal
Aerodynamic Design and Analysis of an Aerial Vehicle Module for Split-Type Flying Cars in Urban Transportation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Responsible AI for Air Traffic Management: Application to Runway Configuration Assistance Tool

1
Aviation Systems Division, NASA Ames Research Center, Moffett Field, CA 94035, USA
2
NASA OSTEM Intern, Iowa State University, Ames, IA 50011, USA
3
Crown Innovations Inc., NASA Ames Research Center, Moffett Field, CA 94035, USA
4
Metis Technology Solutions Inc., NASA Ames Research Center, Moffett Field, CA 94035, USA
*
Author to whom correspondence should be addressed.
Aerospace 2025, 12(10), 872; https://doi.org/10.3390/aerospace12100872 (registering DOI)
Submission received: 28 August 2025 / Revised: 17 September 2025 / Accepted: 22 September 2025 / Published: 27 September 2025
(This article belongs to the Special Issue AI-Driven Innovations in Air Traffic Management and Aviation Safety)

Abstract

The complexity and magnitude of airspace operations are ever increasing, which creates new challenges for air traffic controllers. With the increase in the volume of operations, the size of available data is also increasing. Data-driven AI solutions can provide actionable information for complex decision-making processes that controllers face and assist them in improving the efficiency and safety of operations. However, for such solutions to be trusted by the users and stakeholders, they need to undergo a comprehensive validation process. In this paper, the literature in the development of responsible AI is studied and a subset of the framework is applied to an AI tool proposed for airport runway configuration management. The focus of this study is tackle two main challenges: (1) detection and mitigation of existing bias in the training data and the trained AI tool; and (2) quantification and improvement of the AI tool’s robustness to potential sources of noise in the data. We validate several responsible AI techniques using historical data and simulation studies on three major US airports and quantify their effectiveness in reducing the detected bias and also improving the robustness of the model to adversarial noise in the input data.

1. Introduction

The National Airspace System (NAS) of the United States is facing an ever-increasing trend in both the quantity and complexity of airspace operations. This means that air traffic controllers (ATCOs) are facing an unprecedented increase in workload. On the flip side, the above-mentioned trend has produced a rising amount of operational data that characterize an untapped potential for developing automated solutions to assist ATCOs on a daily basis. With the rise in the amount of data, one area of focus for the research community has been the development of Artificial Intelligence (AI)-/Machine Learning (ML)-based solutions to aid the ATCOs.
Data-driven AI solutions can provide actionable information for complex decision-making problems that ATCOs face and assist them in improving the efficiency and safety of airspace operations. For such solutions to be trusted by the potential users and other stakeholders, they need to undergo a comprehensive validation process before they can be deployed in real-world operations. As a result, one valuable area of research is developing a responsible AI framework, specifically tuned for decision-making processes in Air Traffic Management (ATM). In general, responsible AI refers to the practice of designing, developing, and deploying AI solutions in a way that is ethical, transparent, accountable, and fair. This includes several core principles and concepts such as explainability, interpretability, bias and fairness, robustness, satisfaction and safety, and transparency [1,2].
The main contribution of this paper is to adopt the general responsible AI framework and tune it to applications in the ATM domain. Hence, only a subset of the principles that are the most relevant for the development of AI tools in this domain are emphasized. As the first step, we focus on two core principles that are most relevant to ATM. Our contributions include deploying and evaluating the effectiveness of several techniques to address the following: (1) detection and mitigation of bias in input data and the model’s decision-making mechanism; and (2) robustness of the model’s output to the potential sources of noise in the input data. This paper specifically focuses on the Runway Configuration Assistance (RCA) tool [3,4], developed in collaboration with retired ATCOs and Subject Matter Experts (SMEs) to assist ATCOs in airport runway configuration decision-making. The RCA tool is chosen because it is one of the more mature tools that has been validated against historical data in multiple airports across the NAS and is the ideal candidate for quantifying responsible AI techniques.
Runway configuration management involves selecting the optimal runways and directions of operation for aircraft arrivals and departures. Every airport, depending on its geometry and number of runways, has a set of unique configurations that can be used by the ATCO. The main factors affecting the choice of runway configuration are surface winds, traffic load, meteorological conditions (e.g., cloud ceilings and visibility), and operational procedures (e.g., noise abatement). Once an airport is in an active runway configuration, choosing the optimal time-window to switch the configuration to a different one is a crucial yet challenging decision that ATCOs make multiple times every day. A sub-optimal selection of the runway configuration or poor timing of configuration changes can result in a significant increase in taxi times for aircraft on the surface of the airport, unnecessary aircraft holding (in the air or on the ground, causing long delays), and undesired go-arounds for arriving traffic in the air.
The Runway Configuration Assistance (RCA) tool [3] is an AI solution developed to address the runway configuration management problem. The RCA tool leverages an offline model-free reinforcement learning (RL) method, called Conservative Q-Learning (CQL) [5], to learn a near-optimal policy for runway configuration management solely based on historical data. It is important that learning takes place offline because there is no simulator available to mimic real-world operations for training AI solutions in an online fashion. It is important to emphasize that the RCA tool is not operational and is currently a NASA prototype developed in collaboration with the FAA.
The RCA tool has been validated against historical data from three major US airports, namely, Charlotte Douglas International Airport (CLT), Dallas Fort Worth International Airport (DFW), and Denver International Airport (DEN) [4]. The three airports represent a broad range of complexity associated with runway configuration management, with CLT being “low”, DFW being “mid”, and DEN being “high”. The RCA tool demonstrated significant performance improvement over several baseline methods such as supervised learning [6] and the offline implementation of Deep Q-Network (DQN) [7]. In this paper, a generic framework is developed for (1) detection and mitigation of existing bias in the training data, as well as the trained AI tools, and (2) quantification and improvement of the AI tool’s robustness to different sources of noise in the training data. Then, it is applied to the RCA tool to quantify its effectiveness.

2. Related Work

Significant research has been performed in both areas of bias mitigation and robustness analysis of AI/ML tools. Our goal is to only review a subset of the literature that is most relevant to the topic of this study, and we refer the reader to recent surveys for more comprehensive reviews [8,9,10]. Bias refers to the systemic and unfair influence of factors and variables that are used in a model’s development that can lead to incorrect or discriminatory outcomes and predictions. There are several sources that can cause a systemic bias in a trained AI/ML service, among which the most relevant ones are (1) data bias, arising from the quality, quantity, or representativeness of the training data; (2) feature bias, caused by incomplete data or bias in the developed feature engineering/selection algorithms; (3) model bias, caused by human decisions that affect the design and implementation of the AI/ML services; (4) algorithm bias, arising from a specific algorithm that is used in the training; and (5) operational bias, resulting from the deployment, usage, or interpretation of the models. In this paper, we mostly focus on the detection and mitigation of data, feature, and model biases.
Several statistical techniques can be used to detect bias in AI/ML services. Such bias might be due to the imbalance in the distribution of the training data, class membership, and/or feature attributes. The bias can also exist in models’ predictions, where fairness metrics such as demographic parity and equalized odds or statistical techniques such as the Chi-square test, t-test, and ANOVA can be used to detect it. For example, equalized odds assess whether the predictions are consistent across different features and group memberships of a specific feature, while the Chi-square test can determine whether the differences in the model predictions across different groups of a feature are statistically significant. In the case of runway configuration management, a feature can be wind direction, and a specific group membership of this feature can be north-ward wind. Sometimes an AI service is biased by a feature among all available in the input data, while, at other times, the bias exists only in a specific group membership of features. Details regarding different features within the input data and their group memberships for the case study in this paper are discussed later.
Once bias is detected, different techniques can be used to mitigate it and make the model more robust to other potential sources of bias. Generally, bias mitigation techniques are divided into three categories, depending on when they are applied: (1) pre-processing, which mitigates bias in the training data before it reaches the model; (2) in-processing, which mitigates bias while training a model; and (3) post-processing, which mitigates bias in a trained model during the after-the-fact analysis. Among these three categories, pre-processing and in-processing techniques are significantly more popular and utilized [9]. As a result, we build on pre-existing work and focus mainly on these categories. Among these two, pre-processing techniques are model-agnostic and thus can easily be applied to a wide variety of models without any modification of the underlying algorithms, making them versatile techniques for bias mitigation.
Relabeling is a popular pre-processing technique, which analyzes parts of the data where there is a suspicion that the ground-truth labels might be erroneous, or show a large variance, and corrects their labels. This is specifically applicable to the ATM domain and the RCA tool, because historical data shows great variations among the decisions made by different ATCOs, and this technique can mitigate such variance in the training labels before the models are trained on the data. For example, we observe that, in a specific circumstance (including hour of day, meteorological conditions, and traffic load), different runway configurations are used by different ATCOs. This could be due to several reasons, including the subjectivity of the decision-making process or the effect of other operational variables that are not present in the data used for training of the AI tool. As a result, such variance in the labels of the data would cause a distraction in the training process. An extreme example of such variation is observed in DEN, where ATCO decisions on runway configuration show around 50 % variation [11]. Relabeling identifies such data samples, reviews their circumstance, and corrects their labels. Several techniques have been proposed in the past to apply relabeling, among which the most popular ones are messaging (a ranking technique to determine best candidates for relabeling) [12,13] and k-nearest neighbors [14].
Another popular pre-processing technique is sampling, which is widely utilized for bias mitigation [15,16,17]. This technique changes the distribution of the samples in the training data to increase/decrease the impact of certain classes/features/groups of data and re-weight their importance during the training process. Such a re-distribution of the importance among the training data would result in the model being less biased by the majority representation in the training data. Perturbation [18,19] is another technique that can both reduce the bias of the model and improve its robustness to sources of noise. Perturbation techniques induce small noise into the training data to quantify its effect on the model’s output. Such perturbations can be further included in the training data to improve the robustness of the model to those sources of noise. These perturbations can be random and/or adversarial. Adversarial noises usually use the parameters of the trained AI model to introduce the smallest amount of noise in the input data, which results in the greatest change in the model’s output. As a result, they are considered among the most aggressive sources of perturbation that can hurt the trained AI model.
In-processing techniques mitigate bias during the training process of the AI/ML tool. One of the most popular methods in this category is regularization, which is a technique proposed to penalize discrimination or significant reliance on a specific feature or group membership within the input data [20]. Regularization primarily serves to prevent a model from overfitting on training data and generalize better to unseen data in the operational setting [21]; however, it has been shown to reduce the amount of bias in models’ predictions as well. Among other popular in-processing techniques are representation learning and adversarial learning. Representation learning uses data transformation (either linear or non-linear) to remove the bias from the input data, while maintaining as much information as possible [22,23,24,25,26]. On the other hand, adversarial learning uses concepts from game theory and puts the AI tool in competition with an adversary (a separate AI model) that specifically tries to induce bias in the original AI tool and, through simultaneously training the two models, reduces the overall bias [27,28,29,30,31].

3. Methodology

In this section, the RCA tool is briefly discussed first, then, we introduce a general statistical technique to detect bias in existing models and implement several bias mitigation techniques to reduce the detected bias in the RCA tool. Lastly, a comprehensive analysis is performed to quantify the robustness of the model to different sources of noise and the effect of adversarial training on improving the model’s robustness.

3.1. RCA Tool

The Runway Configuration Assistance (RCA) tool [11] is a previously developed AI solution based on the family of offline model-free RL [32] to provide decision-support for ATCOs in runway configuration management. It frames the problem as a Markov Decision Process (MDP) [33], where the state space S is comprised of relevant information for the ATCO decision-making such as wind conditions, meteorological conditions (cloud ceiling and visibility), traffic load (scheduled arrival and departure aircraft), and time of day. Given this information, a policy for runway configuration selection, π : S A , can be computed based on maximizing long-term expected utilities (or rewards):
V π ( s t ) = max π f U ( s t , π ( s t ) ) + + γ s t + 1 S p ( s t + 1 s t , π ( s t ) ) V π ( s t + 1 )
where γ [ 0 , 1 ) is a discount factor, which tunes the importance of future utilities in comparison to the current ones and is treated as a hyperparameter. p ( s t + 1 s t , π ( s t ) ) is the probability of starting in state s t , taking action π ( s t ) , and ending up in the state s t + 1 according to the transition (dynamics) function f T : S × A S , which defines the dynamics of the states. f U : S × A R is the utility function, which is specifically designed with feedback from the SMEs. It includes traffic throughput, average transit times on the surface of the airport, and the number of go-arounds as a result of high winds. For further details regarding the utility function, refer to [11]. If the AI agent has full knowledge of all components of the MDP, then dynamic programming [33] can be used to find the optimal policy from Equation (1). However, in many real-world applications such as ATM, the transition (dynamics) function is complex, and, as a result, it is not feasible to learn it from limited historical data. Furthermore, any error introduced in learning the model (or simplifying the dynamics) would affect the quality of the policy obtained and its applicability in practice. As a result, the RCA tool relies on a family of offline model-free control, specifically a state-of-the-art Conservative Q-Learning (CQL) algorithm [5], to learn a near-optimal policy solely based on historical data.

3.2. Bias Detection

We develop a general statistical method, similar to the Chi-square test, to quantify the bias. The developed method is generic and can be applied to any supervised learning or RL tasks; hence, it is applicable to the RCA tool. Let us consider the following: (1) the input data X R N × F , where N is the number of instances of data, and F is the number of features (e.g., wind conditions, arrival traffic, etc.); (2) a trained model M θ : X Y ^ , where θ are the parameters of the model, and Y ^ are its outputs (predictions); and (3) the ground-truth labels (e.g., optimal runway configurations) for each input datapoint, i.e., Y. It is assumed that any of the features in the input data can be categorized into groups based on the range of values that they take on. For example, the wind direction is discretized into 8 bins each categorizing 45-degrees. Figure 1 illustrates the group memberships for the wind direction as an example. Further details regarding the features in the training data and their group memberships are provided later in Section 4.1. Lastly, one needs to select a performance metric to quantify the performance of the model; here, F 1 -score (harmonic mean of precision and recall) is used as the performance metric of choice.
Algorithm 1 shows the pseudo-code for the statistical technique developed to quantify the bias. In line 3, the baseline performance of the trained model, M θ , is calculated on each specific group membership of each feature f F , i.e., X g f , where g G f , a set containing all groups of feature f. For example, f is wind direction and g is a specific group within this feature, which can be the east-ward winds (bin 3 in Figure 1). Lines 4–7 calculate the expected performance for each group membership of a feature if the group assignment was completely random. For example, if, for 10% of the data, the wind direction group membership was East, we randomly assign another 10% of the input data to take East group membership for wind direction and calculate the performance of the model. This step is repeated a large number of times to calculate the expected value, μ g , and the standard deviation, σ g , of the expected performance in line 8. Finally, in lines 9–14, if the actual performance of the model, F 1 g * , is between the confidence bound of μ g 2 σ g and μ g + 2 σ g , the conclusion is that there is no statistically significant bias for the group g of feature f. On the other hand, if the performance is outside of this bound, an existing negative or positive bias is quantified, depending on what side of the bound the performance lies. If the performance is below the lower limit, then the model is performing significantly worse for the specific group and is considered as negative bias, while, if the performance is above the upper limit, the performance is better than expected, hence leading to the name positive bias.
Algorithm 1 Bias detection method
Input: data X R N × F , trained model M θ , and labels Y
Output: no/negative/positive bias
  1:
for each feature f F  do
  2:
     for each group membership g G f  do
  3:
           calculate F 1 g * = eval ( M θ ( X g f ) , Y g )
  4:
           for  i = 1 , , I  do
  5:
                 randomly permute group memberships in X f
  6:
                 calculate F 1 g i based on new memberships
  7:
           end for
  8:
           calculate μ g and σ g as mean and standard deviation of F 1 g i i = 1 , , I
  9:
           if  μ g 2 σ g F 1 g * μ g + 2 σ g  then
10:
                there is no bias for group g
11:
           else if  F 1 g * < μ g 2 σ g  then
12:
                there is a negative bias for group g
13:
           else if  F 1 g * > μ g + 2 σ g  then
14:
                there is a positive bias for group g
15:
           end if
16:
     end for
17:
end for

3.3. Bias Mitigation

Among the numerous techniques that exist for bias mitigation, we summarized the most relevant ones in Section 2. In this section, the implementation details of those techniques are explained and they are applied to the RCA tool to mitigate its existing bias and improve its robustness.
The first implemented technique is relabeling, as this is an efficient technique to improve the quality of the training data. Relabeling adjusts the ground-truth labels for data entries that might be erroneous or outliers. This could be due to error in data entries and/or variations in subjective decisions that are present in the historical data. The k-nearest neighbors algorithm is adopted for this purpose, where, for each data instance in the training data, we find the k closest neighbors ( k = 5 is assumed in this paper) to it and adjust its ground-truth labels if the following two conditions are met: (1) at least 80% of the nearest neighbors agree on a different label (i.e., runway configuration); and (2) the average distance of the nearest neighbors to the data instance is less than a threshold (this threshold is defined based on the nature of training data and domain knowledge). Once the training data are relabeled, a model is trained using the modified data and is evaluated to quantify the mitigated bias.
Next, we implement two independent sampling techniques to evaluate their effectiveness in mitigating bias. The first one is an importance sampling scheme designed to balance the class distribution of the training data (referred to as Class Blc in figures). Table 1 provides the class distribution among the data for the three airports, i.e., CLT, DEN, and DFW. The class balancing method assigns weights to data instances in the data loader during the training, with the weights being inversely proportional to the class relative frequency. Intuitively, this gives data from under-represented classes a higher probability of being sampled during training, which improves their representativeness. The second balancing technique focuses on increasing the representativeness of the specific group memberships of features in the data that show negative bias in the baseline model’s performance (referred to as Feature Blc in figures). We refer to the trained model on the original data without any bias mitigation as the baseline model.
Lastly, we implement regularization, which is one of the popular techniques used to mitigate bias and also address overfitting. The RCA tool is a multi-layer feed-forward neural network, and, as a result, dropout [21] is utilized to implement regularization in the training of the model. We use a 50% rate of dropout in each layer of the Q-network.

3.4. Robustness and Adversarial Training

We mainly adopt adversarial training to provide a systemic approach for quantifying and improving the robustness of the RCA tool. Different sources of noise or error can affect the performance of the model and this could be due to inaccuracies in the source data, limitations of sensor precision, etc. The focus of the adversarial training is to improve model’s robustness to different sources of noise and increase its truth-worthiness. Three different types of noise are implemented for the purpose of experiments in this paper. The first and most general source of noise is random noise; given input data X, perturbed data X ˜ is created as follows:
X ˜ Unif ( X + e , x e )
Here, Unif is the uniform distribution and e is the magnitude of the perturbation.
The other two techniques are adversarial noises that use the parameters of the trained model to create a bounded amount of perturbation in the input data that would result in the most deviation in its output. As a result, they cause a stronger perturbation to the input data (compared to the random noise) and direct attacks to hurt the model’s performance. If random noise can be thought to represent data inaccuracies in the average case, then adversarial noise can be regarded as worst-case scenarios. The first adversarial noise is Fast Gradient Sign Method (FGSM) [34], which is a one-step gradient-based algorithm. FGSM generates an adversarial example by adding a small perturbation to the input data that maximizes the loss function and is defined as follows:
X ˜ = X + e · sign x J θ , x , y
where sign ( · ) is the sign function that computes the sign of each component in the input vector and x J θ , x , y is the gradient of the loss function with respect to the input data.
The second methods is Projected Gradient Descent (PGD) [35], which is an iterative algorithm that extends the FGSM approach and is defined as follows:
X ˜ t + 1 = proj B ( x , e ) X ˜ t + α · sign x ˜ t J θ , x ˜ t , y
Here, X ˜ 0 = X , t is the iteration counter, proj B ( x , e ) is the function that projects the perturbed data onto a ball of radius e centered at X, and α is the learning rate. The projection operation ensures that the perturbation remains within the desired threshold of error. PGD can be seen as a stronger attack than FGSM due to its iterative nature of refining the perturbation to find a more effective attack.
First, the robustness of the RCA tool to these sources of noise in the input data is quantified. Then, the training data is augmented with perturbed data and a new model is trained from scratch on the augmented data to evaluate the effectiveness of the adversarial training in improving the robustness of the model.

4. Results and Discussion

In this section, we summarize the findings of applying bias mitigation and adversarial training techniques to the RCA tool and quantify their effectiveness.

4.1. Data

FAA’s Aviation System Performance Metrics (ASPM) reports (https://aspm.faa.gov/), NASA’s Sherlock Data Warehouse (https://sherlock.opendata.arc.nasa.gov/sherlock_open/), Meteorological Aerodrome Reports (METARs) (https://aviationweather.gov/data/metar/), and Terminal Aerodrome Forecast (TAF) reports (https://aviationweather.gov/data/taf/) are fused to create a dataset for training the RCA tool. We use two years of data (2018–2019) from the three chosen US airports (CLT, DEN, and DFW) for the purpose of experimental validation detailed in this paper.
The data is aggregated by a quarter of an hour, meaning that each year results in 35,040 instances of data; overall, we have 70,080 datapoints for each airport. The features included in the analysis are hour of day, wind conditions (direction and speed), cloud ceiling, visibility, arrival demand, and departure demand. Wind direction is categorized into eight bins as shown in Figure 1, wind speed is categorized into six bins (0–5, 5–10, 10–15, 15–20, 20–25, >25 knots), cloud ceiling is categorized into five bins (0–2500, 2500–7500, 7500–15,000, 15,000–30,000 feet, and no ceiling), visibility is categorized into five bins (0–3, 3–6, 6–7.5, 7.5–9, ≥10 miles), and arrival/departure demands are categorized into four bins (0–10, 10–20, 20–30, >30 aircraft). To properly evaluate the effectiveness of the implemented techniques, the entire dataset is divided into 80% training and 20% testing sets, randomly, and each of the techniques are applied only to the training set, keeping the testing set representative of real-world unseen data.

4.2. Bias Mitigation

The main goal of bias mitigation in the case of RCA tool is to mitigate and decrease the negative bias, i.e., cases where the performance is significantly lower than expected. This is due to the fact that higher than expected performance in the ATM domain is not considered an adverse bias. We measure the amount of negative bias as the difference between the actual performance of the baseline model, F 1 g * , and the lower bound of expected performance (refer to Algorithm 1), μ g 2 σ g , i.e., negative bias = min μ g 2 σ g F 1 g * , 0 . Figure 2 shows the average existing negative bias for each feature (the average is performed across different group memberships of a feature), for the three airports of CLT, DEN, and DFW. In this figure, Baseline refers to the original model with no bias mitigation technique applied, and values closer to zero represent less detected negative bias.
Overall, we see some similarities and differences between the three airports. CLT, being the airport with the least complex runway configuration among the three, shows only a significant negative bias in wind conditions. This is intuitive, since wind is the major factor affecting the choice of runway configurations. On the other hand, DEN and DFW, being the more complex airports, show significant bias in most of the features. For example, the hour of day and arrival demand show significant bias for DEN, whereas, for CLT and DFW, they represent negligible bias. On the other hand, the meteorological conditions such as wind, cloud ceiling, and visibility show significant negative bias for DFW. This difference between DEN and DFW could be due to the significant variation in the weather and operational conditions at DEN on an hourly basis, which are related to the location of the airport. This difference is also evident in the usage distribution reported in Table 1, where we can see that six configurations at DEN are used more than 10% of the time, while there are only two configurations at DFW that are utilized more than 10%.
Average changes in the existing bias imposed after applying each of the techniques are reported in Table 2. The most effective technique for reducing bias is relabeling, followed by regularization. Two main messages can be inferred from this finding: (1) The effectiveness of the relabeling technique emphasizes the subjectivity and variability of the ATCOs decision-making, which is present in the historical operational data. This further emphasizes that the ground-truth labels for the data in the case of runway configuration management are highly variable, error-prone, and need careful review. Lastly, the variability among the ATCO decisions might be emphasizing the importance of other features that are affecting the decision-making process, but are not included in the input data of the RCA tool due to unavailability, e.g., noise abatement procedures, operational procedures, etc. (2) Alleviation of bias by regularization shows that the RCA tool might be overly relying on specific features or specific group memberships within a feature that are causing bias in the predictions; increasing the size of data and/or features used in the analysis, along with the use of regularization techniques, should alleviate this source of bias.
Although there still remain significant amounts of negative bias, even after applying a bias mitigation technique, we noticed that the majority of such bias exists in calm wind conditions (wind speeds less than 10 knots). Figure 3 shows the visualization of existing bias (green for no bias, red for negative bias, and blue for positive bias), before (left panel) and after (right panel) applying the relabeling technique to the RCA tool in the cases of CLT (top panels), DEN (middle panels), and DFW (bottom panels). In this figure, wind directions are shown as a radian plot (degrees from north), wind speed is shown as the distance from the center of the circle, and sizes of dots represent the amount of data that is in the training set for each category.
As is illustrated in the figure, most of the existing negative biases under high wind conditions (more than 10 knots) have resolved after applying the relabeling technique for all airports, and the majority of the existing negative bias lies in the realm of calm wind conditions. This is a major finding since, for runway configuration management, key decisions are the ones that happen in the high wind conditions, and the choice of runway configuration in the calm wind conditions is not as important. A lot of other factors can affect the decision-making in calm wind conditions that are not present in the input data, such as operational procedures, noise abatement procedures, airline preferences, and the proximity of the runways to the terminals. Hence, the aggregate mitigation of bias reported in this paper represents the mitigation in general settings, while, in the cases of high wind conditions, the mitigation effects are more significant. Another observation in the case of CLT is that the majority of the remaining negative bias is in the cases of eastward and westward winds, which, due to the north–south orientation of the runways at this airport, the model has no mechanism for addressing the existing bias.

4.3. Adversarial Training

In this section, we quantify the robustness of the RCA tool to random noise and two types of adversarial noise, FGSM and PGD. Then, the perturbed data are included in the training process of the RCA tool and the improvements in the robustness of the model to these sources of noise is quantified. Figure 4 (top panel) shows the drop in performance of the RCA tool (in terms of increase in the cross-entropy loss) as a function of the magnitude of each source of noise, e, as defined in Equations (2)–(4). As expected, PGD has significantly more impact on the model performance compared to the other two approaches, followed by FGSM. In Table 3, three stages of the drop in the performance of the model are identified, i.e., 5%, 10%, and 25%, finding the smallest magnitude level of each noise that causes such a drop in the performance (these values are averaged among the three airports).
Next, we add the perturbed data to the original training set, train a new model on the augmented set, and evaluate its effectiveness on the robustness of the model to the sources of noise. Figure 4 (bottom panel) quantifies this effect for the case of augmenting data with PGD perturbation compared to the original model that does not contain perturbed data in the training. For example, in the case of 10% drop noted in the figure, the training set is perturbed with a magnitude e = 0.04 of PGD noise (according to Table 3), and a new model is trained on the augmented training set that contains the original data plus the perturbed ones. Once the model is trained, its performance is evaluated on the test set.
Overall, we observe that perturbing the training data with noise of some sort (random and/or adversarial) and including the perturbed data in the training of the RCA tool improves its robustness to such noises significantly. This effect is more evident in the case of perturbing the training data with a higher magnitude of PGD noise. This finding might allude to the inherent sources of noise that exist in the operational historical data and show that including noisy data in the training of the AI tools has positive impacts on their robustness to noise in operations. As a result, researchers and practitioners in the field of ATM should not aim to remove all sources of noise and uncertainty from their training data.

5. Conclusions

In this paper, the literature in responsible AI, its different components and principles, and its applicability to the domain of air traffic management are studied. The developed framework mainly addresses two fundamental challenges: (1) bias detection and mitigation in both training data and developed AI models; and (2) model’s robustness to different sources of noise. In order to showcase the effectiveness of the developed framework, we applied it to a previously developed AI decision-support tool for runway configuration management. Using real-world historical data (years 2018–2019) from three major US airports (CLT, DFW, and DEN), the following are shown: (1) simple and scalable bias mitigation techniques such as relabeling and regularization are effective in mitigating existing bias in the AI tool, while not compromising performance. The performance improvement of the relabeling technique further emphasized the importance of data reviews that need to be carried out either through automated approaches or by subject matter experts for the ground-truth data in the domain of air traffic management; (2) adversarial training has a significant effect in improving the robustness of the AI solutions to various sources of adversarial noise. Although the effectiveness of this technique varies in different airports, they showed an overall improvement to the robustness of the model and, hence, are recommended to be adopted by the AI developers in this domain. This paper is the first step towards developing a unified framework for the implementation of responsible AI principles for AI solutions developed to aid air traffic management.

Author Contributions

M.M. and K.M.K. designed the concept and initiated the project. F.M.S. and M.M. developed the bias detection/mitigation methodologies, and P.R. implemented it on the tool and obtained results. M.M. and Z.W. developed the adversarial training methodologies, and Z.W. implemented the robustness analysis and obtained the results. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets presented in this article are not readily available because of restrictions that exist on the data. Requests to access the datasets should be directed to the corresponding author.

Acknowledgments

The authors acknowledge the invaluable support and feedback received from subject matter experts affiliated with the Federal Aviation Administration (FAA)’s Office of NextGen (ANG).

Conflicts of Interest

F.M.S was employed by Crown Innovations Inc, and P.R. was employed by the company Metis Technology Solutions Inc. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Roadmap for Artificial Intelligence Safety Assurance. Federal Aviation Administration. Accessed in Summer of 2024. Available online: https://www.faa.gov/media/82891 (accessed on 1 September 2025).
  2. Stahl, B. Embedding responsibility in intelligent systems: From AI ethics to responsible AI ecosystems. Sci. Rep. 2023, 13, 5. [Google Scholar] [CrossRef] [PubMed]
  3. Memarzadeh, M.; Puranik, T.G.; Kalyanam, K.M.; Ryan, W. Airport runway configuration management with offline model-free reinforcement learning. In Proceedings of the AIAA Scitech 2023 Forum, National Harbor, MD, USA, 23–27 January 2023. [Google Scholar]
  4. Kalyanam, K.M.; Memarzadeh, M.; Crissman, J.; Yang, R.; Tejasen, K.T.J. Applying machine learning tools for runway configuration decision support. In Proceedings of the 11th International Conference on Research in Air Transportation (ICRAT), Singapore, 1–4 July 2024. [Google Scholar]
  5. Kumar, A.; Zhou, A.; Tucker, G.; Levine, S. Conservative Q-learning for offline reinforcement learning. arXiv 2020, arXiv:2006.04779. [Google Scholar] [CrossRef]
  6. Puranik, T.G.; Memarzadeh, M.; Kalyanam, K.M. Predicting airport runway configurations for decision-support using supervised learning. In Proceedings of the 2023 IEEE/AIAA 42nd Digital Avionics Systems Conference (DASC), Barcelona, Spain, 1–5 October 2023. [Google Scholar]
  7. Mnih, V.; Kavukcuoglu, K.; Silver, D. Human-level control through deep reinforcement learning. Nature 2015, 1, 529–533. [Google Scholar] [CrossRef] [PubMed]
  8. Hort, M.; Chen, Z.; Zhang, J.M.; Harman, M.; Sarro, F. Bias mitigation for machine learning classifiers: A comprehensive survey. ACM J. Responsible Comput. 2024, 1, 1–52. [Google Scholar] [CrossRef]
  9. Siddique, S.; Haque, M.A.; Gupta, K.D.; George, R.; Gupta, D.; Faruk, M.J.H. Survey on machine learning biases and mitigation techniques. Digital 2024, 1, 1–68. [Google Scholar] [CrossRef]
  10. Pagano, T.P.; Loureiro, R.B.; Lisboa, F.V.N.; Peixoto, R.M.; Guimaraes, G.A.S.; Cruz, G.O.R.; Araujo, M.M.; Santos, L.L.; Cruz, M.A.S.; Oliveira, E.L.S.; et al. Bias and unfairness in machine learning models: A systematic review on datasets, tools, fairness metrics, and identification and mitigation methods. Big Data Cogn. Comput. 2023, 7, 15. [Google Scholar] [CrossRef]
  11. Memarzadeh, M.; Kalyanam, K.M. Runway configuration assistance: An offline reinforcement learning method for air traffic management. J. Aerosp. Inf. Syst. 2025, 22, 275–287. [Google Scholar] [CrossRef]
  12. Kamiran, F.; Calders, T. Classifying without discriminating. In Proceedings of the IEEE 2nd International Conference on Computer, Control and Communication, Karachi, Pakistan, 17–18 February 2009. [Google Scholar]
  13. Kamiran, F.; Calders, T. Data preprocessing techniques for classification without discrimination. Knowl. Inf. Syst. 2012, 1, 1–33. [Google Scholar] [CrossRef]
  14. Luong, B.T.; Ruggieri, S.; Turini, F. k-NN as an implementation of situation testing for discrimination discovery and prevention. In Proceedings of the 17th ACM SIGKDD, San Diego, CA, USA, 21–24 August 2011; pp. 502–510. [Google Scholar]
  15. Celis, L.E.; Keswani, V.; Vishnoi, N. Data preprocessing to mitigate bias: A maximum entropy based approach. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 1349–1359. [Google Scholar]
  16. Chai, J.; Wang, X. Fairness with adaptive weights. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; pp. 2853–2866. [Google Scholar]
  17. Du, W.; Wu, X. Fair and robust classification under sample selection bias. In Proceedings of the 30th ACM International Conference on Information and Knowledge Management, Gold Coast, Australia, 1–5 November 2021; pp. 2999–3003. [Google Scholar]
  18. Feldman, M.; Friedler, S.A.; Moeller, J.; Scheidegger, C.; Venkatasubramanian, S. Certifying and removing disparate impact. In Proceedings of the 21th ACM SIGKDD, Sydney, Australia, 10–13 August 2015; pp. 259–268. [Google Scholar]
  19. Johndrow, J.E.; Lum, K. An algorithm for removing sensitive information: Application to race-independent recidivism prediction. Ann. Appl. Stat. 2019, 1, 189–220. [Google Scholar] [CrossRef]
  20. Kamishima, T.; Akaho, S.; Asoh, H.; Sakuma, J. Fairness-aware classifier with prejudice remover regularizer. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Bristol, UK, 24–28 September 2012; pp. 35–50. [Google Scholar]
  21. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  22. Zemel, R.; Wu, Y.; Swersky, K.; Pitassi, T.; Dwork, C. Learning fair representations. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 17–19 June 2013; pp. 325–333. [Google Scholar]
  23. Creager, E.; Madras, D.; Jacobsen, J.; Weis, M.; Swersky, K.; Pitassi, T.; Zemel, R. Fair and robust classification under sample selection bias. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 1436–1445. [Google Scholar]
  24. Roh, Y.; Lee, K.; Whang, S.; Suh, C. FR-Train: A mutual information-based approach to fair and robust training. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 8147–8157. [Google Scholar]
  25. Kamani, M.M.; Haddadpour, F.; Forsati, R.; Mahdavi, M. Efficient fair principal component analysis. Mach. Learn. 2022, 111, 3671–3702. [Google Scholar] [CrossRef]
  26. Gupta, U.; Ferber, A.; Dilkina, B.; Steeg, G.V. Controllable guarantees for fair outcomes via contrastive information estimation. AAAI Conf. Artif. Intell. 2021, 1, 7610–7619. [Google Scholar] [CrossRef]
  27. Dalvi, N.; Domingos, P.; Sanghai, S.; Verma, D. Adversarial classification. In Proceedings of the tenth ACM SIGKDD, Seattle, WA, USA, 22–25 August 2004; pp. 99–108. [Google Scholar]
  28. Lowd, D.; Meek, C. Adversarial learning. In Proceedings of the eleventh ACM SIGKDD, Chicago, IL, USA, 21–24 August 2005; pp. 641–647. [Google Scholar]
  29. Iosiidis, V.; Fetahu, B.; Ntoutsi, E. Fae: A fairness-aware ensemble framework. In Proceedings of the IEEE International Conference on Big Data, Los Angeles, CA, USA, 9–12 December 2019; pp. 1375–1380. [Google Scholar]
  30. Oneto, L.; Doninini, M.; Elders, A.; Pontil, M. Taking advantage of multitask learning for fair classification. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, Honolulu, HI, USA, 27–28 January 2019; pp. 227–237. [Google Scholar]
  31. Calders, T.; Verwer, S. Three naive Bayes approaches for discrimination-free classification. Data Min. Knowl. Discov. 2010, 1, 277–292. [Google Scholar] [CrossRef]
  32. Levine, S.; Kumar, A.; Tucker, G.; Fu, J. Offline reinforcement learning: Tutorial, review, and perspectives on open problems. arXiv 2020, arXiv:2005.01643. [Google Scholar] [CrossRef]
  33. Sutton, R.S.; Barto, A.G. Introduction to Reinforcement Learning; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  34. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  35. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar]
Figure 1. Discretization of wind direction into 8 bins.
Figure 1. Discretization of wind direction into 8 bins.
Aerospace 12 00872 g001
Figure 2. Mitigation of negative bias based on different techniques.
Figure 2. Mitigation of negative bias based on different techniques.
Aerospace 12 00872 g002
Figure 3. Illustrating existing bias in wind conditions.
Figure 3. Illustrating existing bias in wind conditions.
Aerospace 12 00872 g003
Figure 4. This figure shows (top panel) the effect of different sources of noise on the performance of the RCA tool and (bottom panel) the improvements made in the robustness of the model by adding perturbed data (according to PGD method) to the training process.
Figure 4. This figure shows (top panel) the effect of different sources of noise on the performance of the RCA tool and (bottom panel) the improvements made in the robustness of the model by adding perturbed data (according to PGD method) to the training process.
Aerospace 12 00872 g004
Table 1. Major runway configurations and their frequency based on 2018–2019 data for CLT, DEN, and DFW.
Table 1. Major runway configurations and their frequency based on 2018–2019 data for CLT, DEN, and DFW.
Configuration [Arr/Dep]Usage [%]
CLT
N/N60.8
S/S39.2
DEN
SE/SE18.8
S/S15
N/NEW14.5
S/SEW12.6
N/N12.3
NE/NE11.7
NW/NW8.6
SW/SW3.4
E/E1.6
NS/EW1.2
W/W0.3
DFW
SSE/S61.5
NNW/NNW21.3
S/S7.6
N/NNW5.1
NNW/N3
N/N1.1
SSE/NNW0.2
NNW/S0.1
NW/NW0.1
Table 2. Percentage changes in the existing bias by each technique. The percentages are averaged for the three airports. Bold text shows the best bias mitigation technique for each feature.
Table 2. Percentage changes in the existing bias by each technique. The percentages are averaged for the three airports. Bold text shows the best bias mitigation technique for each feature.
FeatureClass BlcFeat. BlcReg.Relab.
Hour+18%−23%−11%−14%
Wind−17%−17%−36%−43%
Cloud−11%−23%−26%−30%
Visibility+20%−22%−55%−53%
Arrival+20%+25%−29%−39%
Departure−13%−14%−6%−12%
Average+3%−12%−27%−32%
Table 3. Quantified magnitude of each source of noise for specific drop in the model’s performance.
Table 3. Quantified magnitude of each source of noise for specific drop in the model’s performance.
Method5% Drop10% Drop25% Drop
Random0.140.20.4
FGSM0.020.040.09
PGD0.020.040.08
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Memarzadeh, M.; Wang, Z.; Masrour Shalmani, F.; Razzaghi, P.; Kalyanam, K.M. Responsible AI for Air Traffic Management: Application to Runway Configuration Assistance Tool. Aerospace 2025, 12, 872. https://doi.org/10.3390/aerospace12100872

AMA Style

Memarzadeh M, Wang Z, Masrour Shalmani F, Razzaghi P, Kalyanam KM. Responsible AI for Air Traffic Management: Application to Runway Configuration Assistance Tool. Aerospace. 2025; 12(10):872. https://doi.org/10.3390/aerospace12100872

Chicago/Turabian Style

Memarzadeh, Milad, Zili Wang, Farzan Masrour Shalmani, Pouria Razzaghi, and Krishna M. Kalyanam. 2025. "Responsible AI for Air Traffic Management: Application to Runway Configuration Assistance Tool" Aerospace 12, no. 10: 872. https://doi.org/10.3390/aerospace12100872

APA Style

Memarzadeh, M., Wang, Z., Masrour Shalmani, F., Razzaghi, P., & Kalyanam, K. M. (2025). Responsible AI for Air Traffic Management: Application to Runway Configuration Assistance Tool. Aerospace, 12(10), 872. https://doi.org/10.3390/aerospace12100872

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop