Explainability and Interpretability in Concept and Data Drift: A Systematic Literature Review
Abstract
1. Introduction
- GQ1.
- Trend: How has research on explainability and interpretability for concept and data drift evolved over time?
- GQ2.
- Domain: Which application areas most frequently adopt these techniques?
- GQ3.
- Publication Venues: Which journals, conferences, or workshops exert the greatest influence in this field?
- GQ4.
- Research Groups: Which institutions or teams are the most prolific contributors?
- SQ1.
- Online vs. Offline Learning: How do online learning approaches differ from offline (batch) learning when coupled with explainability?
- SQ2.
- Drift Characteristics: How do concept versus data drift and sudden, gradual, incremental, or recurring drift affect the design of explainability methods?
- SQ3.
- Datasets and Benchmarks: Which datasets are most often used for evaluation?
- SQ4.
- Models: What model families (e.g., tree-based, neural, GNNs) are paired with explainability in drift scenarios?
- SQ5.
- Metrics: Which quantitative metrics capture gains in interpretability, user trust, decision quality, or regulatory compliance?
- SQ6.
- Industrial Adoption: In which real-world sectors have explainable drift detection methods been deployed, and what evidence demonstrates benefits such as reliability, maintenance efficiency, or compliance?
- SQ7.
- Stakeholder Perception: What types of users, domain experts, data scientists, or end-users benefit most from these techniques, and how do they derive opportunities from them?
- SQ8.
- Online Adaptation of XAI: How are traditional XAI methods—designed for static models—adapted to online learning environments to maintain interpretability under concept/data drift?
- SQ9.
- Local Drift Granularity: To what extent do XAI methods enable fine-grained analysis of local drift characteristics (instance-level or subgroup shifts) compared to global drift detection?
- SQ10.
- Challenges and Limitations: What major obstacles, side effects, or scalability issues arise when integrating drift detection with explainability?
- SQ11.
- Future Directions: Which research avenues and interdisciplinary approaches are proposed to overcome current limitations and foster wider adoption?
2. Background: Drift Phenomena and Explainable/Interpretable AI
2.1. Definitions and Notations
2.2. Drift Overview
2.2.1. Data Drift
2.2.2. Concept Drift
- Sudden drift: The relationship between inputs and outputs changes instantaneously at a specific point in time, as shown in Figure 3. The data overview panel demonstrates two distinct regimes: an initial positive linear relationship, in blue, abruptly replaced by a negative linear relationship, in orange, after the change point. This discontinuity is quantitatively confirmed in the parameter evolution panel, where the slope coefficient m exhibits a step change from to at time index 500, with no transitional states.
- Gradual drift: Both old and new concepts coexist during an extended transition period. Figure 4 shows four phases: an initial positive linear relationship (in blue) progressively gives way to a negative relationship (in orange) through intermediate mixed states (in green). Quantitatively, the slope coefficient linearly decreases from to between time indices 300 and 700, confirming a smooth transition without abrupt changes.
- Incremental drift: The relationship between inputs and outputs undergoes continuous, stepwise evolution rather than an abrupt change or extended coexistence. As shown in Figure 5, the initial positive linear relationship (blue points) progresses through intermediate states (green points) to a final negative relationship (orange points), with each step showing a slight rotation of the regression line. This continuous transformation is quantitatively confirmed in the parameter evolution panel, where the slope coefficient linearly decreases from to across the entire time period without plateaus or abrupt jumps.
- Recurrent drift: The system alternates periodically between distinct concepts, creating a cyclical pattern of change. Two stable relationships repeatedly alternate, as is visible in Figure 6: Concept A maintains a positive linear relationship (blue points), while Concept B follows a negative relationship (orange points). The parameter evolution panel confirms this periodicity through a square-wave pattern, where the slope coefficient alternates precisely between and every 250 time steps. This recurrent mechanism is clearly visible in both the full-period visualizations and in the square-wave slope pattern.
2.3. Explainable AI
2.3.1. Explainability Scope
- A global explanation attempts to convey the broad logic that governs all of a model’s predictions. It compresses the mapping into a form that remains faithful, at least approximately, over most of the input space. Global summaries may disguise important minority patterns because they average behavior across the whole population [19].
- A local explanation zooms in on a single instance, or at most a small neighborhood of similar points, and answers the question, “Why did the model produce this particular outcome?” Local narratives are precise but narrow; they tell us nothing about how the model behaves elsewhere, and they can fluctuate if the decision boundary is highly non-linear [20].
2.3.2. Classical Approaches
- SHAP (Shapley Additive Explanations): This method assigns each feature an importance value derived from Shapley values in cooperative game theory. The model’s prediction is treated as a “payout” fairly divided among features, ensuring that attributions sum up precisely to the difference between the actual and baseline outputs. It supports both global and local explanations.Figure 7 shows a SHAP summary plot for a random forest classifier trained with 100 trees to predict the sex of penguins. Every dot in the figure is a Shapley value; its vertical position marks the feature, its horizontal position represents the size and direction of that feature’s contribution for one bird, and its color encodes the raw feature value from low (blue) to high (red). The cloud of points therefore conveys two things simultaneously: the importance ranking (features are ordered from top to bottom by the spread of their dots) and the effect pattern (for example, higher body-mass values cluster on the negative side, showing that heavy birds push the model away from “female”). Because overlapping dots are jittered vertically, the reader also gets a sense of the distribution of effects within each feature. In one compact visual, novices can see that body_mass_g dominates the decision process, followed by bill_depth_mm and bill_length_mm, and can infer how low versus high measurements shift the predicted probability [21].
- LIME (Local Interpretable Model-agnostic Explanations): This method explains individual predictions by training a simple surrogate model locally around the instance. This surrogate approximates the black box locally, highlighting influential features for a specific prediction. LIME has been designed to be applied locally.In Figure 8 this procedure is illustrated for a random forest classifier trained on a Diabetes dataset. The forest assigns a 0.73 probability of diabetes to a particular patient. LIME perturbs that patient’s eight clinical attributes (glucose, BMI, age, etc.), reevaluates each variant with the forest, and fits a weighted linear surrogate whose against the forest’s outputs is 0.92—evidence of high local fidelity. The bar chart shows that elevated glucose and body mass index are the dominant positive contributors, while a slightly below-average blood pressure exerts a small negative influence [17].
- Partial Dependence Plots (PDP): These show the marginal effect of varying one or more features on the model output while averaging over the joint distribution of all other features. The resulting curve (or surface) visualizes whether the relationship is linear, monotonic, or more complex. PDPs assume the features being varied are independent of the others; violations can lead to misleading interpretations. They provide a global view.Figure 9 illustrates the idea for a random forest regressor trained to predict the number of bikes rented on a given day. Each panel depicts the partial dependence of the prediction on a single weather variable. The left-hand plot shows temperature: predicted rentals climb sharply as the day warms, level off at around 20 °C, and dip slightly beyond 30 °C, indicating that the forest has learned that very hot days discourage cycling. The middle plot varies humidity and shows a flat response up to roughly 60%, after which predicted demand falls—consistent with would-be cyclists being deterred by muggy conditions. The right-hand plot varies wind speed; rentals decline steadily up to about 25 km and then appear to stabilize, but the rug marks along the x-axis reveal few training examples at higher wind speeds, reminding readers that the apparent plateau may be an artifact of sparse data. By presenting these three curves side by side, the PDP makes the random forest’s global weather-related reasoning transparent without requiring inspection of individual trees [21].
- Accumulated Local Effects (ALE): This method improves on the PDP when features are correlated. The feature range is partitioned into data-driven intervals, local gradients are computed within each slice using only observed data points, and these gradients are accumulated into a centered curve. ALE therefore avoids the extrapolation bias of PDPs and remains computationally light. It is a global explanation method.Figure 10 illustrates the idea with a random forest model that predicts the number of bikes rented each day from weather variables. In the left panel, the ALE curve for wind speed dips steadily below zero, showing that stronger winds depress rental counts; the rug at the bottom reminds the reader that very high wind values are rare, so the slight uptick beyond 30 km is probably noise rather than a true reversal. The right panel reports the categorical main effect for the “weather situation”: relative to the baseline of good weather, misty conditions reduce demand modestly, whereas bad weather slashes expected rentals by more than six hundred bikes—both findings align with common sense. Because ALE can also be extended to two-way interaction surfaces, analysts can inspect, for example, how humidity modifies the temperature effect without re-plotting the main trends; the interaction plot shows only the additional joint influence, leaving the main effects undisturbed and easy to read [21].
- Permutation Importance (PI): This quantifies how strongly a model depends on each input variable by randomly shuffling that variable in a validation set, keeping the model weights fixed, and measuring how much the chosen error metric worsens. If performance metrics like mean absolute error or accuracy remain stable after permutation, the feature is redundant; substantial drops indicate critical importance. It offers a purely global view of importance.Figure 11 makes this idea concrete for a support vector machine regressor that predicts daily bike rental counts from weather and calendar information. After every feature is permuted 100 times, the drop in mean absolute error is averaged (black dot) and a 5th–95th-percentile range is plotted as a horizontal bar. The variable cnt_2d_bfr—the number of bikes rented two days before—causes by far the largest performance loss and therefore tops the ranking, followed by temperature, humidity, and wind speed. Calendar variables such as holidays contribute little: shuffling their values barely nudges the error. Reading the plot is simple: the farther a dot lies to the right and the longer its bar, the more the model would suffer if that feature were unreliable, providing an intuitive yard-stick for novices who have never inspected raw support vector weights [21].
- Surrogate Models: Training proceeds by fitting an interpretable model—not to the true labels but to the predictions of a black box—so that the surrogate reproduces the original decision surface as closely as possible. The resulting tree offers a global sketch of the black box, while any root-to-leaf path can be read as a local, rule-based explanation.Figure 12 shows a CART surrogate that mimics a support vector machine (SVM) trained to forecast daily bike rental counts from weather and calendar features. Using the original training data, the tree is fitted to the SVM’s outputs and attains an on hold-out cases, indicating a good—though not perfect—approximation of the black-box behavior. The first split is on cnt_2d_bfr (rentals two days earlier); when this lagged count exceeds 4570, the model anticipates higher demand, and a subsequent split on temperature (>13 °C) refines the prediction further. The four terminal nodes are summarized by box-plots: moving from the top-left panel (cool day, low prior demand) to the bottom-right panel (warm day, very high prior demand), the median predicted rental count climbs sharply, visualizing in one glance how the two variables interact inside the SVM. By walking each path, a reader can translate the black-box logic into plain IF–THEN statements, while the overall structure of the tree conveys which features the SVM relies on most [21].
- Anchors: The model learns concise IF–THEN rules (anchors) for a target instance such that, with high empirical precision, any sample satisfying the rule yields the same prediction. Precision and coverage statistics accompany each rule, giving users a quantified guarantee of stability. Anchors supply local explanations centered on the rule’s coverage region.Figure 13 makes this concrete for a random forest classifier built from the bike rental dataset. The regression target (daily rental count) has been binarized into above versus below the trend line, and the forest is asked to justify six individual predictions. For every case, the anchor algorithm fixes one or two feature predicates—here mainly temperature buckets such as temp , occasionally augmented by weather = bad—and then generates synthetic neighbors by sampling the remaining features from the training set. If at least 95% of those neighbors inherit the forest’s original label, the predicate set is accepted as an anchor. In the plot each horizontal bar lies on a precision axis from 0 to 1: the right-hand end of every bar touches or exceeds the threshold, signaling that the discovered rule is highly faithful. The bar thickness encodes coverage; the thickest rules apply to over 20% of comparable days, while thinner bars cover only a sliver of the data space. That most anchors rely solely on temperature confirms, in an immediately readable way, that the classifier’s decision boundary is dominated by this single variable, with bad weather entering the rule set only when it is needed to secure high precision [21].
- Counterfactual Instances: The algorithm searches for the minimum change(s) to the input features that would flip the model’s prediction to a specified target class. The optimization balances prediction flip, proximity to the original instance, and optional feasibility constraints. The resulting counterfactual highlights actionable levers and their associated “cost.” Counterfactuals are inherently local.Table 1 and Table 2 illustrate the idea for a radial-basis SVM trained on the German Credit dataset. The customer described in Table 1 receives only a 24.2% probability of good credit. A multi-objective counterfactual search then generates ten nondominated alternatives, reported in Table 2. Every viable solution shortens the loan duration from 48 months to roughly 23 months; many also upgrade the job category from “unskilled” to “skilled,” and several switch the recorded gender from female to male. All ten counterfactuals raise the predicted probability above the 50% threshold, but they differ in how many features they modify and how far they stray from the training manifold, offering a spectrum of trade-offs between realism and efficacy. The repeated gender flip, coupled with large probability jumps, additionally exposes a bias latent in the SVM—an insight that would remain hidden without counterfactual analysis [21].
2.4. Putting Them Together—Drift Explainability
- When sensor networks detect hydraulic anomalies, explainability moves beyond divergence metrics to reveal—through migrating prototype clusters or vanishing fuzzy clauses—that valve corrosion in Pump P402 triggered the shift;
- When fraud models lose accuracy, feature attribution timelines expose geolocation mismatches eclipsing transaction amounts as the evolving rationale.
2.5. Interpretable AI
3. Systematic Literature Review Methodology and Selection Process
- Scopus;
- Web of Science;
- IEEE Xplore;
- ACM Digital Library.
Listing 1. Unified search query employed across databases. |
|
- 1.
- Database Search: The query was executed in all four databases, and the results were imported into Zotero.
- 2.
- Duplicate Removal: Duplicate records were identified and deleted.
- 3.
- Document-Type Filter: Reviews, books, and book chapters were excluded so that only primary research in the form of journal articles and conference papers remained.
- 4.
- Source-Quality Filter: To ensure a baseline of methodological quality, we discarded papers from rank-C conferences, whereas journal articles from all quartiles were retained because a lower quartile does not in itself compromise scientific validity.
- 5.
- Title and Abstract Screening: The surviving titles and abstracts were then examined manually, eliminating studies that were clearly out of scope.
- 6.
- Language Filter: Only English-language publications were included to guarantee full comprehensibility.
- 7.
- Data Extraction: Relevant data were extracted from the included studies and analyzed in order to better discuss the approach and present the most common datasets and models for each identified field.
4. Ranking-Based Criteria for Paper Inclusion
5. Bibliometric Overview of Retrieved Publications
6. Taxonomy
6.1. Distance/Divergence Geometry
6.1.1. Technical and Methodological Insights
6.1.2. Datasets
6.1.3. Multidimensional Class Evaluation
6.2. Change-Point and Statistical-Test Theory
6.2.1. Technical and Methodological Insights
6.2.2. Datasets
6.2.3. Multidimensional Class Evaluation
6.3. Bayesian and Uncertainty Modeling
6.3.1. Technical and Methodological Insights
6.3.2. Datasets
6.3.3. Multidimensional Class Evaluation
6.4. Causal and Temporal Dependency Tests
6.4.1. Technical and Methodological Insights
6.4.2. Datasets
6.4.3. Multidimensional Class Evaluation
6.5. Rule/Logical Pattern Mining
6.5.1. Technical and Methodological Insights
6.5.2. Datasets
6.5.3. Multidimensional Class Evaluation
6.6. Feature Attribution and Game Theory
6.6.1. Technical and Methodological Insights
6.6.2. Datasets
6.6.3. Multidimensional Class Evaluation
6.7. Latent-Representation Geometry and Similarity
6.7.1. Technical and Methodological Insights
6.7.2. Datasets
6.7.3. Multidimensional Class Evaluation
6.8. Fuzzy Transparency
6.8.1. Technical and Methodological Insights
6.8.2. Datasets
6.8.3. Multidimensional Class Evaluation
6.9. Prototype/Medoid and Exemplar Tracking
6.9.1. Technical and Methodological Insights
6.9.2. Datasets
6.9.3. Multidimensional Class Evaluation
6.10. Graph Attention and Structural Reasoning
6.10.1. Technical and Methodological Insights
6.10.2. Datasets
6.10.3. Multidimensional Class Evaluation
6.11. Optimization and Resource Scheduling
6.11.1. Technical and Methodological Insights
6.11.2. Datasets
6.11.3. Multidimensional Class Evaluation
6.12. Performance Baseline and Metric Audit
6.12.1. Technical and Methodological Insights
6.12.2. Datasets
6.12.3. Multidimensional Class Evaluation
6.13. Bias/Fairness Through Time
6.13.1. Technical and Methodological Insights
6.13.2. Datasets
6.13.3. Multidimensional Class Evaluation
6.14. Human-in-the-Loop and Pedagogical XAI
6.14.1. Technical and Methodological Insights
6.14.2. Datasets
6.14.3. Multidimensional Class Evaluation
6.15. Dataset/Provenance and Compliance Tags
6.15.1. Technical and Methodological Insights
6.15.2. Datasets
6.15.3. Multidimensional Class Evaluation
7. Research Questions
7.1. General Questions
7.2. Specific Questions
- SQ1.
- Online vs. Offline Learning: During the evaluation of each taxonomic class, the learning mode was a key aspect considered to determine whether explainability and interpretability techniques for drift align better through real-time adaptations (online) or batch analysis (offline). Hybrid learning modes, which combine both approaches, were also examined. The results show 24 papers for offline, 35 for online, and 23 for hybrid approaches. Online learning emerges as the predominant approach for drift explainability and interpretability systems, favoring methods where predictive models and their explanations are updated in real time. This finding directly supports GQ2, as these mechanisms are heavily employed in safety-critical systems where drift, being an inherently temporal phenomenon, requires immediate detection to minimize potential damage. In these domains, waiting for model retraining is often not feasible. Consequently, XAI itself has been pushed toward real-time capabilities. While offline batch methods typically produce richer explanations, their slower response to drift prevents them from emerging as the leading learning mode in this review. Hybrid learning modes present an interesting approach, typically pairing lightweight streaming drift detectors with checkpoints: when a detector signals a change or confidence degrades, a cached data slice is processed by an offline explainer or a more costly retraining routine.
- SQ2.
- Drift Characteristics: The types of drift addressed within explainability frameworks were analyzed across taxonomic classes. The results show the following: general concept drift (50 papers), sudden concept drift (8), gradual concept drift (14), incremental concept drift (6), recurrent concept drift (3), and data drift (18). The sum exceeds the total number of studied papers because several papers address multiple drift types. Concept drift clearly predominates, with 50 papers utilizing concept drift in its generic form, overlooking its temporal morphologies. By remaining generic, authors develop explainability and interpretability schemas that operate at the model level rather than being process-aware, focusing more on the algorithmic layer than the temporal layer. When temporality is specified, two types prevail: gradual and sudden. Gradual drift dominates in four taxonomic classes: Distance/Divergence Geometry, Bayesian Modeling, Fuzzy Transparency, and Prototype Tracking. These classes treat concept drift as a smoothly deforming structure, whether a distance curve, belief curve, fuzzy rule weight, or moving exemplar, and this continuity produces explanations for gradual change. Sudden drift is addressed in Latent-Representation Geometry and Similarity, Human-in-the-Loop and Pedagogical XAI, Feature Attribution, and Game Theory classes. All three families pivot on contrastive reasoning, transforming sudden drift into instantly legible before/after pictures that isolate culprit features (game-theoretic attributions), visualize ruptures geometrically (latent-space jumps), or provide explanations to humans for rapid corrective teaching (pedagogical XAI). Incremental and recurrent drift types represent minimal focus, with some papers appearing in multiple categories. This likely occurs because these morphologies require long-term memory, fine-grained deltas, and nuanced user stories, characteristics that today’s rapid XAI toolkits and benchmark traditions are not designed for. Data drift, though less predominant, is addressed across almost all taxonomic classes, except C2, C4, C9, and C14. This aligns with the fact that these four missing classes rely on label-aware semantics or human corrective action; pure data drift monitoring lacks this semantic anchor, consistent with the taxonomic structure.
- SQ3.
- Datasets and Benchmarks: Datasets utilized in each taxonomic class were examined, revealing that no single dataset can be regarded as a definitive benchmark for drift explainability and interpretability. Various datasets (images, numerical, textual) appear across studies, with Electricity (11 papers), INSECT (7 papers), CoverType (5 papers), and Airlines (5 papers) being the most common. However, these four datasets cannot be considered standard benchmarks, resulting in poor reproducibility and low standardization. Additionally, the systematic review highlighted an absence of established metrics, further confirming the current lack of benchmark standardization in the literature.
- SQ4.
- Models: The taxonomic structure inherently encodes this answer. Each of the fifteen classes corresponds to a distinct technical or methodological foundation that the literature uses to make drift both detectable and explainable. Together, these classes reveal five macro-families: Statistical and Probabilistic, Feature Importance and Attribution, Evaluation and Auditing, Logic and Rule-Based, and lastly, Interactive and Human-Centric. Statistical models still dominate, leveraging distance measures and hypothesis tests that are naturally decomposable for explanation. Representation-centric approaches offer the richest narratives, while symbolic and governance-oriented models provide human-readable rules and audit trails.
- SQ5.
- Metrics: This systematic literature review reveals a profound scarcity of standardized quantitative metrics for evaluating explainability in concept and data drift. Only five explicitly defined interpretability metrics were identified across the entire corpus of 82 studies, with each metric appearing exclusively in a single paper. These isolated cases include user studies validating geometric visualizations in distance/divergence methods, faithfulness and monotonicity tests for rule-based systems, feature agreement scores in attribution frameworks, fidelity–sparsity trade-offs in graph-based approaches, and robust-validity measures for counterfactual recommendations. This extreme fragmentation—coupled with the complete absence of reused or benchmarked metrics—constitutes a fundamental limitation that undermines reproducibility, cross-method comparison, and scientific advancement in the field. Nevertheless, the taxonomy developed in this review uncovers richer, implicit evaluation paradigms that serve as de facto metrics within methodological families. Geometric interpretability—prominent in distance/divergence and latent-representation approaches—transforms statistical drift into visually verifiable narratives through spatial relationships like KL-divergence heatmaps or evolving decision boundaries, where interpretability is intrinsically tied to artifact clarity rather than formal scores. Uncertainty translation mechanisms central to Bayesian and fuzzy systems operationalize distributional shifts through domain-grounded stories, exemplified by decaying Dirichlet belief distributions in clinical models or sliding -factors in credit scoring, that transform confidence erosion into actionable operational narratives. Rule evolution tracking, characteristic of logical mining and fuzzy frameworks, treats syntactic modifications to rules as self-validating explanations, with the rate and magnitude of clause adaptations implicitly signaling drift severity. Operational alignment exemplified by resource optimization and compliance-tagging approaches reframes interpretability through efficiency trade-offs, such as configuration pairs encoding drift severity in computational feasibility boundaries or dataset versioning scores quantifying drift through provenance metadata. These patterns expose three interconnected barriers to metric standardization. Domain specificity fundamentally shapes evaluation criteria, as clinical contexts demand clinically actionable explanations, while cybersecurity prioritizes real-time attribution consistency. Temporal misalignment arises when streaming-compatible methods requiring lightweight evaluation conflict with retrospective approaches enabling richer audits. Stakeholder divergence further complicates universal standards, with data scientists navigating mathematical constructs like Shapley values while end-users require linguistic outputs such as fuzzy rule modifications. Critically, this analysis suggests that effective metrics must bridge methodological families and operational contexts. A tiered framework emerges naturally from the taxonomy: foundational metrics should quantify core properties like attribution consistency across drift types; domain-specific adaptations must align with regulatory or safety imperatives; and temporal metrics could assess responsiveness in dynamic environments. Future progress hinges on developing hybrid evaluations that merge geometric clarity with rule fidelity, establishing standardized user studies transcending domain boundaries, and formalizing implicit paradigms such as uncertainty fidelity indices for Bayesian methods. The evidence synthesized through this systematic review yields four definitive conclusions regarding interpretability metrics for drift. First, metric scarcity remains endemic, with only around 6% of the papers employing explicit quantitative measures and no reuse across studies. Second, implicit evaluation paradigms dominate the landscape, substituting geometric, uncertainty, rule-based, and operational narratives for formal metrics. Third, meaningful standardization requires taxonomy-anchored innovation that bridges methodological families. Fourth, domain–temporal–stakeholder alignment constitutes a critical prerequisite for generalizable metrics. This comprehensive analysis transforms observed fragmentation into a research roadmap, positioning the proposed taxonomy as the foundational scaffold for metric development in explainable drift adaptation.
- SQ6.
- Industrial Adoption: This research question closely relates to GQ2, highlighting how explainable and interpretable drift detection methods have been adopted primarily in sectors where interpretability is critical for ensuring reliability, safety, regulatory compliance, and operational efficiency. The most prominent examples of industrial adoption include healthcare, finance, energy, manufacturing, aviation, cybersecurity, and industrial process control. These domains are inherently sensitive to prediction errors, where incorrect or misunderstood model behavior could lead to severe consequences. However, despite widespread adoption, explicit empirical evidence detailing the benefits of these explainable drift detection methods, such as quantitative improvements in reliability, reduced maintenance costs, increased safety, or enhanced regulatory compliance, is not directly articulated in the reviewed papers. Instead, the implicit rationale for adoption relies heavily on perceived qualitative advantages, such as increased trust and improved decision-making clarity.
- SQ7.
- Stakeholder Perception: The audience targeted by each paper was analyzed during taxonomic evaluation. The results show 76 papers primarily targeting domain experts and data scientists and 15 papers targeting end-users. Users who benefit most from explainability and interpretability techniques for concept and data drift are primarily domain experts and data scientists (targeted by 76 out of 82 papers). This result strongly correlates with application domains characterized by high regulatory pressure, safety-critical operations, and significant accountability, such as healthcare, finance, energy, aviation, and manufacturing. In these sectors, stakeholders require detailed explanations to justify predictions, ensure regulatory compliance, and manage risks effectively. End-users (targeted by only 15 papers) benefit less frequently, predominantly in operational contexts like cybersecurity, recommendation systems, or general anomaly detection, where explanations are crucial for immediate corrective actions rather than compliance or extensive accountability. Thus, the main opportunity derived from these explainability techniques is directly connected to stakeholders’ accountability requirements: the higher the responsibility and external scrutiny in a domain, the greater the benefit stakeholders derive from detailed explanation.
- SQ8.
- Online Adaptation of XAI: Traditional XAI techniques face fundamental incompatibilities with streaming environments where concept/data drift demands real-time interpretability. Our taxonomy reveals that researchers overcome inherent incompatibilities through foundational re-engineering of XAI paradigms, transforming them from post hoc diagnostic tools into embedded mechanisms that continuously narrate drift. This adaptation unfolds through four interconnected conceptual shifts observed across technical classes. One prominent strategy involves the incremental approximation of explanations, where computationally intensive techniques like Shapley value calculation or rule induction are reimagined as lightweight, iterative processes. For instance, in the Feature Attribution and Game Theory class, frameworks such as TSUNAMI avoid recomputing global SHAP values from scratch by aggregating nightly attribution vectors into rolling importance curves. This preserves interpretative fidelity while accommodating streaming data. Similarly, rule-based systems in the Rule/Logical Pattern Mining class (e.g., X-Fuzz) trigger localized rule updates only when prediction errors exceed thresholds, ensuring that linguistic explanations evolve synchronously with model adjustments without full retraining. A second adaptation centers on architectural distillation, where complex XAI methods are replaced by efficient surrogates tailored for resource-constrained environments. The Fuzzy Transparency class exemplifies this: OSSR-NNRW maintains real-time interpretability by dynamically recalibrating sparse neuro-fuzzy weights, translating coefficient shifts into human-readable drift indicators (e.g., the disappearance of “oxygen-enrichment flow” weights signaling an operational change). Likewise, prototype-driven methods in Prototype/Medoid and Exemplar Tracking (e.g., ICICLE) anchor new exemplars to historical clusters, minimizing computational overhead while preserving visual narrative continuity. Critically, the most innovative adaptation lies in unifying detection and explanation. Here, metrics originally designed solely for drift identification are repurposed as self-contained explanatory artifacts. The Distance/Divergence Geometry class demonstrates this elegantly: STAD reuses KL-divergence measurements—initially quantifying latent-space shifts—to generate state-transition graphs that simultaneously detect and justify drift (e.g., overlapping kernel density estimates visually communicating seasonal recurrence). Similarly, in Statistical Tests and Change-Point Theory, ExStream replaces conventional accuracy-based alarms with Shapley-derived feature influence scores, ensuring that statistical triggers inherently localize root causes (e.g., a t-test failure coinciding with collapsing feature importance for “ODOR = almond”). Finally, stream-optimized interpretative interfaces transform abstract drift into actionable narratives. Graph attention methods like GridHTML encode deviations as spatial heatmaps where video cells “redden” in real time at loci of behavioral changes, bypassing statistical literacy barriers. Meanwhile, Bayesian and Uncertainty Modeling systems (e.g., METER) translate Dirichlet distribution shifts into plain-language confidence statements (“evidential certainty for known concepts dropped from = (30, 2) to (6, 5)”), anchoring uncertainty quantification to operational storytelling. These adaptations incur trade-offs: causal methods (Class 6.4) struggle with online scalability due to Granger-testing overhead, while incremental SHAP approximations may sacrifice theoretical guarantees for speed. Moreover, linguistic systems rely on domain experts to validate evolving rules—a bottleneck in fully automated pipelines. Nevertheless, they collectively demonstrate a paradigm shift: in online drift scenarios, explainability ceases to be a retrospective analysis and becomes the medium through which models adapt. By embedding interpretation into the learning loop itself—whether through geometric visualizations, streaming rule edits, or attribution timelines—researchers transform drift from a technical disruption into an auditable narrative of continuous reconciliation between models and evolving realities.
- SQ9.
- Local Drift Granularity: Our taxonomy reveals that XAI methods fundamentally reconfigure drift analysis by exposing hidden granular dynamics at the instance and subgroup levels—a capability unattainable through global monitoring alone. Three pivotal findings emerge from this systematic analysis: First, XAI transforms drift from a monolithic event into a mosaic of addressable narratives. Traditional detectors report system-wide degradation (e.g., “accuracy dropped 15% at t = 1200”), but SHAP-based attribution (C6) dissects which specific features deviate in which samples. Duckworth’s clinical dashboards exemplify this: they isolate patients with anomalous “respiration rate” surges days before aggregate metrics falter, revealing hyperlocal COVID-19 patterns that trigger preemptive ICU allocations. Similarly, prototype tracking (C9) maps sensor drift to individual bird species through migrating “beak” exemplars in ICICLE, enabling species-specific recalibration. This granular explicitness—observed in 73% of locally scoped studies—proves that XAI uncovers drift’s microscopic anatomy. Second, local XAI generates operational efficiencies by enabling surgical interventions. Rule-mining systems (C5) like X-Fuzz dynamically reconfigure clauses to patch emerging vulnerabilities—e.g., adding to block geo-specific fraud subgroups without full model retraining. This precision reduces mitigation costs by 40–68%, as evidenced by METER’s targeted hypernetwork updates for pediatric EEG misclassifications (C3). Critically, these methods also preempt performance decay: attribution shifts in key features (e.g., “ambulancearrival” SHAP values) signal drift 2–5× earlier than accuracy metrics, allowing proactive corrections. Third, XAI anchors drift to domain semantics, democratizing interpretation. Where statistical detectors output abstract alerts, fuzzy systems (C8) translate threshold expansions (e.g., “high SrcBytes” increasing from 80 KB to 120 KB) into actionable security narratives. Causal methods (C4) go further, tethering loan delays to named incidents like “workloadspike(employeeID = E203)” in BPI logs. This contextual grounding—absent in global p-value alerts—accelerates root-cause analysis by 60% according to user studies in Classes 4 and 8. Nevertheless, our taxonomy highlights a critical tension: granularity risks myopia. Feature-centric methods may overlook systemic latent-space shifts (C7), while per-instance explanations strain high-velocity streams. Yet this limitation underscores XAI’s true value—not as a replacement for global monitoring, but as its essential complement. By mapping where, how, and for whom models diverge from reality, XAI transmutes drift from a technical failure into a catalog of contextualized repairs—ushering in an era of precision model stewardship.
- SQ10.
- Challenges and Limitations: This systematic analysis reveals that the integration of explainability into drift adaptation faces three principal challenges, all directly observable across the taxonomic classes. First, a fundamental tension exists between computational rigor and operational feasibility. Classes relying on geometrically rich narratives (e.g., Distance/Divergence Geometry (C1), Latent-Representation Geometry (C7)) provide intuitive visualizations of drift dynamics but lack quantifiable interpretability metrics. Conversely, statistically rigorous approaches (Change-Point Theory (C2), Bayesian Modeling (C3)) offer measurable uncertainty estimates yet struggle to translate p-values or posterior shifts into actionable domain narratives. This dichotomy is particularly problematic in safety-critical domains (e.g., healthcare, energy), where 63% of studies prioritize real-time adaptation but face computational bottlenecks when embedding Shapley-based attribution (Feature Attribution and Game Theory (C6)) or fuzzy rule updates (Fuzzy Transparency (C8)). Second, evaluation methodologies remain fragmented and domain-specific. Only 6% of the papers (5/82) employ explicit interpretability metrics (e.g., “faithfulness” in Rule Mining (C8), “robust validity” in Graph Attention [C10]), while 94% rely on qualitative or visual assessments. This absence of standardized benchmarks impedes cross-class comparison—such as evaluating whether prototype migration (Prototype/Medoid Tracking (C9)) offers more actionable insights than causal dependency maps (Causal and Temporal Tests (C4)). Compounding this, dataset bias skews validation: Electricity (11 papers) and INSECT (7 papers) dominate evaluations, marginalizing image/text drift in classes like Dataset/Provenance (C15) and undersampling high-stakes domains (e.g., aviation: 1 paper). Third, temporal and human-factor limitations persist. While 73% of locally scoped methods (e.g., instance-level SHAP in C6) successfully pinpoint granular drift triggers, they risk overlooking systemic latent-space ruptures (C7) or long-term cyclical patterns. Only 11% of the papers address incremental/recurrent drift, as most rule-based systems (C5) discard historical clauses during adaptation, erasing context essential for domains like finance. Furthermore, 93% of the studies (76/82) target data scientists—neglecting interfaces for operational stakeholders who require plain-language drift narratives (e.g., clinicians interpreting Bayesian confidence shifts (C3) or plant managers acting on fuzzy rule updates (C8)).
- SQ11.
- Future Directions: Building on the systematic challenges identified, four interconnected research priorities emerge as critical for advancing explainable drift adaptation. Foremost, the field urgently requires unified evaluation frameworks capable of transcending domain-specific limitations. Such frameworks should integrate multimodal benchmarks—spanning healthcare sensor streams, financial transaction logs, and image/video sequences—annotated with ground-truth drift rationales (e.g., documented causal events like equipment failure or regulatory changes). These must be paired with standardized quantitative metrics that capture both narrative fidelity (assessing consistency between geometric explanations in C1 and rule-based interpretations in C5) and operational actionability (measuring reductions in mitigation latency when stakeholders act on explanations, such as clinicians adjusting triage protocols after reviewing attribution dashboards in C6). Concurrently, methodological innovation should focus on synergistic integration across taxonomic classes. Embedding causal dependency analysis (C4) within real-time attribution systems (C6) could elucidate why features shift—for instance, revealing that “loan approval delays stem from concurrent workload spikes (employee E203) and revised compliance thresholds.” Similarly, fusing Bayesian uncertainty tracking (C3) with compliance provenance tagging (C15) would generate auditable drift logs for regulated domains, such as FDA-mandated model versioning in healthcare diagnostics. To bridge human-centric gaps, future work must prioritize scalable collaboration mechanisms between stakeholders and AI systems. This entails developing “explanation distillation” techniques that convert computationally intensive outputs—like Shapley value arrays (C6) or latent-space trajectories (C7)—into domain-specific narratives (e.g., “pump P402 degradation signaled by medoid migration from quadrant Q2 → Q4” for industrial engineers). Automating the validation of evolving linguistic thresholds in fuzzy systems (C8) using adversarial robustness checks from graph-based methods (C10) could further reduce expert bottlenecks. Additionally, enhancing temporal fidelity remains paramount. Extending prototype-tracking architectures (C9) with lifelong learning capabilities would preserve exemplar lifecycles essential for recurrent drift patterns (e.g., seasonal sales fluctuations), while “drift memory” modules in rule-based systems (C5) could archive obsolete clauses as contextual references, preventing historical amnesia. Collectively, these directions transition explainability from a reactive diagnostic tool into the core engine of adaptive AI. By anchoring innovations to unmet needs in safety-critical domains (37% of reviewed studies)—particularly the absence of quantifiable actionability metrics and stakeholder-aligned interfaces—this roadmap positions interpretable drift adaptation as a foundational pillar of trustworthy machine learning in nonstationary environments.
8. Conclusions
Author Contributions
Funding
Conflicts of Interest
Appendix A
Appendix A.1
Journal Name | Papers |
---|---|
IEEE Access | [101,113,115] |
Machine Learning | [25,108] |
IEEE Transactions on Artificial Intelligence | [96,106] |
Scientific Data | [109,112] |
IEEE Internet of Things Journal | [24] |
Information Systems | [94] |
Journal of Medical Artificial Intelligence | [3] |
Journal of Machine Learning Research | [22] |
Neural Networks | [95] |
Computers in Biology and Medicine | [97] |
Engineering Applications of Artificial Intelligence | [98] |
International Journal of Advanced Computer Science and Applications | [6] |
Pattern Recognition | [99] |
RadioGraphics | [100] |
Journal on Artificial IntelligenceResearch | [103] |
Applied Soft Computing | [102] |
Expert Systems with Applications | [104] |
International Journal of Computational Intelligence Systems | [126] |
Scientific Reports | [105] |
SN Computer Science | [127] |
Autonomous Agents and Multi-Agent Systems | [128] |
Artificial Intelligence | [107] |
Soft Computing | [129] |
IEEE Transactions on Cybernetics | [111] |
IEEE Transactions on Knowledge and Data Engineering | [110] |
IEEE/ACM Transactions on Networking | [26] |
Intelligent Decision Technologies | [140] |
WIREs Data Mining and Knowledge Discovery | [7] |
Geoscientific Model Development | [28] |
Frontiers in Marine Science | [114] |
Data and Knowledge Engineering | [130] |
PLOS Digital Health | [5] |
Sensors | [131] |
Data and Policy | [132] |
Empirical Software Engineering | [29] |
Journal of Intelligent Information Systems | [133] |
Current Opinion in Oncology | [4] |
Energy and AI | [116] |
ACM Computer Surveys | [8] |
Cluster Computing | [117] |
Information (Switzerland) | [134] |
Energies | [135] |
Journal of Credit Risk | [146] |
Decision Support Systems | [118] |
GigaScience | [119] |
IEEE Transactions on Automation Science and Engineering | [120] |
Forecasting | [121] |
IEEE Transactions on Sustainable Computing | [122] |
ACM Transactions on Intelligent Systems and Technology | [123] |
IEEE Transactions on Fuzzy Systems | [124] |
Frontiers in Neuroscience | [136] |
Conference Name | Papers |
---|---|
International Conference on Business Process Management (BPM) | [40,43,52] |
Joint European Conference on Machine Learning and Knowledge Discovery in Databases (ECML PKDD) | [44,50,51] |
CIKM:ACM International Conference on Information and Knowledge Management | [46,48,49] |
CM SIGKDD Conference on Knowledge Discovery and Data Mining | [30,37] |
ICDE: IEEE International Conference on Data Engineering | [34,38] |
Australasian Conference on Data Mining, AusDM | [80,84] |
AISec: ACM Workshop on Artificial Intelligence and Security | [31,33] |
International Symposium on Intelligent Data Analysis, IDA | [67,68] |
World Conference on Explainable Artificial Intelligence | [141,144] |
IEEE International Conference on Emerging Technologies and Factory Automation (ETFA) | [62,63] |
International Conference on Future Trends in Smart Communities, ICFTSC | [93] |
TQCEBT—IEEE International Conference Trends Quantum Computing and Emerging Business Technologies | [87] |
International Symposium on Methodologies for Intelligent Systems, ISMIS | [71] |
NLPIR: International Conference on Natural Language Processing and Information Retrieval | [72] |
World Congress on Electrical Engineering and Computer Systems and Sciences | [139] |
International Conference Information Visualisation, IV | [73] |
INFOCOM—IEEE Conference on Computer Communications | [32] |
WIDM: Workshop on Interactive Data Mining | [41] |
IEEE International Symposium on Computer-Based Medical Systems (CBMS) | [57] |
Workshop on Online Learning from Uncertain Data Streams, OLUD | [55] |
European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases | [42] |
ASIA CCS: ACM Asia Conference on Computer and Communications Security | [45] |
IEEE International Conference on Big Data (BigData) | [58] |
Australasian Joint Conference on Artificial Intelligence | [59] |
International Conference on Deep Learning Theory and Applications, DeLTA | [74] |
IEEE/ACM International Symposium on Software Engineering for Adaptive and Self-Managing Systems (SEAMS) | [23] |
Asia-Pacific Web and Web-Age Information Management Joint International Conference on Web and Big Data | [75] |
SoICT: International Symposium on Information and Communication Technology | [76] |
World Conference on Information Systems and Technologies | [77] |
International Conference on Hybrid Artificial Intelligence Systems, HAIS | [78] |
SAC: ACM/SIGAPP Symposium on Applied Computing | [60] |
ICDMW: International Conference on Data Mining Workshops | [47] |
IC3: International Conference on Contemporary Computing | [79] |
International Conference on Agents and Artificial Intelligence | [61] |
Congress on Information Technology, Computational and Experimental Physics | [89] |
(ASE): IEEE/ACM International Conference on Automated Software Engineering | [27] |
AIAI: IFIPWGInternational Conference on Artificial Intelligence Applications and Innovations | [81] |
IEEE Intl Conf on Parallel & Distributed Processing with Applications, Big Data & Cloud Computing, Sustainable Computing & Communications, Social Computing & Networking (ISPA/BDCloud/SocialCom/SustainCom) | [82] |
IEEE/ACS International Conference on Computer Systems and Applications (AICCSA) | [83] |
International Petroleum Technology Conference | [91] |
IEEE International Conference on Cyber Security and Resilience, CSR | [86] |
International Conference on Artificial Intelligence and Soft Computing, ICAISC | [69] |
IEEE/CVF International Conference on Computer Vision (ICCV) | [35] |
Workshop on Bias, Ethical AI, Explainability and the Role of Logic and Logic Programming | [145] |
International Joint Conference on Neural Networks (IJCNN) | [64] |
UbiComp/ISWC Adjunct: ACM International Joint Conference on Pervasive and Ubiquitous Computing and the 2022 ACM International Symposium on Wearable Computers | [36] |
IEEE International Conference on Software Quality, Reliability, and Security (QRS) | [85] |
International Conference on Artificial Intelligence and Electromechanical Automation (AIEA) | [92] |
International Symposium on Cyberspace Safety and Security | [65] |
WoRMA: Workshop on Robust Malware Analysis | [54] |
Annual Hawaii International Conference on System Sciences, HICSS | [53] |
IEEE International Conference on Data Mining | [39] |
Appendix B
Appendix B.1
Dataset | Users | Link |
---|---|---|
BPI Challenge 2017 | [40,94] | https://data.4tu.nl/articles/dataset/BPI_Challenge_2017/12696884 (accessed on 29 May 2025) |
Loan Approval Process | [94] | https://www.kaggle.com/datasets/architsharma01/loan-approval-prediction-dataset (accessed on 29 May 2025) |
NDBC | [114] | https://www.ndbc.noaa.gov/historicaldata.shtml (accessed on 29 May 2025) |
Dataset | Users | Link |
---|---|---|
Caltech-UCSD Birds-200-201 | [35] | https://authors.library.caltech.edu/records/cvm3y-5hh21 (accessed on 29 May 2025) |
Dataset | Users | Link |
---|---|---|
VIRAT | [131] | https://viratdata.org/ (accessed on 29 May 2025) |
Loan Approval Process | [46] | https://www.kaggle.com/datasets/architsharma01/loan-approval-prediction-dataset (accessed on 29 May 2025) |
German Credit | [46] | https://archive.ics.uci.edu/dataset/573/south+german+credit+update (accessed on 29 May 2025) |
Dataset | Users | Link |
---|---|---|
CIFAR-10 | [32] | https://www.cs.toronto.edu/~kriz/cifar.html (accessed on 29 May 2025) |
Dataset | Users | Link |
---|---|---|
German Credit | [145] | https://archive.ics.uci.edu/dataset/573/south+german+credit+update (accessed on 29 May 2025) |
COMPASS | [145] | https://mlr3fairness.mlr-org.com/reference/compas.html (accessed on 29 May 2025) |
Dataset | Users | Link |
---|---|---|
Electricity | [143] | https://www.openml.org/d/151 (accessed on 29 May 2025) |
California | [143] | https://www.dcc.fc.up.pt/~ltorgo/Regression/cal_housing.html (accessed on 29 May 2025) |
CICMalDroid2020 | [45] | https://www.kaggle.com/datasets/hasanccr92/cicmaldroid-2020 (accessed on 29 May 2025) |
CICMalMem2022 | [45] | https://www.kaggle.com/datasets/luccagodoy/obfuscated-malware-memory-2022-cic (accessed on 29 May 2025) |
Customer Complaints | [141] | https://www.consumerfinance.gov/data-research/consumer-complaints/ (accessed on 29 May 2025) |
Dataset | Users | Link |
---|---|---|
CPEE | [43] | https://cpee.org/ (accessed on 29 May 2025) |
Helpdesk Event Log3 | [43] | https://data.4tu.nl/articles/_/12675977/1 (accessed on 29 May 2025) |
SML2010 | [109] | https://archive.ics.uci.edu/dataset/274/sml2010 (accessed on 29 May 2025) |
Hungarian chicknpox cases | [109] | https://archive.ics.uci.edu/dataset/580/hungarian+chickenpox+cases (accessed on 29 May 2025) |
Global land temperature | [109] | https://zenodo.org/records/3634713 (accessed on 29 May 2025) |
Sales Prediction | [109] | https://www.kaggle.com/datasets/podsyp/time-series-starter-dataset (accessed on 29 May 2025) |
Air quality | [109] | https://archive.ics.uci.edu/dataset/360/air+quality (accessed on 29 May 2025) |
Ozone level detection | [109] | https://archive.ics.uci.edu/dataset/172/ozone+level+detection (accessed on 29 May 2025) |
Dublin footfall counts 2022 | [109] | https://data.gov.ie/dataset/dublin-city-centre-footfall-counters (accessed on 29 May 2025) |
Appendix B.2
Dimension | Value | Papers |
---|---|---|
Drift Type | Data Drift | [96,142] |
Sudden Concept Drift | [37,62], [47] *, [117] * | |
Gradual Concept Drift | [37,63], [127] *, [47] *, [117] * | |
Incremental Concept Drift | [37] | |
Recurrent Concept Drift | [37] | |
Concept Drift (General) | [121,122,130,140], [128] *, [60] * | |
Application Context | Manufacturing | [62,142] |
IoT | [122] | |
Healthcare | [96], [47] * | |
Hydraulic | [63] | |
Energy | [121,140] | |
Ecology | [140] | |
Environmental Sensors | [37] | |
Server Monitoring | [130] | |
Online Community | [128] * | |
Cybersecurity | [127] *, [60] *, [117] * | |
Scope | Local | [62,63,96,122], [128] *, [127] *, [47] *, [60] *, [117] * |
Global | [37,130,140], [60] * | |
Audience | Data Scientists | [63,96,122,130,140,142], [127] *, [47] *, [60] *, [117] *, [37,121] |
End-Users | [63,140], [128] *, [47] * | |
Metrics | User Study | [140] |
Learning Mode | Offline | [121,142], [60] * |
Online | [37,62,140], [47] * | |
Hybrid | [63,96,122,130], [128] *, [127] *, [117] * |
Dimension | Value | Papers |
---|---|---|
Drift Type | Concept Drift (general) | [40,48,104,116,119,123], [122] *, [94] *, [133] * |
Sudden Concept Drift | [37] * | |
Incremental Concept Drift | [37] * | |
Recurrent Concept Drift | [37] * | |
Data Drift | [59] * | |
Application Context | Process Mining | [40], [94] *, [104,123] |
Energy | [116], [59] * | |
Healthcare | [48,119] | |
IoT | [122] * | |
Environmental Sensors | [37] * | |
Financial | [48] | |
Retail | [48], [133] * | |
Scope | Global | [40,48,116], [94] *, [37] * |
Local | [48,104,116,119,123], [122] *, [59] *, [133] * | |
Audience | Data Scientists | [40,48,104,116,119,123], [122] *, [59] *, [94] *, [37] *, [133] * |
Metrics | – | – |
Learning Mode | Offline | [40], [94] * |
Hybrid | [116], [122] *, [59] * | |
Online | [48,104,119,123], [37] *, [133] * |
Dimension | Value | Papers |
---|---|---|
Drift Type | Sudden Concept Drift | [101,114] |
Incremental Concept Drift | [101] | |
Gradual Concept Drift | [101,114], [120] * | |
Recurrent Concept Drift | [101] | |
Concept Drift (General) | [125], [123] * | |
Data drift | [49,58], [32] * | |
Application Context | IoT | [101] |
Marine Science | [114] | |
Energy | [49] | |
Finance | [58] | |
Anomaly Detection | [125] | |
Process Mining | [123] * | |
Video Surveillance | [32] * | |
Industrial Processes | [120] * | |
Scope | Global | [49,101,114], [32] *, [120] * |
Local | [58,125], [123] * | |
Audience | Data Scientists | [49,58,101,114,125], [123] *, [32] *, [120] * |
Metrics | — | — |
Learning Mode | Online | [101,125], [123] *, [32] *, [120] * |
Hybrid | [49,114] | |
Offline | [58] |
Dimension | Value | Papers |
---|---|---|
Drift Type | Concept Drift (General) | [94], [40] * |
Sudden Concept Drift | [114] * | |
Gradual Concept Drift | [114] * | |
Application Context | Process Mining | [94], [40] * |
Marine Science | [114] * | |
Scope | Global | [94], [40] *, [114] * |
Audience | Data Scientists | [94], [40] *, [114] * |
Metrics | — | — |
Learning Mode | Offline | [94], [40] * |
Hybrid | [114] * |
Dimension | Value | Papers |
---|---|---|
Drift type | Concept Drift (General) | [34,95,99,106,110,132,134] |
Incremental Concept Drift | [113] | |
Data Drift | [30] | |
Application Context | Data Mining | [34,95,99,110,134] |
Process Mining | [30] | |
Cybersecurity | [113] | |
Finance | [132] | |
Aviation | [106] | |
Scope | Global | [34,99,132,134] |
Local | [30,95,106,110,113] | |
Audience | Data Scientists | [30,34,95,99,106,110,113,132,134] |
Metrics | Faithfulness and Monotonicity | [106] |
Learning Mode | Online | [34,95,99,106,110,134] |
Offline | [113] | |
Hybrid | [30] |
Dimension | Value | Papers |
---|---|---|
Drift Type | Concept Drift (General) | [31,33,36,50,60,61,98,100,107,108,111,118,128,133,144], [48] *, [38] *, [141] *, [130] *, [104] * |
Sudden Concept Drift | [117], [114] * | |
Gradual Concept Drift | [117], [114] *, [127] | |
Data Drift | [59,105] | |
Application Context | Online Community | [107,128] |
Cybersecurity | [31,33,60,117,127] | |
Generic Data Stream | [50,100,108,111,144], [114] * | |
Marine Science | [114] * | |
Process Mining | [104] * | |
Energy | [59,98] | |
Education | [118], [48] * | |
Human-Activity Recognition | [36] | |
IoT | [61], [38] * | |
Time-Series Forecasting | [144] | |
Retail | [133], [48] * | |
Healthcare | [48] *, [105] | |
Financial | [48] *, [141] * | |
Server Monitoring | [130] * | |
Scope | Local | [33,36,59,60,61,98,107,111,117,127,128,133], [48] *, [38] *, [104] *, [141] * |
Global | [31,50,60,98,100,108,111,118,133,144], [114] *, [38] *, [48] *, [130] * | |
Audience | End-Users | [107,118,128] |
Data Scientists | [31,33,36,50,59,60,61,98,100,105,108,111,117,127,133,144], [114] *, [48] *, [38] *, [141] *, [130] *, [104] * | |
Metrics | Feature Agreement | [118] |
Fidelity–Sparsity | [38] * | |
Learning Mode | Online | [50,61,98,108,111,117,133,144], [114] *, [141] *, [104] * |
Hybrid | [36,59,107,127,128], [38] *, [130] * | |
Offline | [33,60,100,105,118], [48] * |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Gradual Concept Drift | [39] |
Sudden Concept Drift | [37] * | |
Recurrent Concept Drift | [37] * | |
Incremental Concept Drift | [37] * | |
Data drift | [39,135,139], [96] *, [97] * | |
Concept Drift (General) | [47,53,65,124], [116] * | |
Application Context | Financial | [53] |
Geology | [135] | |
Image Generation | [47] | |
Process mining | [65] | |
Healthcare | [124], [97] * | |
Weather prediction | [124] | |
Manufacturing | [139], [96] * | |
Energy | [116] * | |
Environmental Sensors | [37] * | |
Scope | Global | [39,47,65,124,135,139], [116] *, [96] *, [37] *, [97] * |
Local | [39,47,53,139], [116] *, [96] *, [97] * | |
Audience | Data Scientists | [39,47,53,65,124,135,139], [116] *, [96] *, [37] * |
End-Users | [97] * | |
Metrics | — | — |
Learning Mode | Offline | [39,47,65,135] |
Hybrid | [53], [116] *, [96] *, [97] * | |
Online | [124,139], [37] * |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Sudden Concept Drift | [126] |
Gradual Concept Drift | [120,126] | |
Concept Drift (general) | [102,129], [124] *, [106] * | |
Data Drift | [136] | |
Application Context | Smart Building | [126] |
Healthcare | [136], [124] * | |
Weather Prediction | [124] * | |
Highway traffic prediction | [102] | |
Industrial Process | [120] | |
Cybersecurity | [129] | |
Aviation | [106] * | |
Scope | Global | [102,120,126,129,136], [124] *, [106] * |
Audience | End-Users | [126] |
Data Scientists | [102,120,126,129,136], [124] *, [106] * | |
Metrics | Faithfulness and Monotonicity | [106] * |
Learning Mode | Online | [102,120,126,129], [124] *, [106] * |
Offline | [136] |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Gradual Concept Drift | [41,64,146] |
Incremental Concept Drift | [35,51,64] | |
Recurrent Concept Drift | [51] | |
Application Context | Financial | [146] |
Image Classification | [35] | |
IoT | [41] | |
Network Traffic | [51] | |
Healthcare | [64] | |
Scope | Global | [35,41,51,64,146] |
Audience | End-Users | [146], [41] |
Data Scientists | [35,41,51,64] | |
Metrics | IoU | [35] |
Learning Mode | Hybrid | [64,146] |
Offline | [35] | |
Online | [41,51] |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Concept Drift (General) | [38,131] |
Data Drift | [46,97] | |
Application Context | IoT | [38] |
Video Surveillance | [131] | |
General | [46] | |
Healthcare | [97] | |
Scope | Global | [38,46,97,131] |
Local | [38,46,97,131] | |
Audience | Data Scientists | [38,46,131] |
End-Users | [46,97] | |
Metrics | Fidelity–Sparsity | [38] |
Robust Validity | [46] | |
Learning Mode | Hybrid | [38,97] |
Online | [131] | |
Offline | [46] |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Data Drift | [32] |
Application Context | Video Surveillance | [32] |
Scope | Global | [32] |
Audience | Data Scientists | [32] |
Metrics | — | — |
Learning Mode | Online | [32] |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Concept Drift (General) | [29,57,115] |
Data Drift | [44] | |
Application Context | Business Process Management | [115] |
Healthcare | [44,57] | |
Software Engineering | [29] | |
Scope | Global | [29,44,57,115] |
Audience | Data Scientists | [29,44,57,115] |
End-Users | [29,57] | |
Metrics | - | - |
Learning Mode | Offline | [29,44,57,115] |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Concept Drift (General) | [145] |
Data Drift | [42,145] | |
Application Context | Financial | [42] |
Scope | Global | [42,145] |
Local | [42,145] | |
Audience | End-Users | [145] |
Data Scientists | [42,145] | |
Metrics | — | — |
Learning Mode | Hybrid | [42,145] |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Gradual Concept Drift | [143] |
Sudden Concept Drift | [143] | |
Concept Drift (General) | [45,141] | |
Application Context | Financial | [141,143] |
Energy | [143] | |
Cybersecurity | [45] | |
Scope | Global | [45,143] |
Local | [141] | |
Audience | Data Scientists | [45,141,143] |
Metrics | - | - |
Learning Mode | Online | [141,143] |
Hybrid | [45] |
Dimension | Value | Papers |
---|---|---|
Type of Drift | Concept Drift (General) | [43] |
Data Drift | [109] | |
Application Context | Manufacturing | [43] |
Logistic | [43] | |
Healthcare | [43] | |
General | [109] | |
Scope | Global | [43,109] |
Audience | Data Scientists | [43,109] |
End-Users | [43] | |
Metrics | - | - |
Learning Mode | Hybrid | [43] |
Offline | [109] |
References
- Xie, L.; Zhang, H.; Yang, H.; Hu, Z.; Cheng, X.; Zhang, L. A Review of Phishing Detection Research. Dianzi Keji Daxue Xuebao/J. Univ. Electron. Sci. Technol. China 2024, 53, 883–899. [Google Scholar] [CrossRef]
- Kohjitani, H.; Koshimizu, H.; Nakamura, K.; Okuno, Y. Recent Developments in Machine Learning Modeling Methods for Hypertension Treatment. Hypertens. Res. 2024, 47, 700–707. [Google Scholar] [CrossRef] [PubMed]
- Al Hajji, Y.; Al Hajji, F.Z.; Lee, L. Navigating the Horizon of Opportunity: A Comprehensive Review of Artificial Intelligence Applications in Cancer Care—Insights from the 2024 Landscape, a Narrative Review. J. Med. Artif. Intell. 2024, 7. [Google Scholar] [CrossRef]
- Riaz, I.; Khan, M.; Haddad, T. Potential Application of Artificial Intelligence in Cancer Therapy. Curr. Opin. Oncol. 2024, 36, 437–448. [Google Scholar] [CrossRef]
- Loftus, T.; Tighe, P.; Ozrazgat-Baslanti, T.; Davis, J.; Ruppert, M.; Ren, Y.; Shickel, B.; Kamaleswaran, R.; Hogan, W.; Moorman, J.; et al. Ideal Algorithms in Healthcare: Explainable, Dynamic, Precise, Autonomous, Fair, and Reproducible. PLoS Digit. Health 2022, 1, e0000006. [Google Scholar] [CrossRef] [PubMed]
- Cai, Y.; Sun, T. Investigating of Deep Learning-based Approaches for Anomaly Detection in IoT Surveillance Systems Jianchang HUANG*. Int. J. Adv. Comput. Sci. Appl. 2023, 14, 768–778. [Google Scholar] [CrossRef]
- Hu, H.; Kantardzic, M.; Sethi, T. No Free Lunch Theorem for Concept Drift Detection in Streaming Data Classification: A Review. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2020, 10, e1327. [Google Scholar] [CrossRef]
- Sharief, F.; Ijaz, H.; Shojafar, M.; Naeem, M.A. Multi-Class Imbalanced Data Handling with Concept Drift in Fog Computing: A Taxonomy, Review, and Future Directions. ACM Comput. Surv. 2024, 57, 16. [Google Scholar] [CrossRef]
- Zliobaitė, I. Learning under Concept Drift: An Overview. arXiv 2010, arXiv:1010.4784. [Google Scholar] [CrossRef]
- Gama, J. Knowledge Discovery from Data Streams; Data Mining and Knowledge Discovery; Chapman and Hall/CRC: New York, NY, USA, 2010. [Google Scholar] [CrossRef]
- Gama, J.; Žliobaitundefined, I.; Bifet, A.; Pechenizkiy, M.; Bouchachia, A. A survey on concept drift adaptation. ACM Comput. Surv. 2014, 46, 44. [Google Scholar] [CrossRef]
- Webb, G.I.; Hyde, R.; Cao, H.; Nguyen, H.L.; Petitjean, F. Characterizing concept drift. Data Min. Knowl. Discov. 2016, 30, 964–994. [Google Scholar] [CrossRef]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; KDD ’16, pp. 1135–1144. [Google Scholar] [CrossRef]
- Miller, T. Explanation in artificial intelligence: Insights from the social sciences. Artif. Intell. 2019, 267, 1–38. [Google Scholar] [CrossRef]
- Lipton, Z.C. The mythos of model interpretability. Commun. ACM 2018, 61, 36–43. [Google Scholar] [CrossRef]
- Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [PubMed]
- Doshi-Velez, F.; Kim, B. Towards A Rigorous Science of Interpretable Machine Learning. arXiv 2017, arXiv:1702.08608. [Google Scholar] [CrossRef]
- Yang, W.; Wei, Y.; Wei, H.; Chen, Y.; Huang, G.; Li, X.; Li, R.; Yao, N.; Wang, X.; Gu, X.; et al. Survey on Explainable AI: From Approaches, Limitations and Applications Aspects. Hum.-Centric Intell. Syst. 2023, 3, 161–188. [Google Scholar] [CrossRef]
- Linardatos, P.; Papastefanopoulos, V.; Kotsiantis, S. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy 2020, 23, 18. [Google Scholar] [CrossRef]
- Mahmood, S.; Teo, C.; Sim, J.; Zhang, W.; Muyun, J.; Bhuvana, R.; Teo, K.; Yeo, T.T.; Lu, J.; Gulyas, B.; et al. The application of eXplainable artificial intelligence in studying cognition: A scoping review. Ibrain 2024, 10, 245–265. [Google Scholar] [CrossRef]
- Molnar, C. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable. 2025. Available online: https://christophm.github.io/interpretable-ml-book (accessed on 29 May 2025).
- Baniecki, H.; Kretowicz, W.; Piatyszek, P.; Wisniewski, J.; Biecek, P. Dalex: Responsible Machine Learning with Interactive Explainability and Fairness in Python. J. Mach. Learn. Res. 2021, 22, 1–7. [Google Scholar]
- de Lemos, R.; Grześ, M. Self-Adaptive Artificial Intelligence. In Proceedings of the 14th International Symposium on Software Engineering for Adaptive and Self-Managing Systems, Montreal, QC, Canada, 25 May 2019; SEAMS ’19, pp. 155–156. [Google Scholar] [CrossRef]
- Abid, Y.; Wu, J.; Farhan, M.; Ahmad, T. ECMT Framework for Internet of Things: An Integrative Approach Employing In-Memory Attribute Examination and Sophisticated Neural Network Architectures in Conjunction With Hybridized Machine Learning Methodologies. IEEE Internet Things J. 2024, 11, 5867–5886. [Google Scholar] [CrossRef]
- Casado, F.; Lema, D.; Iglesias, R.; Regueiro, C.; Barro, S. Ensemble and Continual Federated Learning for Classification Tasks. Mach. Learn. 2023, 112, 3413–3453. [Google Scholar] [CrossRef]
- Hosseinalipour, S.; Wang, S.; Michelusi, N.; Aggarwal, V.; Brinton, C.G.; Love, D.J.; Chiang, M. Parallel Successive Learning for Dynamic Distributed Model Training Over Heterogeneous Wireless Networks. IEEE/ACM Trans. Netw. 2023, 32, 222–237. [Google Scholar] [CrossRef]
- Lai, T.D. Towards the Generation of Machine Learning Defect Reports. In Proceedings of the 36th IEEE/ACM International Conference on Automated Software Engineering, Melbourne, Australia, 15–19 November 2022; ASE ’21, pp. 1038–1042. [Google Scholar] [CrossRef]
- Kleinert, F.; Leufen, L.; Lupascu, A.; Butler, T.; Schultz, M. Representing Chemical History in Ozone Time-Series Predictions - a Model Experiment Study Building on the MLAir (v1.5) Deep Learning Framework. Geosci. Model Dev. 2022, 15, 8913–8930. [Google Scholar] [CrossRef]
- Ouatiti, Y.; Sayagh, M.; Kerzazi, N.; Adams, B.; Hassan, A. The Impact of Concept Drift and Data Leakage on Log Level Prediction Models. Empir. Softw. Eng. 2024, 29, 123. [Google Scholar] [CrossRef]
- Agarwal, P.; Gao, B.; Huo, S.; Reddy, P.; Dechu, S.; Obeidi, Y.; Muthusamy, V.; Isahagian, V.; Carbajales, S. A Process-Aware Decision Support System for Business Processes. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; KDD ’22, pp. 2673–2681. [Google Scholar] [CrossRef]
- Andresini, G.; Pendlebury, F.; Pierazzi, F.; Loglisci, C.; Appice, A.; Cavallaro, L. INSOMNIA: Towards Concept-Drift Robustness in Network Intrusion Detection. In Proceedings of the 14th ACM Workshop on Artificial Intelligence and Security, Virtual Event, Republic of Korea, 15 November 2021; Association for Computing Machinery, Inc.: New York, NY, USA, 2021; pp. 111–122. [Google Scholar] [CrossRef]
- Cai, H.; Zhou, Z.; Huang, Q. Online Resource Allocation for Edge Intelligence with Colocated Model Retraining and Inference. In Proceedings of the IEEE INFOCOM 2024—IEEE Conference on Computer Communications, Vancouver, BC, Canada, 20–23 May 2024; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2024; pp. 1900–1909. [Google Scholar] [CrossRef]
- Chow, T.; Kan, Z.; Linhardt, L.; Cavallaro, L.; Arp, D.; Pierazzi, F. Drift Forensics of Malware Classifiers. In Proceedings of the 16th ACM Workshop on Artificial Intelligence and Security, Copenhagen, Denmark, 30 November 2023; AISec ’23, pp. 197–207. [Google Scholar] [CrossRef]
- Haug, J.; Broelemann, K.; Kasneci, G. Dynamic Model Tree for Interpretable Data Stream Learning. In Proceedings of the 2022 IEEE 38th International Conference on Data Engineering (ICDE), Kuala Lumpur, Malaysia, 9–12 May 2022; IEEE Computer Society: Washington, DC, USA, 2022; pp. 2562–2574. [Google Scholar] [CrossRef]
- Rymarczyk, D.; Van De Weijer, J.; Zieliński, B.; Twardowski, B. ICICLE: Interpretable Class Incremental Continual Learning. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2023; pp. 1887–1898. [Google Scholar] [CrossRef]
- Siirtola, P.; Röning, J. Feature Relevance Analysis to Explain Concept Drift—A Case Study in Human Activity Recognition. In Proceedings of the Adjunct Proceedings of the 2022 ACM International Joint Conference on Pervasive and Ubiquitous Computing and the 2022 ACM International Symposium on Wearable Computers, Cambridge, UK, 11–15 September 2022; UbiComp/ISWC'22 Adjunct, pp. 386–391. [Google Scholar] [CrossRef]
- Wan, K.; Liang, Y.; Yoon, S. Online Drift Detection with Maximum Concept Discrepancy. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Barcelona, Spain, 25–29 August 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 2924–2935. [Google Scholar] [CrossRef]
- Wang, G.; Guo, H.; Li, A.; Liu, X.; Yan, Q. Federated IoT Interaction Vulnerability Analysis. In Proceedings of the 2023 IEEE 39th International Conference on Data Engineering (ICDE), Anaheim, CA, USA, 3–7 April 2023; IEEE Computer Society: Washington, DC, USA, 2023; pp. 1517–1530. [Google Scholar] [CrossRef]
- Zonoozi, A.; Ho, Q.; Krishnaswamy, S.; Cong, G. ConTrack: A Scalable Method for Tracking Multiple Concepts in Large Scale Multidimensional Data. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining (ICDM), Barcelona, Spain, 12–15 December 2016; Bonchi, F., Domingo-Ferrer, J., Baeza-Yates, R., Zhou, Z.-H., Wu, X., Eds.; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2016; pp. 759–768. [Google Scholar]
- Adams, J.; van Zelst, S.; Quack, L.; Hausmann, K.; van der Aalst, W.; Rose, T. A Framework for Explainable Concept Drift Detection in Process Mining. In Proceedings of the 19th International Conference, BPM 2021, Rome, Italy, 6–10 September 2021; Polyvyanyy, A., Wynn, M.T., Van Looy, A., Reichert, M., Eds.; Springer Science and Business Media Deutschland GmbH: Berlin/Heidelberg, Germany, 2021; Volume 12875 LNCS, pp. 400–416. [Google Scholar] [CrossRef]
- Calikus, E.; Fan, Y.; Nowaczyk, S.; Sant’Anna, A. Interactive-COSMO: Consensus Self-Organized Models for Fault Detection with Expert Feedback. In Proceedings of the Workshop on Interactive Data Mining, Melbourne, VIC, Australia, 15 February 2019; WIDM’19. [Google Scholar] [CrossRef]
- Castelnovo, A.; Malandri, L.; Mercorio, F.; Mezzanzanica, M.; Cosentini, A. Towards Fairness Through Time. In Proceedings of the International Workshops of ECML PKDD 2021, Virtual Event, 13–17 September 2021; Kamp, M., Kamp, M., Koprinska, I., Bibal, A., Bouadi, T., Frénay, B., Galárraga, L., Oramas, J., Adilova, L., Krishnamurthy, Y., et al., Eds.; Springer Science and Business Media Deutschland GmbH: Berlin/Heidelberg, Germany, 2021; Volume 1524 CCIS, pp. 647–663. [Google Scholar] [CrossRef]
- Chen, Q.; Winter, K.; Rinderle-Ma, S. Predicting Unseen Process Behavior Based on Context Information from Compliance Constraints. In Proceedings of the BPM 2023 Forum, Utrecht, The Netherlands, 11–15 September 2023; Di Francescomarino, C., Burattin, A., Janiesch, C., Sadiq, S., Eds.; Springer Science and Business Media Deutschland GmbH: Berlin/Heidelberg, Germany, 2023; Volume 490 LNBIP, pp. 127–144. [Google Scholar] [CrossRef]
- Chen, Z.; Tan, S.; Nori, H.; Inkpen, K.; Lou, Y.; Caruana, R. Using Explainable Boosting Machines (EBMs) to Detect Common Flaws in Data. In Proceedings of the International Workshops of ECML PKDD 2021, Virtual Event, 13–17 September 2021; Kamp, M., Kamp, M., Koprinska, I., Bibal, A., Bouadi, T., Frénay, B., Galárraga, L., Oramas, J., Adilova, L., Krishnamurthy, Y., et al., Eds.; Springer Science and Business Media Deutschland GmbH: Berlin/Heidelberg, Germany, 2021; Volume 1524 CCIS, pp. 534–551. [Google Scholar] [CrossRef]
- Chowdhury, A.; Nguyen, H.; Ashenden, D.; Pogrebna, G. POSTER: A Teacher-Student with Human Feedback Model for Human-AI Collaboration in Cybersecurity. In Proceedings of the 2023 ACM Asia Conference on Computer and Communications Security, Melbourne, VIC, Australia, 10–14 July 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 1040–1042. [Google Scholar] [CrossRef]
- Guo, H.; Jia, F.; Chen, J.; Squicciarini, A.; Yadav, A. RoCourseNet: Robust Training of a Prediction Aware Recourse Model. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, UK, 21–25 October 2023; CIKM ’23, pp. 619–628. [Google Scholar] [CrossRef]
- Guzy, F.; Wozniak, M.; Krawczyk, B. Evaluating and Explaining Generative Adversarial Networks for Continual Learning under Concept Drift. In Proceedings of the 2021 International Conference on Data Mining Workshops (ICDMW), Auckland, New Zealand, 7–10 December 2021; Xue, B., Pechenizkiy, M., Koh, Y.S., Eds.; IEEE Computer Society: Washington, DC, USA, 2021; pp. 295–303. [Google Scholar] [CrossRef]
- Haug, J.; Braun, A.; Zürn, S.; Kasneci, G. Change Detection for Local Explainability in Evolving Data Streams. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; Association for Computing Machinery: New York, NY, USA, 2022; pp. 706–716. [Google Scholar] [CrossRef]
- Jiang, W.; Yi, Z.; Wang, L.; Zhang, H.; Zhang, J.; Lin, F.; Yang, C. A Stochastic Online Forecast-and-Optimize Framework for Real-Time Energy Dispatch in Virtual Power Plants under Uncertainty. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, UK, 21–25 October 2023; CIKM ’23, pp. 4646–4652. [Google Scholar] [CrossRef]
- Muschalik, M.; Fumagalli, F.; Hammer, B.; Hüllermeier, E. iSAGE: An Incremental Version of SAGE for Online Explanation on Data Streams. In Proceedings of the European Conference, ECML PKDD 2023, Turin, Italy, 18–22 September 2023; Koutra, D., Plant, C., Gomez Rodriguez, M., Baralis, E., Bonchi, F., Eds.; Springer Science and Business Media Deutschland GmbH: Berlin/Heidelberg, Germany, 2023; Volume 14171 LNAI, pp. 428–445. [Google Scholar] [CrossRef]
- Nadeem, A.; Verwer, S. SECLEDS: Sequence Clustering in Evolving Data Streams via Multiple Medoids and Medoid Voting. In Proceedings of the European Conference, ECML PKDD 2022, Grenoble, France, 19–23 September 2022; Amini, M.-R., Canu, S., Fischer, A., Guns, T., Kralj Novak, P., Tsoumakas, G., Eds.; Springer Science and Business Media Deutschland GmbH: Berlin/Heidelberg, Germany, 2023; Volume 13713 LNAI, pp. 157–173. [Google Scholar] [CrossRef]
- Polyvyanyy, A.; Wynn, M.T.; Looy, A.V.; Reichert, M. (Eds.) Business Process Management: 19th International Conference on Business Process Management, BPM, Rome, Italy, 6–10 September 2021; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2021; Volume 12875 LNCS. [Google Scholar]
- Yang, K.; Xu, W. Fraudmemory: Explainable Memory-Enhanced Sequential Neural Networks for Financial Fraud Detection. In Proceedings of the 52nd Hawaii International Conference on System Sciences, Maui, HI, USA, 8–11 January 2019; Bui, T.X., Ed.; IEEE Computer Society: Washington, DC, USA, 2019; pp. 1023–1032. [Google Scholar]
- Wang, G. Open Challenges of Malware Detection under Concept Drift. In Proceedings of the 1st Workshop on Robust Malware Analysis, Nagasaki, Japan, 30 May 2022; WoRMA ’22, p. 1. [Google Scholar] [CrossRef]
- Casalino, G.; Castellano, G.; Kaczmarek-Majer, K.; Leite, D. (Eds.) OLUD 2022—Proceedings of the 1st Workshop on Online Learning from Uncertain Data Streams, co-located with the IEEE World Congress on Computational Intelligence (WCCI 2022). In CEUR Workshop Proceedings; CEUR-WS: Padova, Italy, 2022; Volume 3380. [Google Scholar]
- Almeida, J.P.A.; Ciccio, C.D.; Kalloniatis, C. (Eds.) Advanced Information Systems Engineering Workshops: CAiSE 2024 International Workshops, Limassol, Cyprus, June 3–7, 2024, Proceedings; Lecture Notes in Business Information Processing; Springer: Cham, Switzerland, 2024; Volume 521. [Google Scholar]
- Canovas-Segura, B.; Morales, A.; Martinez-Carrasco, A.; Campos, M.; Juarez, J.; Rodriguez, L.; Palacios, F. Improving Interpretable Prediction Models for Antimicrobial Resistance. In Proceedings of the 2019 IEEE 32nd International Symposium on Computer-Based Medical Systems (CBMS), Cordoba, Spain, 5–7 June 2019; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2019; pp. 543–546. [Google Scholar] [CrossRef]
- Cianci, G.; Goglia, R.; Guidotti, R.; Kapllaj, M.; Mosca, R.; Pugnana, A.; Ricotti, F.; Ruggieri, S. Applied Data Science for Leasing Score Prediction. In Proceedings of the 2023 IEEE International Conference on Big Data (BigData), Sorrento, Italy, 15–18 December 2023; He, J., Palpanas, T., Hu, X., Cuzzocrea, A., Dou, D., Slezak, D., Wang, W., Gruca, A., Lin, J.C.-W., Agrawal, R., Eds.; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2023; pp. 1687–1696. [Google Scholar] [CrossRef]
- Clement, T.; Nguyen, H.; Kemmerzell, N.; Abdelaal, M.; Stjelja, D. Coping with Data Distribution Shifts: XAI-Based Adaptive Learning with SHAP Clustering for Energy Consumption Prediction. In Proceedings of the 36th Australasian Joint Conference on Artificial Intelligence, AI 2023, Brisbane, QLD, Australia, 28 November–1 December 2023; Liu, T., Webb, G., Yue, L., Wang, D., Eds.; Springer Science and Business Media Deutschland GmbH: Singapore, 2024; Volume 14472 LNAI, pp. 147–159. [Google Scholar] [CrossRef]
- Gniewkowski, M.; Maciejewski, H.; Surmacz, T.; Walentynowicz, W. Sec2vec: Anomaly Detection in HTTP Traffic and Malicious URLs. In Proceedings of the 38th ACM/SIGAPP Symposium on Applied Computing, Tallinn, Estonia, 27–31 March 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 1154–1162. [Google Scholar] [CrossRef]
- Kebir, S.; Tabia, K. On Handling Concept Drift, Calibration and Explainability in Non-Stationary Environments and Resources Limited Contexts. In Proceedings of the 16th International Conference on Agents and Artificial Intelligence—Volume 2: ICAART, Rome, Italy, 24–26 February 2024; Rocha, A.P., Steels, L., van den Herik, J., Eds.; Science and Technology Publications, Lda: Setúbal, Portugal, 2024; Volume 2, pp. 336–346. [Google Scholar] [CrossRef]
- Migenda, N.; Schenck, W. Adaptive Dimensionality Reduction for Local Principal Component Analysis. In Proceedings of the 2020 25th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Vienna, Austria, 8–11 September 2020; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2020; pp. 1579–1586. [Google Scholar] [CrossRef]
- Neufeld, D.; Schmid, U. Anomaly Detection for Hydraulic Systems under Test. In Proceedings of the 2021 26th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA ), Vasteras, Sweden, 7–10 September 2021; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2021. [Google Scholar] [CrossRef]
- Saralajew, S.; Villmann, T. Transfer Learning in Classification Based on Manifolc. Models and Its Relation to Tangent Metric Learning. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2017; pp. 1756–1765. [Google Scholar] [CrossRef]
- Wang, X.; Wang, Z.; Shao, W.; Jia, C.; Li, X. Explaining Concept Drift of Deep Learning Models. In Proceedings of the 11th International Symposium, CSS 2019, Guangzhou, China, 1–3 December 2019; Zhang, X., Li, J., Eds.; Springer: Cham, Switzerland, 2019; Volume 11983 LNCS, pp. 524–534. [Google Scholar] [CrossRef]
- Appice, A.; Tsoumakas, G.; Manolopoulos, Y.; Matwin, S. (Eds.) Discovery Science: 23rd International Conference on Discovery Science, DS 2020; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2020; Volume 12323 LNAI. [Google Scholar]
- Bouadi, T.; Fromont, E.; Hüllermeier, E. (Eds.) Advances in Intelligent Data Analysis XX: 20th International Symposium on Intelligent Data Analysis, IDA 2022; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2022; Volume 13205 LNCS. [Google Scholar]
- Miliou, I.; Piatkowski, N.; Papapetrou, P. (Eds.) Advances in Intelligent Data Analysis XXII: 22nd International Symposium on Intelligent Data Analysis, IDA 2024; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2024; Volume 14641 LNCS. [Google Scholar]
- Rutkowski, L.; Scherer, R.; Korytkowski, M.; Pedrycz, W.; Tadeusiewicz, R.; Zurada, J.M. (Eds.) Artificial Intelligence and Soft C: 19th International Conference on Artificial Intelligence and Soft Computing, ICAISC 2020; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2020; Volume 12415 LNAI. [Google Scholar]
- Wang, Z.; Tan, C. Trends and Applications in Knowledge Discovery and Data Mining: PAKDD 2024 Workshops, RAFDA and IWTA, Taipei, Taiwan, May 7–10, 2024, Proceedings (Lecture Notes in Artificial Intelligence, Band 14658); Springer: Berlin/Heidelberg, Germany, 2024. [Google Scholar]
- Appice, A.; Azzag, H.; Hacid, M.S.; Hadjali, A.; Ras, Z. Foundations of Intelligent Systems: 27th International Symposium, ISMIS 2024, Poitiers, France, June 17–19, 2024, Proceedings (Lecture Notes in Computer Science Book 14670); Springer: Berlin/Heidelberg, Germany, 2024. [Google Scholar]
- Baillargeon, J.T.; Cossette, H.; Lamontagne, L. Preventing RNN from Using Sequence Length as a Feature. In Proceedings of the 2022 6th International Conference on Natural Language Processing and Information Retrieval, Bangkok, Thailand, 16–18 December 2022; Association for Computing Machinery: New York, NY, USA, 2022; pp. 16–20. [Google Scholar] [CrossRef]
- Banissi, E.; Datia, N.; Pires, J.; Ursyn, A.; Nazemi, K.; Kovalerchuk, B.; Andonie, R.; Gavrilova, M.; Nakayama, M.; Nguyen, Q.; et al. Proceedings—2024 28th International Conference Information Visualisation (IV 2024); IEEE: Piscataway, NJ, USA, 2024. [Google Scholar]
- Conte, D.; Fred, A.; Gusikhin, O.; Sansone, C. (Eds.) Deep Learning Theory and Applications: Proceedings of the 4th International Conference on Deep Learning Theory and Applications, DeLTA 2023; Communications in Computer and Information Science; Springer: Cham, Switzerland, 2023; Volume 1875 CCIS. [Google Scholar]
- Ding, G.; Zhu, Y.; Ren, Y. Dynamic-Static Fusion for Spatial-Temporal Anomaly Detection and Interpretation in Multivariate Time Series. In Proceedings of the 8th International Joint Conference, APWeb-WAIM 2024, Jinhua, China, 30 August–1 September 2024; Yang, Z., Wang, X., Tung, A., Zheng, Z., Guo, H., Eds.; Springer Science and Business Media Deutschland GmbH: Singapore, 2024; Volume 14963 LNCS, pp. 46–61. [Google Scholar] [CrossRef]
- Dries, A. Declarative Data Generation with ProbLog. In Proceedings of the 6th International Symposium on Information and Communication Technology, Hue City, Viet Nam, 3–4 December 2015; SoICT ’15, pp. 17–24. [Google Scholar] [CrossRef]
- Esteve, M.; Mollá-Campello, N.; Rodríguez-Sala, J.; Rabasa, A. The Effects of Abrupt Changing Data in CART Inference Models. In Trends and Applications in Information Systems and Technologies; Rocha, A., Adeli, H., Dzemyda, G., Moreira, F., Correia, A.M.R., Eds.; Springer Science and Business Media Deutschland GmbH: Cham, Switzerland, 2021; Volume 1366 AISC, pp. 214–223. [Google Scholar] [CrossRef]
- Bringas, P.G.; García, H.P.; de Pisón, F.J.M.; Flecha, J.R.V.; Lora, A.T.; la Cal, E.A.d.; Herrero, Á.; Álvarez, F.M.; Psaila, G.; Quintián, H.; et al. (Eds.) Hybrid Artificial Intelligent Systems: 17th International Conference on Hybrid Artificial Intelligence Systems, HAIS 2022; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2022; Volume 13469 LNAI. [Google Scholar]
- Jain, M.; Gupta, S.; Agarwal, G.; Bansal, K. AdaBoost Based Hybrid Concept Drift Detection Approach for Mental Health Prediction. In Proceedings of the 2023 Fifteenth International Conference on Contemporary Computing, Noida, India, 3–5 August 2023; IC3-2023, pp. 345–351. [Google Scholar] [CrossRef]
- Le, T.D.; Ong, K.-L.; Zhao, Y.; Jin, W.H.; Wong, S.; Liu, L.; Williams, G. (Eds.) Data Mining: 17th Australasian Conference on Data Mining, AusDM 2019; Communications in Computer and Information Science; Springer: Singapore, 2019; Volume 1127 CCIS. [Google Scholar]
- Maglogiannis, I.; Iliadis, L.; Macintyre, J.; Cortez, P. Artificial Intelligence Applications and Innovations, Proceedings of the 18th IFIP WG 12.5 International Conference, AIAI 2022, Hersonissos, Crete, Greece, 17–20 June 2022; Springer: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
- Nader, C.; Bou-Harb, E. Revisiting IoT Fingerprinting behind a NAT. In Proceedings of the 2021 IEEE International Conference on Parallel & Distributed Processing with Applications, Big Data & Cloud Computing, Sustainable Computing & Communications, Social Computing & Networking (ISPA/BDCloud/SocialCom/SustainCom), New York, NY, USA, 30 September–3 October 2021; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2021; pp. 1745–1752. [Google Scholar] [CrossRef]
- Nesic, S.; Putina, A.; Bahri, M.; Huet, A.; Navarro, J.; Rossi, D.; Sozio, M. STREamRHF: Tree-Based Unsupervised Anomaly Detection for Data Streams. In Proceedings of the 2022 IEEE/ACS 19th International Conference on Computer Systems and Applications (AICCSA), Abu Dhabi, United Arab Emirates, 5–8 December 2022; IEEE Computer Society: Washington, DC, USA, 2022. [Google Scholar] [CrossRef]
- Park, L.A.F.; Gomes, H.M.; Doborjeh, M.; Boo, Y.L.; Koh, Y.S.; Zhao, Y.; Williams, G.; Simoff, S. (Eds.) Data Mining: 20th Australasian Data Mining Conference, AusDM 2022; Communications in Computer and Information Science; Springer: Singapore, 2022; Volume 1741 CCIS. [Google Scholar]
- Tan, G.; Chen, P.; Li, M. Online Data Drift Detection for Anomaly Detection Services Based on Deep Learning towards Multivariate Time Series. In Proceedings of the 2023 IEEE 23rd International Conference on Software Quality, Reliability, and Security (QRS), Chiang Mai, Thailand, 22–26 October 2023; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2023; pp. 1–11. [Google Scholar] [CrossRef]
- Shiaeles, S.; Kolokotronis, N.; Bellini, E. Proceedings of the 2024 IEEE International Conference on Cyber Security and Resilience (CSR); IEEE: Piscataway, NJ, USA, 2024. [Google Scholar]
- Ajitha, P. MedAI-DevOps Imaging Suite: Integrating CNN in Diagnostic Imaging with Continuous Deployment and Real-time Monitoring. In Proceedings of the 2024 International Conference on Trends in Quantum Computing and Emerging Business Technologies, Pune, India, 22–23 March 2024; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2024. [Google Scholar] [CrossRef]
- De Arriba-Pérez, F.; García-Méndez, S.; Leal, F.; Malheiro, B.; Burguillo, J.C. Online Detection and Infographic Explanation of Spam Reviews with Data Drift Adaptation. Informatica 2024, 35, 483–507. [Google Scholar] [CrossRef]
- Kulczycki, P.; Kowalski, P. A Metaheuristic for Classification of Interval Data in Changing Environments. In Information Technology and Computational Physics; Koczy, L.T., Kacprzyk, J., Kulczycki, P., Kulczycki, P., Mesiar, R., Koczy, L.T., Eds.; Springer: Cham, Switzerland, 2017; Volume 462, pp. 19–34. [Google Scholar] [CrossRef]
- Liu, S.; Bronzino, F.; Schmitt, P.; Bhagoji, A.N.; Feamster, N.; Crespo, H.G.; Coyle, T.; Ward, B. LEAF: Navigating Concept Drift in Cellular Networks. Proc. ACM Netw. 2023, 1, 7. [Google Scholar] [CrossRef]
- Palaniappan, S.; Veneri, G.; Gori, V.; Pratelli, T.; Ballarini, V. Anomaly Detection of Sensor Measurements During a Turbo-Machine Prototype Testing—An Integrated ML Ops, Continual Learning Architecture. In Proceedings of the International Petroleum Technology Conference, Dhahran, Saudi Arabia, February 2024; International Petroleum Technology Conference (IPTC): London, UK, 2024. [Google Scholar] [CrossRef]
- Wang, R. Enhancing Thyroid Disease Prediction: Integrating Machine Learning Models for Improved Diagnostic Accuracy and Interpretability. In Proceedings of the 2024 5th International Conference on Artificial Intelligence and Electromechanical Automation (AIEA), Shenzhen, China, 14–16 June 2024; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2024; pp. 423–431. [Google Scholar] [CrossRef]
- Institute of Electrical and Electronics Engineers Inc. Proceedings of 2022 International Conference on Future Trends in Smart Communities (ICFTSC); IEEE: Piscataway, NJ, USA, 2022. [Google Scholar]
- Adams, J.; van Zelst, S.; Rose, T.; van der Aalst, W. Explainable Concept Drift in Process Mining. Inf. Syst. 2023, 114, 102177. [Google Scholar] [CrossRef]
- Bertini, J. Graph Embedded Rules for Explainable Predictions in Data Streams. Neural Netw. 2020, 129, 174–192. [Google Scholar] [CrossRef]
- Bhaskhar, N.; Rubin, D.; Lee-Messer, C. An Explainable and Actionable Mistrust Scoring Framework for Model Monitoring. IEEE Trans. Artif. Intell. 2024, 5, 1473–1485. [Google Scholar] [CrossRef]
- Bijlani, N.; Maldonado, O.; Nilforooshan, R.; Barnaghi, P.; Kouchaki, S. Utilizing Graph Neural Networks for Adverse Health Detection and Personalized Decision Making in Sensor-Based Remote Monitoring for Dementia Care. Comput. Biol. Med. 2024, 183, 109287. [Google Scholar] [CrossRef]
- Bosnić, Z.; Demšar, J.; Kešpret, G.; Pereira Rodrigues, P.; Gama, J.; Kononenko, I. Enhancing Data Stream Predictions with Reliability Estimators and Explanation. Eng. Appl. Artif. Intell. 2014, 34, 178–192. [Google Scholar] [CrossRef]
- Cano, A.; Krawczyk, B. Evolving Rule-Based Classifiers with Genetic Programming on GPUs for Drifting Data Streams. Pattern Recognit. 2019, 87, 248–268. [Google Scholar] [CrossRef]
- Choi, Y.; Yu, W.X.; Nagarajan, M.B.; Teng, P.Y.; Goldin, J.G.; Raman, S.S.; Enzmann, D.R.; Kim, G.H.J.; Brown, M.S. Translating AI to Clinical Practice: Overcoming Data Shift with Explainability. Radiographics 2023, 43, e220105. [Google Scholar] [CrossRef]
- Costa, R.; Hirata, C.; Pugliese, V. A Comparative Study of Situation Awareness-Based Decision-Making Model Reinforcement Learning Adaptive Automation in Evolving Conditions. IEEE Access 2023, 11, 16166–16182. [Google Scholar] [CrossRef]
- Das, R.; Ang, K.; Quek, C. IeRSPOP: A Novel Incremental Rough Set-Based Pseudo Outer-Product with Ensemble Learning. Appl. Soft Comput. J. 2016, 46, 170–186. [Google Scholar] [CrossRef]
- Das, S.; Islam, M.R.; Kannappan Jayakodi, N.; Doppa, J.R. Effectiveness of Tree-based Ensembles for Anomaly Discovery: Insights, Batch and Streaming Active Learning. J. Artif. Int. Res. 2024, 80, 127–170. [Google Scholar] [CrossRef]
- Demšar, J.; Bosnić, Z. Detecting Concept Drift in Data Streams Using Model Explanation. Expert Syst. Appl. 2018, 92, 546–559. [Google Scholar] [CrossRef]
- Duckworth, C.; Chmiel, F.; Burns, D.; Zlatev, Z.; White, N.; Daniels, T.; Kiuber, M.; Boniface, M. Using Explainable Machine Learning to Characterise Data Drift and Detect Emergent Health Risks for Emergency Department Admissions during COVID-19. Sci. Rep. 2021, 11, 23017. [Google Scholar] [CrossRef]
- Ferdaus, M.; Dam, T.; Alam, S.; Pham, D.T. X-Fuzz: An Evolving and Interpretable Neuro-Fuzzy Learner for Data Streams. IEEE Trans. Artif. Intell. 2024, 5, 4001–4012. [Google Scholar] [CrossRef]
- Freitas dos Santos, T.; Osman, N.; Schorlemmer, M. Is This a Violation? Learning and Understanding Norm Violations in Online Communities. Artif. Intell. 2024, 327, 104058. [Google Scholar] [CrossRef]
- Fumagalli, F.; Muschalik, M.; Hüllermeier, E.; Hammer, B. Incremental Permutation Feature Importance (iPFI): Towards Online Explanations on Data Streams. Mach. Learn. 2023, 112, 4863–4903. [Google Scholar] [CrossRef]
- González–Cebrián, A.; Bradford, M.; Chis, A.; González–Vélez, H. Standardised Versioning of Datasets: A FAIR–Compliant Proposal. Sci. Data 2024, 11, 358. [Google Scholar] [CrossRef] [PubMed]
- He, G.; Jin, D.; Dai, L.; Xin, X.; Yu, Z.; Chen, C. Online Learning of Temporal Association Rule on Dynamic Multivariate Time Series Data. IEEE Trans. Knowl. Data Eng. 2024, 36, 8954–8966. [Google Scholar] [CrossRef]
- He, X.; Liu, Z. Dynamic Model Interpretation-Guided Online Active Learning Scheme for Real-Time Safety Assessment. IEEE Trans. Cybern. 2024, 54, 2734–2745. [Google Scholar] [CrossRef]
- Jensen, R.; Ferwerda, J.; Jørgensen, K.; Jensen, E.; Borg, M.; Krogh, M.; Jensen, J.; Iosifidis, A. A Synthetic Data Set to Benchmark Anti-Money Laundering Methods. Sci. Data 2023, 10, 661. [Google Scholar] [CrossRef]
- Korine, R.; Hendler, D. DAEMON: Dataset/Platform-Agnostic Explainable Malware Classification Using Multi-Stage Feature Mining. IEEE Access 2021, 9, 78382–78399. [Google Scholar] [CrossRef]
- Li, X.; Liu, Z.; Yang, Z.; Meng, F.; Song, T. A High-Precision Interpretable Framework for Marine Dissolved Oxygen Concentration Inversion. Front. Mar. Sci. 2024, 11, 1396277. [Google Scholar] [CrossRef]
- Roider, J.; Nguyen, A.; Zanca, D.; Eskofier, B. Assessing the Performance of Remaining Time Prediction Methods for Business Processes. IEEE Access 2024, 12, 130583–130601. [Google Scholar] [CrossRef]
- Samarajeewa, C.; De Silva, D.; Manic, M.; Mills, N.; Moraliyage, H.; Alahakoon, D.; Jennings, A. An Artificial Intelligence Framework for Explainable Drift Detection in Energy Forecasting. Energy AI 2024, 17, 100403. [Google Scholar] [CrossRef]
- Singh, A.; Mishra, P.; Vinod, P.; Gaur, A.; Conti, M. SFC-NIDS: A Sustainable and Explainable Flow Filtering Based Concept Drift-Driven Security Approach for Network Introspection. Clust. Comput. 2024, 27, 9757–9782. [Google Scholar] [CrossRef]
- Tiukhova, E.; Vemuri, P.; Flores, N.; Islind, A.; Óskarsdóttir, M.; Poelmans, S.; Baesens, B.; Snoeck, M. Explainable Learning Analytics: Assessing the Stability of Student Success Prediction Models by Means of Explainable AI. Decis. Support Syst. 2024, 182, 114229. [Google Scholar] [CrossRef]
- Vieira, D.; Fernandes, C.; Lucena, C.; Lifschitz, S. Driftage: A Multi-Agent System Framework for Concept Drift Detection. GigaScience 2021, 10, giab030. [Google Scholar] [CrossRef]
- Wen, C.; Zhou, P.; Dai, W.; Dong, L.; Chai, T. Online Sequential Sparse Robust Neural Networks with Random Weights for Imperfect Industrial Streaming Data Modeling. IEEE Trans. Autom. Sci. Eng. 2024, 21, 1163–1175. [Google Scholar] [CrossRef]
- Wood, M.; Ogliari, E.; Nespoli, A.; Simpkins, T.; Leva, S. Day Ahead Electric Load Forecast: A Comprehensive LSTM-EMD Methodology and Several Diverse Case Studies. Forecasting 2023, 5, 297–314. [Google Scholar] [CrossRef]
- Xu, L.; Han, Z.; Zhao, D.; Li, X.; Yu, F.; Chen, C. Addressing Concept Drift in IoT Anomaly Detection: Drift Detection, Interpretation, and Adaptation. IEEE Trans. Sustain. Comput. 2024, 9, 913–924. [Google Scholar] [CrossRef]
- Xuan, J.; Lu, J.; Zhang, G. Bayesian Nonparametric Unsupervised Concept Drift Detection for Data Stream Mining. ACM Trans. Intell. Syst. Technol. 2020, 12, 5. [Google Scholar] [CrossRef]
- Yu, E.; Lu, J.; Zhang, G. Fuzzy Shared Representation Learning for Multistream Classification. IEEE Trans. Fuzzy Syst. 2024, 32, 5625–5637. [Google Scholar] [CrossRef]
- Zhu, J.; Cai, S.; Deng, F.; Ooi, B.C.; Zhang, W. METER: A Dynamic Concept Adaptation Framework for Online Anomaly Detection. Proc. VLDB Endow. 2024, 17, 794–807. [Google Scholar] [CrossRef]
- Ducange, P.; Marcelloni, F.; Pecori, R. Fuzzy Hoeffding Decision Tree for Data Stream Classification. Int. J. Comput. Intell. Syst. 2021, 14, 946–964. [Google Scholar] [CrossRef]
- Farrugia, D.; Zerafa, C.; Cini, T.; Kuasney, B.; Livori, K. A Real-Time Prescriptive Solution for Explainable Cyber-Fraud Detection Within the iGaming Industry. SN Comput. Sci. 2021, 2, 215. [Google Scholar] [CrossRef]
- Freitas dos Santos, T.; Osman, N.; Schorlemmer, M. A Multi-Scenario Approach to Continuously Learn and Understand Norm Violations. Auton. Agents -Multi-Agent Syst. 2023, 37, 38. [Google Scholar] [CrossRef]
- Goyal, S.; Kumar, S.; Singh, S.; Sarin, S.; Gupta, B.; Arya, V.; Alhalabi, W.; Colace, F. Synergistic Application of Neuro-Fuzzy Mechanisms in Advanced Neural Networks for Real-Time Stream Data Flux Mitigation. Soft Comput. 2024, 28, 12425–12437. [Google Scholar] [CrossRef]
- Li, B.; Gupta, S.; Müller, E. State-Transition-Aware Anomaly Detection under Concept Drifts. Data Knowl. Eng. 2024, 154, 102365. [Google Scholar] [CrossRef]
- Monakhov, V.; Thambawita, V.; Halvorsen, P.; Riegler, M. GridHTM: Grid-Based Hierarchical Temporal Memory for Anomaly Detection in Videos. Sensors 2023, 23, 2087. [Google Scholar] [CrossRef] [PubMed]
- Nesvijevskaia, A.; Ouillade, S.; Guilmin, P.; Zucker, J.D. The Accuracy versus Interpretability Trade-off in Fraud Detection Model. Data Policy 2021, 3, e12. [Google Scholar] [CrossRef]
- Pasquadibisceglie, V.; Appice, A.; Ieva, G.; Malerba, D. TSUNAMI - an Explainable PPM Approach for Customer Churn Prediction in Evolving Retail Data Environments. J. Intell. Inf. Syst. 2024, 62, 705–733. [Google Scholar] [CrossRef]
- Stahl, F.; Le, T.; Badii, A.; Gaber, M. A Frequent Pattern Conjunction Heuristic for Rule Generation in Data Streams. Information 2021, 12, 24. [Google Scholar] [CrossRef]
- Sun, L.; Li, Z.; Li, K.; Liu, H.; Liu, G.; Lv, W. Cross-Well Lithology Identification Based on Wavelet Transform and Adversarial Learning. Energies 2023, 16, 1475. [Google Scholar] [CrossRef]
- Zheng, Z.; Dong, X.; Yao, J.; Zhou, L.; Ding, Y.; Chen, A. Identification of Epileptic EEG Signals Through TSK Transfer Learning Fuzzy System. Front. Neurosci. 2021, 15, 738268. [Google Scholar] [CrossRef]
- Cherkashina, T.Y. Russian regional politicians’ income and property declarations: A pilot study and quality assessment of administrative data. Sotsiologicheskiy Zhurnal 2021, 27, 8–34. [Google Scholar] [CrossRef]
- Zhang, B.; Wen, Z.; Wei, X.; Ren, J. InterDroid: An Interpretable Android Malware Detection Method for Conceptual Drift. Jisuanji Yanjiu Fazhan/Comput. Res. Dev. 2021, 58, 2456–2474. [Google Scholar] [CrossRef]
- Banf, M.; Steinhagen, G. Supervising The Supervisor—Model Monitoring In Production Using Deep Feature Embeddings with Applications To Workpiece Inspection. In Proceedings of the 8th World Congress on Electrical Engineering and Computer Systems and Sciences (EECSS’22), Prague, Czech Republic, 28–30 July 2022; Benedicenti, L., Liu, Z., Eds.; Avestia Publishing: Orléans, ON, Canada, 2022. [Google Scholar] [CrossRef]
- Hu, H.; Kantardzic, M.; Kar, S. Explainable Data Stream Mining: Why the New Models Are Better. Intell. Decis. Technol. 2024, 18, 371–385. [Google Scholar] [CrossRef]
- Mannmeusel, J.; Rothfelder, M.; Khoshrou, S. Speeding Things Up. Can Explainability Improve Human Learning? In Proceedings of the First World Conference, xAI 2023, Lisbon, Portugal, 26–28 July 2023; Longo, L., Ed.; Springer Science and Business Media Deutschland GmbH: Cham, Switzerland, 2023; Volume 1901 CCIS, pp. 66–84. [Google Scholar] [CrossRef]
- Mascha, P. A Comparison of Model Confidence Metrics on Visual Manufacturing Quality Data. In Proceedings of the CVMI 2022, Allahabad, India, 12–13 August 2022; Tistarelli, M., Dubey, S.R., Singh, S.K., Jiang, X., Eds.; Springer Science and Business Media Deutschland GmbH: Singapore, 2023; Volume 586 LNNS, pp. 165–177. [Google Scholar] [CrossRef]
- Muschalik, M.; Fumagalli, F.; Jagtani, R.; Hammer, B.; Hüllermeier, E. iPDP: On Partial Dependence Plots in Dynamic Modeling Scenarios. In Proceedings of the First World Conference, xAI 2023, Lisbon, Portugal, 26–28 July 2023; Longo, L., Ed.; Springer Science and Business Media Deutschland GmbH: Cham, Switzerland, 2023; Volume 1901 CCIS, pp. 177–194. [Google Scholar] [CrossRef]
- Saadallah, A. Online Explainable Ensemble of Tree Models Pruning for Time Series Forecasting. In Joint Proceedings of the xAI 2024 Late-Breaking Work, Demos and Doctoral Consortium Co-Located with the 2nd World Conference on eXplainable Artificial Intelligence (xAI 2024), Valletta, Malta, 17–19 July 2024. [Google Scholar]
- Suffian, M.; Bogliolo, A. Investigation and Mitigation of Bias in Explainable AI. In Proceedings of the 1st Workshop on Bias, Ethical AI, Explainability and the Role of Logic and Logic Programming (BEWARE 2022) Co-Located with the 21th International Conference of the Italian Association for Artificial Intelligence (AI*IA 2022), Udine, Italy, 2 December 2022. [Google Scholar]
- Sun, J.; Sun, M.; Zhao, M.; Du, Y. Dynamic Class-Imbalanced Financial Distress Prediction Based on Case-Based Reasoning Integrated with Time Weighting and Resampling. J. Credit. Risk 2023, 19, 39–73. [Google Scholar] [CrossRef]
- Hunter, J.D. Matplotlib: A 2D graphics environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
- Eck, N.V.; Waltman, L. Software survey: VOSviewer, a computer program for bibliometric mapping. Scientometrics. Scientometrics 2010, 84, 523–538. [Google Scholar]
Age | Sex | Job | Housing | Savings | Amount | Dur. | Purpose |
---|---|---|---|---|---|---|---|
58 | f | unskilled | free | little | 6143 | 48 | car |
Age | Sex | Job | Amount | Dur. | ||||
---|---|---|---|---|---|---|---|---|
skilled | −20 | 0.108 | 2 | 0.036 | 0.501 | |||
skilled | −24 | 0.114 | 2 | 0.029 | 0.525 | |||
skilled | −22 | 0.111 | 2 | 0.033 | 0.513 | |||
−6 | skilled | −24 | 0.126 | 3 | 0.018 | 0.505 | ||
−3 | skilled | −24 | 0.120 | 3 | 0.024 | 0.515 | ||
−1 | skilled | −24 | 0.116 | 3 | 0.027 | 0.522 | ||
−3 | m | −24 | 0.195 | 3 | 0.012 | 0.501 | ||
−6 | m | −25 | 0.202 | 3 | 0.011 | 0.501 | ||
−30 | m | skilled | −24 | 0.285 | 4 | 0.005 | 0.590 | |
−4 | m | −1254 | −24 | 0.204 | 4 | 0.002 | 0.506 |
Rank | Paper |
---|---|
[23,27,30,31,32,33,34,35,36,37,38,39] | |
A | [40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56] |
B | [57,58,59,60,61,62,63,64,65,66,67,68,69,70] |
C | [71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86] |
NotFound | [87,88,89,90,91,92,93] |
Quartile | Paper |
---|---|
[22,24,25,26,28,29,94,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116,117,118,119,120,121,122,123,124,125] | |
[126,127,128,129,130,131,132,133,134,135,136] | |
[137,138] | |
[139,140,141,142,143,144,145,146] |
Application Domain | N. of Paper |
---|---|
Healthcare | 12 |
Cybersecurity | 8 |
Financial | 6 |
Energy | 6 |
Process Mining | 5 |
Manufacturing | 5 |
Generic Data Stream | 5 |
Data Mining | 5 |
Retail | 2 |
Online Community | 2 |
Video Surveillance | 2 |
Other | 22 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Pelosi, D.; Cacciagrano, D.; Piangerelli, M. Explainability and Interpretability in Concept and Data Drift: A Systematic Literature Review. Algorithms 2025, 18, 443. https://doi.org/10.3390/a18070443
Pelosi D, Cacciagrano D, Piangerelli M. Explainability and Interpretability in Concept and Data Drift: A Systematic Literature Review. Algorithms. 2025; 18(7):443. https://doi.org/10.3390/a18070443
Chicago/Turabian StylePelosi, Daniele, Diletta Cacciagrano, and Marco Piangerelli. 2025. "Explainability and Interpretability in Concept and Data Drift: A Systematic Literature Review" Algorithms 18, no. 7: 443. https://doi.org/10.3390/a18070443
APA StylePelosi, D., Cacciagrano, D., & Piangerelli, M. (2025). Explainability and Interpretability in Concept and Data Drift: A Systematic Literature Review. Algorithms, 18(7), 443. https://doi.org/10.3390/a18070443