Next Article in Journal
Ultrasound-Assisted Urea-Water Solution (AdBlue) Droplets Vaporization: A Mathematical Model for Film and Volumetric Regimes with Implications in NOx Emission Control
Previous Article in Journal
Correction: Hameed et al. Green Synthesis of Zinc Oxide (ZnO) Nanoparticles from Green Algae and Their Assessment in Various Biological Applications. Micromachines 2023, 14, 928
Previous Article in Special Issue
An Artificial Intelligence QRS Detection Algorithm for Wearable Electrocardiogram Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Correction and Long-Term Drift Compensation in MOS Gas Sensor Arrays Using Iterative Random Forests and Incremental Domain-Adversarial Networks

1
School of Big Data and Basic Sciences, Shandong Institute of Petroleum and Chemical Technology, Dongying 257000, China
2
School of Natural Resources and Surveying, Nanning Normal University, Nanning 530001, China
*
Author to whom correspondence should be addressed.
Micromachines 2025, 16(9), 991; https://doi.org/10.3390/mi16090991 (registering DOI)
Submission received: 1 August 2025 / Revised: 26 August 2025 / Accepted: 27 August 2025 / Published: 29 August 2025
(This article belongs to the Special Issue AI-Driven Design and Optimization of Microsystems)

Abstract

Sensor arrays serve a crucial role in various fields such as environmental monitoring, industrial process control, and medical diagnostics, yet their reliability remains challenged by sensor drift and noise contamination. This study presents a novel framework for real-time data error correction and long-term drift compensation utilizing an iterative random forest-based error correction algorithm paired with an Incremental Domain-Adversarial Network (IDAN). The iterative random forest algorithm leverages the collective data from multiple sensor channels to identify and rectify abnormal sensor responses in real time. The IDAN integrates domain-adversarial learning principles with an incremental adaptation mechanism to effectively manage temporal variations in sensor data. Experiments utilizing the metal oxide semiconductor gas sensor array drift dataset demonstrate that the combination of these approaches significantly enhances data integrity and operational efficiency, achieving a robust and good accuracy even in the presence of severe drift. This study underscores the efficacy of integrating advanced artificial intelligence techniques for the ongoing evolution of sensor array technology, paving the way for enhanced monitoring systems capable of sustaining high levels of performance over extended time periods.

1. Introduction

Sensor arrays have become indispensable in a variety of critical sectors, including environmental monitoring, industrial process control, medical diagnostics, and food safety assurance [1]. These systems offer the capability to acquire multi-dimensional and high-throughput data, enabling real-time detection and quantitative analysis of physical, chemical, or biological parameters in complex environments [2,3]. Despite their widespread adoption and promising potential, the reliability and long-term stability of sensor arrays remain a significant challenge in practical applications. One of the most pervasive issues encountered in sensor array deployments is sensor drift, a phenomenon characterized by the gradual, systematic deviation of sensor responses from their original calibrated baseline over time [4,5]. Sensor drift can result from various factors, including sensor aging, material degradation, fouling by sample matrices, changes in environmental conditions, electronic component instability, and even manufacturing inconsistencies or batch-to-batch variability [6]. As a result, the sensor outputs may no longer accurately reflect the true properties of the measured samples as time progresses. In addition to drift, sensor arrays are also susceptible to noise contamination arising from environmental interference or intrinsic electronic noise [7]. The combined effects of drift and noise can severely impair measurement precision and consistency, leading to unreliable data, erroneous trend interpretation, and ultimately, faulty decision-making. Given the increasing reliance on sensor arrays for high-stakes decision-making and automated process control, it is both technologically and economically imperative to develop robust, adaptive approaches that can compensate for long-term drift and maintain data integrity throughout the lifespan of a sensor system [8].
Traditionally, drift compensation strategies for sensor arrays have relied on periodic manual recalibration and the application of statistical signal processing techniques, such as baseline correction [5,9], calibration curve updates [10,11], principal component analysis (PCA) [5], and outlier filtering [12]. While these approaches can provide partial suppression of drift and noise, especially in short-term or controlled environments, they are often inadequate for addressing the complex, nonlinear, and unpredictable patterns of drift observed in real-world, long-term deployments. Moreover, conventional calibration routines are typically labor-intensive, time-consuming, and may interrupt continuous sensor operation [13,14]. These limitations make traditional methods difficult to scale for automated systems and reduce their practicality in real-time, high-throughput, or cost-sensitive industrial scenarios.
With the recent advancements in artificial intelligence (AI), machine learning (ML) and deep learning (DL) approaches have opened up new possibilities for accurate and adaptive sensor data processing [15,16]. Many researchers have proposed AI-based techniques, such as support vector machines (SVM) [7], artificial neural networks (ANN) [8], and recurrent neural networks (RNN) [17], for sensor drift compensation, anomaly detection, and signal prediction. These data-driven methods can learn complex temporal dependencies and intrinsic relationships from historical data, thus providing more flexible and effective solutions for dynamic, real-world environments. Vergara’s seminal work introduced the Gas Sensor Array Drift (GSAD) dataset, which quickly became a pivotal benchmark for studying long-term sensor drift compensation, offering a comprehensive, publicly available testbed that provided a foundational machine learning-based solution and catalyzed advancements in developing robust methodologies to maintain sensor system reliability over extended operational periods [6]. We extended the research based on Vergara’s work by proposing a novel data preprocessing method and an inertial machine learning framework for drift compensation in sensor systems, achieving significant improvements in classification accuracy and extending sensor lifespan without additional sensor production costs [7]. Liang proposed a multibranch long short-term memory-attention ensemble classification network (MLAEC-Net) to address sensor drift in electronic noses, improving recognition accuracy and model generalization by integrating LSTM, attention mechanisms, and classifier ensembles [18]. Bhadola proposed a multiplex network framework for analyzing sensor array data, leveraging PCA and planar maximally filtered graphs (PMFG) to investigate the evolving topological properties and sensor drift effects across multiple gases over time [19].
In this study, we propose a real-time data error correction and prediction framework for long-term drift compensation in sensor arrays. On one hand, we developed an iterative random forest-based data error correction algorithm to automatically identify and correct abnormal or drifting sensor responses in real time. On the other hand, we propose an incremental domain-adversarial network, which integrates the principles of domain-adversarial learning with an incremental adaptation mechanism. The combined framework ensures robust data reliability and supports long-term, uninterrupted sensor operation.
Section 2 describes the sensor array dataset and outlines specific data defects caused by long-term deployment. Section 3 presents details of the iterative random forest data error correction method. Section 4 introduces the proposed prediction algorithm based on a deep learning network named Incremental Domain-Adversarial Network (IDAN) with an incremental adaptation mechanism. Section 5 reports comparative experimental results and analyses. Section 6 discusses the implications, advantages, and limitations of the proposed approach. Finally, Section 7 concludes this paper.

2. Dataset

The Gas Sensor Array Drift (GSAD) dataset represents a monumental milestone in the field, standing as the most important and uniquely suited dataset for the study and development of advanced algorithms specifically addressing long-term sensor drift compensation. It is, in essence, the definitive and practically sole viable choice for thoroughly investigating long-term drift phenomena and rigorously evaluating adaptive learning algorithms in electronic nose systems [6]. Compiled over a period exceeding three years at the University of California, San Diego, this dataset’s unparalleled extensive temporal coverage, systematic sampling protocol, and realistic operational conditions solidify its status as an ideal and indispensable platform for researchers aiming to evaluate the long-term reliability and adaptability of pattern recognition techniques in the face of sensor aging and environmental variability. Data acquisition was performed using an array composed of sixteen metal-oxide semiconductor (MOS) gas sensors, evenly distributed among four commercially available sensor models (TGS2600, TGS2602, TGS2610, and TGS2620) [20]. This configuration ensures not only redundancy but also diversity in gas response characteristics, thereby enhancing robustness in the recorded signals. During the experimental campaign, the sensor array was repeatedly exposed to six representative volatile organic compounds—ethanol, ethylene, ammonia, acetaldehyde, acetone, and toluene—each administered at multiple controlled concentration levels. The measurement protocol included strictly timed exposure and cleaning phases, which minimized confounding effects and maximized the reproducibility of the observations. Each data sample in the GSAD dataset consists of a 128-dimensional feature vector reflecting extensive feature extraction from raw sensor signals, with eight key features derived from dynamic response profiles [7], such as amplitude, response and recovery time, and other physically relevant metrics computed for every sensor channel; the 128-column features are shown in Table 1.
Label information corresponding to the type and concentration of the analyte is provided for each record, facilitating both classification and regression analyses. The GSAD dataset consists of 13,910 samples systematically organized into 10 distinct batches based on their acquisition period, with a detailed distribution provided in Table 2. Such batching captures the chronological progression of sensor drift, offering a realistic scenario for investigating time-dependent performance degradation and the efficacy of drift mitigation methods. The GSAD’s design and scope make it particularly valuable for research on transfer learning, domain adaptation, and online learning as applied to chemical sensing.
The GSAD dataset is presented as being automatically collected; however, our previous research [7] revealed data integrity issues within it, which we believe represents a pioneering comprehensive identification. These observed anomalies, including missing data, sign errors, misplaced decimal points, and outliers, are likely symptomatic of signal corruption by noise or transmission channel malfunctions. Subsequently, Dennler identified a latent property in the GSAD dataset that limits its suitability for gas classification studies, revealing that temporal clustering and residual drift in the baseline response can significantly distort classification results [21]. These problems may distort statistical characteristics, impair data reliability, and compromise the effectiveness of subsequent analysis or machine learning models. Despite these recognized limitations, the GSAD dataset, with its extensive application and comprehensive time span, remains an indispensable resource. Crucially, its inherent data integrity challenges and identified temporal characteristics provide a robust and realistic testbed. This unique context allows us to thoroughly validate and demonstrate the effectiveness and resilience of our real-time calibration and long-term drift compensation methods, particularly in addressing severe real-world data imperfections. Section 3 will thus visualize and address these data errors using advanced error-correction algorithms, highlighting their impact and facilitating significant improvements in data quality.

3. Real-Time Data Error Correction Method Based on Random Forest

Real-time correction of sensor drift and measurement error is critical in ensuring the reliability of long-term sensor array deployments. Traditional calibration techniques often fail to generalize across diverse operational conditions or are computationally expensive for online adaptation. To address these challenges, we propose an online random forest regression framework for real-time error correction in sensor arrays, which leverages the collective information from all sensor channels to compensate for drift and outliers dynamically and robustly. The correction framework operates in two fundamental stages: (1) model training and validation and (2) online error correction and prediction. The framework is systematically designed for each sensor channel, exploiting the multivariate interdependence of the array structure.

3.1. Model Training and Validation

3.1.1. Polarity Correction

For each incoming batch, basic physical constraints, primarily sign correction based on channel grouping, are enforced to eliminate implausible negative readings and harmonize channel orientation. To account for inherent sensor polarity and installation configuration, a sign correction rule is applied as follows:
x i , j x i , j   or   x i , j , depending on channel group ,
If 1 i % 8 5 , the value is replaced by its absolute value, thereby ensuring that channels in this group only record non-negative values. Conversely, for channels where i % 8 = 0 or 6 i % 8 7 , the negative absolute value is assigned, systematically maintaining their expected negative polarity.

3.1.2. Model Training

Let the sensor array contain N channels. As the GSAD dataset was introduced earlier, we directly include it in our experimental setup, where it consists of N = 128 sensor channels (i.e., i 1 , 2 , , N ). Each channel corresponds to a distinct sensor in the array, capturing parallel observations of the sensing environment.
For each individual sensor channel i , a dedicated random forest regressor [22] is trained. The regressor for channel i takes as input the measurements from all other channels but excludes the target channel measurement itself from the predictor set. The model learns to infer or reconstruct the reference signal for channel i based on information from the rest of the array.
A dedicated random forest regressor is trained, using the remaining channels and auxiliary metadata (excluding identifiers and the target channel itself) as predictive features. The random forest learns the mapping
f ^ i : X \ i y i ,
where y i represents the reference measurements (either ground truth or values corrected in the first iteration) for channel i , and X \ i denotes the matrix of features excluding channel i , the batch index, and other meta-variables.
The model for each channel i is optimized to minimize the expected mean squared error (MSE) [23] between predicted and reference values across all training instances. The objective function is
f ^ i = arg min f F Ε y i , j f X \ i , j 2 ,
where F is the function space corresponding to all possible random forest mappings, y i , j is the j -th reference of channel i , X \ i , j is the corresponding feature vector for the j -th sample for channel i excluding channel i , and E is the expectation for the training dataset.

3.1.3. Model Validation

The training was conducted on a selected calibration batch (batch = 1), with parallel instantiation of 128 random forest models ( n e s t i m a t o r s = 100 ), each validated using in-sample metrics such as the coefficient of determination (R2) [24] and MSE:
M S E i = 1 n j = 1 n y i , j y ^ i , j 2 ,
R i 2 = 1 j y i , j y ^ i , j 2 j y i , j y ¯ i , j 2 ,
where y ^ i , j is the predicted value and y ¯ i , j is the mean of observed values for channel i . The model, performance metrics, and diagnostic results were persistently stored for subsequent deployment and analysis, ensuring full reproducibility of the modeling process. The performance of the models trained on different batches of data across eight dimensions is shown in Table 3. As shown in Table 3, the random forest model trained using batch 1 as the training set exhibited superior performance, while the models trained with batch 1–2 or batch 1–3 showed relatively poorer performance in the F2, F5, and F8 dimensions. A detailed analysis of these results is provided in Section 3.3.

3.2. Real-Time Error Correction

Upon deployment, the trained models were used for online correction of newly acquired sensor data. The process involved the following steps.

3.2.1. Data Preprocessing

To maintain the robustness and reliability of the prediction process, the same sign correction and channel orientation harmonization approach described in Section 3.1.1 was systematically applied prior to model inference. Although the predictive models had been previously trained on data subjected to these physical constraints, unforeseen noise or anomalies in the input data could still compromise prediction accuracy or even lead to model degradation. By replicating the preprocessing steps from the training phase, model performance is safeguarded against the adverse effects of noise, sensor polarity errors, and installation inconsistencies, thereby enhancing the generalizability and stability of the deployed models.

3.2.2. Error Prediction

For each sensor channel i , the corresponding random forest model f ^ i predicts the expected true value y ^ i , j given the current feature vector X \ i , j :
y ^ i , j = f ^ i X \ i , j ,
Since the system was designed for online deployment, data was processed sequentially as it was generated by the sensors. As a result, at most time points, the feature matrix X consisted of only a single record, corresponding to the current observation X j . That is, the input to each prediction step was the latest reading, rather than a batch of samples.

3.2.3. Correction Decision Rule

A relative residual check was conducted to determine whether the observed value should be replaced. Specifically, for each nonzero observation
r i , j = y ^ i , j x i , j x i , j ,
the measurement was deemed erroneous and substituted with the model prediction
x i , j c o r r = y ^ i , j   if   r i , j > θ x i , j   otherwise ,
In this study, the threshold parameter θ was set to 0.05 during the experimental phase. This value signified that a deviation of up to 5% between the predicted value and the original value is considered acceptable. If the deviation exceeded this 5% threshold, it was deemed a significant difference, prompting a correction through replacement to address notable outliers or decimal point inaccuracies. It is important to note that this specific θ value was chosen empirically and was not determined based on statistical criteria. However, in practical applications, the selection of θ can be refined according to statistical considerations, and its value is also highly relevant to the specific problem at hand. For instance, if the numerical distribution exhibits significant volatility, a larger value for θ might be more appropriate. This refinement could potentially involve leveraging confidence intervals from historical data or optimizing based on residual distributions.

3.2.4. Iterative Correction

The data correction process was executed in two consecutive rounds. The rationale for this iterative approach lies in the inherently interconnected nature of sensor arrays, wherein measurement error or drift in one channel not only affects its own reading but may also implicitly influence regression predictions for other channels via shared latent factors or correlated dynamics. In the first iteration, the random forest models were applied to identify and correct substantial deviations in each channel, utilizing the current values of the remaining channels as predictive features. However, since the input features themselves may still have contained residual anomalies, the predictions in this pass may not have fully eradicated all data defects. By performing a second iteration over the now partially corrected data matrix, the feature vectors supplied to each regressor were of improved fidelity, thereby enabling more accurate and consistent restoration in the subsequent pass. This cascading correction strategy ensures that mutual error propagation is constrained, and the correction is no longer limited by initially corrupted inputs, leading to superior overall data quality.
The error correction module was designed using 128 independent random forest models, each corresponding to a specific sensor channel. These models were trained offline, and the primary operation during actual deployment was inference. For each incoming data sample from a single sensor channel, the prediction produced by a trained random forest model involved traversing a set of n e s t i m a t o r s decision trees. Each individual tree traversal exhibited a complexity proportional to the depth of the tree. In the approach proposed in this paper, we configured n e s t i m a t o r s = 100 , with the maximum tree depth calculated automatically during training. However, through empirical observations, we found that the maximum depth of the trees in the constructed models was 16. Consequently, the inference complexity for a single prediction on a single channel was approximately O n e s t i m a t o r s × d m a x _ d e p t h , or O 100 × 16 .
The iterative correction process for each sensor channel involved two such inference steps, resulting in a computational complexity of 2 × O 100 × 16 . This represents a relatively small constant operation count. Furthermore, the total computational cost of correcting a full array snapshot remained fixed and independent of the historical data volume. From an empirical standpoint, when all 128 models resided in memory, the observed time for correcting one complete sensor array data entry (i.e., data from all 128 channels) was approximately 1.56 milliseconds, enabling high-speed real-time operation. This parallel processing capability was highly efficient. However, even under severe memory constraints where the 128 models had to be loaded serially for each operation, a complete correction could still be accomplished in approximately 7.8 s, demonstrating a viable processing option. This performance level unequivocally confirms the real-time capability of the proposed error correction method.

3.3. Effect of Error Correction

Compared with the method proposed in reference [7], the approach introduced in this study does not require researchers to manually fit empirical correction formulas, thereby offering a more objective and universal error correction framework. This method can be readily applied to both online and offline data calibration across a wide range of sensor types, significantly improving its generalizability and practical applicability. To visually illustrate the sensor drift behavior and how the proposed random forest correction method reshapes the data distribution, we compared the original data features with those of the corrected dataset. We trained random forest prediction models using batch 1, batch 1–2, and batch 1–3 as training sets, and subsequently compared the corrected results with both the raw sensor data and the outputs obtained using the methodology outlined in reference [7]. Taking the information collected by the S5 sensor as an example, the results after the five processing methods mentioned above are shown in Figure 1. Within each subplot of Figure 1, eight distinct scatter plots are shown, corresponding to feature values F1 through F8. The dataset is organized such that different gas types are first grouped and then arranged in chronological order, with each gas type identified by a unique color. In Figure 1, these colors specifically represent Ethanol, Ethylene, Ammonia, Acetaldehyde, Acetone, and Toluene. The horizontal axis is labeled as the sorted record index, and the vertical axis indicates the feature values.
As can be observed from Figure 1, substantial outliers in Figure 1a severely deteriorated the data distribution, where the use of MinMax normalization [25] compressed most values to zero, resulting in significant data distortion. In Figure 1b, the empirical correction method from reference [7] did alleviate some of these issues, especially in terms of reducing the overall number of outliers and better representing the actual data dispersions. Nonetheless, apparent problems persist in features F3, F4, F5, and F8, where the majority of values remain compressed below 0.5 or even 0.3, indicating that the data spread is still not optimal and considerable information loss occurs. Figure 1c, which illustrates the results of random forest-based correction using batch 1 as the training set, demonstrates a significantly improved data distribution. Here, the values for each feature are more evenly distributed, and the influence of outliers is notably mitigated. This reflects the strong learning and generalization capabilities of the random forest model under this configuration, allowing it to more effectively capture nonlinear relationships and compensate for the effects of noise or drift present in the raw data. However, a further increase in the training set, as shown in panel Figure 1d,e, did not yield better correction performance. In fact, the data quality declined considerably, as observed from the increasing data compression and the re-emergence of distributional distortions. The main reason can be attributed to the accumulation of noise and drift as more batches were included in the training set, which eventually hindered the model’s error correction capacity and may have even introduced additional bias.
Based on these observations, in the subsequent experiments, we selected the corrected dataset shown in Figure 1c—achieved by applying the random forest error correction model trained solely on batch 1 data—as the primary dataset for subsequent analysis and model development. This choice ensured an optimal balance between effective error correction and the retention of key data distribution characteristics.

4. Incremental Domain-Adversarial Network

We propose an Incremental Domain-Adversarial Network (IDAN) which integrates the principles of domain-adversarial learning [26] with an incremental adaptation mechanism [27]. This approach addresses the need for efficient data error correction and drift mitigation by utilizing learning paradigms that generalize effectively over dynamic data distributions and can rapidly adapt to new domains with minimal retraining cost. The design of IDAN is depicted in Figure 2.

4.1. Network Architecture

The IDAN consists of three major modules:
  • Feature Extractor G f . The feature extractor processes an input vector x 128 (to accommodate the sensor array) using a stack of convolutional and pooling layers, followed by a fully connected transformation to obtain a compact latent representation h = G f x 64 .
  • Label Predictor G y . This branch, implemented as a fully connected layer, predicts the task class y ^ = G y h , supporting multi-class classification ( y 1 , 2 . , C , with C = 6 in this experiment).
  • Domain Classifier G d . Also a fully connected layer, the domain classifier aims to infer from which domain or batch ( d 1 , 2 , , K ) the feature embedding stems, promoting domain-invariant representation learning.
The full forward calculation can be summarized as
h = G f x ; y ^ = G y h ; d ^ = G d h ,
where y ^ , d ^ denote the predicted outputs for class and domain, respectively.

4.2. Domain-Adversarial Training

To allow the feature extractor to learn domain-invariant embeddings, we employ domain-adversarial training by simultaneously minimizing the task classification loss L o s s c l a s s and the domain classification loss L o s s d o m a i n :
L o s s c l a s s = E x , y log P y | x ,
L o s s d o m a i n = E x , d log P d | x ,
L o s s t o t a l = L o s s c l a s s + λ L o s s d o m a i n ,
where λ is a hyperparameter balancing classification and domain-adversarial objectives (set to 1 in this study for simplicity). Both loss terms use the standard cross-entropy form.

4.3. Incremental Adaptation Mechanism

Standard domain-adversarial networks are trained jointly on data from multiple domains and are not inherently incremental. IDAN introduces an online adaptation mechanism as follows:
  • Initial Training. We train the network using labeled data from the first two sensor batches (domains). Each data sample is labeled as belonging to its batch for domain-discriminative learning. Let D 1 , D 2 denote these batches.
  • Dynamic Domain Expansion. Upon the arrival of a new batch D k + 1 , IDAN extends its domain classifier G d to accommodate the new domain by expanding the output layer and copying previous weights. Given a previously trained model with K domains, the domain classifier is extended to K + 1 outputs; new weights for the additional domain are randomly initialized, while the existing ones are preserved to maintain prior knowledge. For a domain classifier with parameters W d K × h , the extension is
    W d n e w = W d w K + 1 ,
    where w K + 1 ~ N 0 , σ 2 I .
  • Online Fine-Tuning. For each incoming batch, data is first normalized using parameters fitted from the initial training batches to ensure consistent feature scaling. The new batch is incrementally introduced in small chunks (buffer size is 50 in this experiment), mixed with a reservoir of historical samples (i.e., pool). For each chunk, IDAN is fine-tuned using combined current and historic batches, adapting to the new domain while retaining knowledge of previous ones. The combined sample update for each mini-batch can be formalized as
    X t r a i n = x ( i ) , y ( i ) , d ( i ) i I n d e x h i s t I n d e x c h u n k ,
    where I n d e x h i s t denotes the set of indices for the samples selected from the historical data pool, and I n d e x c h u n k denotes the set of indices for the samples in the currently incoming data chunk. This incremental learning continues for each subsequent batch, enabling real-time drift compensation as new domains emerge.

5. Experimental Comparison and Analysis

5.1. Experimental Datasets and Environment

For the purpose of investigating long-term sensor drift, two distinct experimental datasets based on the GSAD dataset, namely Dataset1 and Dataset2, were constructed based on specific partitioning rules:
  • Dataset1: Batch 1 is exclusively used as the training set, while batches k (where k = 2, 3, …, 10) are each used independently as the test set. This sequential evaluation is designed to assess the impact of temporal drift by examining the model’s predictive performance as the time gap between training and testing increases.
  • Dataset2: Batches 1 and 2 are jointly used to form the training set, and evaluation is conducted separately on batches k (k = 3, 4, …, 10). This approach aims to determine whether the inclusion of more initial data can mitigate the adverse effects of temporal drift and improve generalization to later, temporally distant batches.
Notably, due to the absence of toluene gas samples in batches 3 to 5, configurations such as using batches 1–3, 1–4, or 1–5 as the training set were not considered, as this would undermine the completeness and validity of the classification task. Additionally, the framework posits that training data stems from either laboratory-collected or manually calibrated samples; therefore, the use of extended historical data for training would not reflect practical application scenarios.
All experimental analyses were conducted within an Anaconda environment (Python 3.11, Individual Edition) on a platform equipped with an Intel i7-14700 3.40 GHz CPU and 64 GB RAM. The implementation leverages several third-party Python libraries, specifically NumPy (v1.26.4), Pandas (v2.0.3), SciPy (v1.16.0), Scikit-learn (v1.2.2), Matplotlib (v3.7.2), and Torch (v2.1.2+cu121) to ensure reproducibility and scalability of experimentation.

5.2. Experimental Results on Dataset1

To comprehensively evaluate the effectiveness of the proposed IDAN model, extensive experiments were conducted on Dataset1. We compared IDAN with several representative baseline methods, including Support Vector Machine (SVM) [6], Inertial-SVM (ISVM) [7], Convolutional Neural Network (CNN) [28], 1D Convolutional Neural Network (1DCNN) [29], TimesNet [30], and Long Short-Term Memory Network (LSTM) [31]. For all comparative experiments, the same data partitioning scheme was adopted as described in Section 5.1, with batch 1 used for model training and batches 2–10 employed for independent testing.
The primary evaluation metric employed in this study is classification accuracy (ACC), which provides a clear comparison of different models’ abilities to handle sensor drift over time. Table 4 summarizes the ACC results for all methods across the different test batches.
To better visualize the influence of sensor drift and the temporal trend in model performance, we also present the ACC results in Figure 3. The figure illustrates, for each test batch, the accuracy trajectories of IDAN and all comparator models. It is evident that the performance gap widens over later batches, underscoring the effectiveness and robustness of IDAN in mitigating the adverse effects of long-term drift.

5.3. Experimental Results on Dataset2

In addition, we conducted a series of experiments on Dataset2, in which both batches 1 and 2 were merged to form a training set, while each of batches 3–10 was used as a separate test set. The same set of comparison methods (SVM, ISVM, CNN, 1DCNN, TimesNet, LSTM) was used, and the experimental settings remained identical to those described previously. Table 5 and Figure 4 present the experimental results on Dataset2.

5.4. Experimental Results Analysis

On Dataset1, which represents the scenario with the training set temporally furthest from most test batches, IDAN achieved an overall accuracy of 71.34%, surpassing the next-best baseline (ISVM, 52.07%) by more than 19 percentage points. While the other baseline methods exhibited significant accuracy decline as test batches shifted further in time, IDAN maintained a much steadier performance, highlighting its superior resilience to temporal drift. In test batches 7–10, IDAN consistently outperformed the second-best model by a margin of 14–35% in terms of accuracy. The performance of IDAN on Dataset1 is presented in Table 6, where the Precision (PR), Recall (RC), and F1-scores (F1) are computed using the “weighted” average rule.
On Dataset2, a similar pattern emerges. The inclusion of an additional training batch benefitted all models, yet IDAN continued to lead with the best overall accuracy of 76.29%. While traditional and deep learning baselines such as SVM and CNN improved notably on earlier test batches (e.g., batches 3–5), their results were unstable and dropped sharply in later sets (e.g., batch 10). In contrast, IDAN not only maintained high accuracy amidst severe drift, but also preserved robustness as the distribution shifted. Even during batches 7 to 10, the performance degradation remained minimal, demonstrating its adaptability to new patterns and enhanced generalization to distant temporal domains. The performance of IDAN on Dataset2 is shown in Table 7, with the performance metrics consistent with those reported in Table 6.
The traditional methods had inherent limitations in drift-prone environments. Classical models like SVM and ISVM displayed restricted resistance; deep learning approaches (CNN, 1DCNN, LSTM) tended to overfit to the training window and falter on evolved distributions. TimesNet, despite being tailored for time series, also lagged behind IDAN, likely due to insufficient modeling of complex, non-stationary drift typical in real-world sensor data. In addition, extending the training set (as in Dataset2) alleviated accuracy loss to some degree for all methods, confirming the value of expanded calibration. However, this positive effect soon plateaued and failed to fully address long-term drift, as reflected in the later test batch results. Consequently, IDAN’s edge became more prominent as the temporal gap widened, reinforcing its practicality for long-term monitoring tasks where frequent recalibration is impractical.
The superior performance of IDAN, particularly its sustained accuracy across later batches, is a direct consequence of its integrated domain-adversarial learning and incremental adaptation mechanisms. Unlike traditional models which struggle with shifting data distributions, the domain-adversarial component of IDAN explicitly forces the feature extractor to learn representations that are invariant to domain changes, effectively decoupling gas classification from temporal drift. Furthermore, the incremental adaptation mechanism allows IDAN to continuously update its knowledge with new data batches without catastrophic forgetting, a critical capability for maintaining performance over extended deployment periods.

6. Discussion

In terms of error correction, this study proposes a real-time data error correction method based on random forests, which is capable of effectively addressing measurement anomalies and compensating for sensor drift to a certain degree. This approach introduces an iterative framework that exploits the collective information from multiple sensor channels, thereby dynamically mitigating both drift and the presence of outliers. Moreover, the two-stage correction process continuously refines the predictions by utilizing inter-channel complementarity, thus reducing the propagation of errors across the sensor network.
In terms of drift compensation and online prediction, while existing strategies, such as SVM and more conventional deep learning architectures, have shown merit within static environments, they often struggle to maintain high performance levels when faced with the complexities and variabilities of long-term drift. Methods such as ISVM yield improved results compared to SVM, yet still fall short of the robustness and accuracy afforded by IDAN. The proposed IDAN framework represents a significant advancement in sensor drift compensation. Its ability to achieve robust and high accuracy under severe and long-term drift conditions addresses a major bottleneck in deploying reliable sensor arrays for critical applications. By simultaneously optimizing both classification and domain discrimination tasks, the IDAN ensures that learned representations are robust to variations often observed in real-world settings. The iterative fine-tuning process inherent in IDAN serves to bolster its resilience against drift. This iterative approach, implemented through an online adaptation mechanism, enables the model to continually update its learned parameters and maintain relevance as new data batches are introduced. This not only reduces the need for costly and time-consuming manual recalibration but also enables the development of truly autonomous and self-adapting sensor systems. The combination of real-time error correction and continuous drift compensation through incremental domain adaptation is especially crucial for industrial and environmental monitoring where sensor data is constantly evolving and immediate, accurate insights are required.
Despite the promising outcomes presented, it is crucial to acknowledge some limitations inherent to our current framework. The reliance on periodic adjustments via model retraining imposes a burden on computational resources, particularly in scenarios involving extensive datasets or limited processing capabilities. Future work may focus on optimizing the error correction algorithm and exploring more sophisticated unsupervised or semi-supervised learning techniques that can balance computational load with model fidelity.

7. Conclusions

This study demonstrates the effectiveness of real-time data error correction and prediction with online learning for long-term drift compensation in sensor arrays through the application of IDAN. By integrating domain-adversarial learning with an iterative error correction approach, IDAN not only outperforms traditional methods in handling sensor drift but also remains adaptable in the face of evolving measurement conditions. This dual capability ensures sustained data integrity and reliability over extended deployment periods, making it an essential model for contemporary sensor applications. Furthermore, while the current study focuses primarily on gas sensor arrays, the principles underlying the IDAN, coupled with real-time data error correction, can be extended to other sensor types, including temperature and humidity sensors. Future explorations can examine the applicability of our framework across multi-modal sensor systems, potentially enriching the robustness and accuracy of data across a diverse spectrum of sensing technologies.

Author Contributions

Conceptualization, X.D. and S.H.; methodology, X.D. and S.H.; software, X.D. and S.H.; validation, X.D. and S.H.; formal analysis, X.D. and S.H.; investigation, X.D. and S.H.; resources, X.D. and S.H.; data curation, X.D. and S.H.; writing—original draft preparation, X.D. and S.H.; writing—review and editing, X.D. and S.H.; visualization, X.D. and S.H.; supervision, X.D. and S.H.; project administration, X.D. and S.H.; funding acquisition, X.D. and S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Dongying City Scientific Development Fund Project, grant number DJ2022010; the Guangxi Higher Education Young and Middle-aged Teachers’ Research Capability Enhancement Project, grant number 2023KY0400; the Undergraduate Teaching Reform Project of Nanning Normal University, grant number 2025JGX049; and the Horizontal Project of Shandong Institute of Petroleum and Chemical Technology, grant number 2025HKJ0006.

Data Availability Statement

The dataset and codes reported in this study are publicly available in the GitHub repository at https://github.com/dongxr2/DriftComp-IDAN/ (accessed on 28 July 2025).

Acknowledgments

During the preparation of this manuscript, the authors used GPT-4o to assist in translating Chinese draft manuscripts into English.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Albert, K.J.; Lewis, N.S.; Schauer, C.L.; Sotzing, G.A.; Stitzel, S.E.; Vaid, T.P.; Walt, D.R. Cross-Reactive Chemical Sensor Arrays. Chem. Rev. 2000, 100, 2595–2626. [Google Scholar] [CrossRef]
  2. Askim, J.R.; Mahmoudi, M.; Suslick, K.S. Optical Sensor Arrays for Chemical Sensing: The Optoelectronic Nose. Chem. Soc. Rev. 2013, 42, 8649–8682. [Google Scholar] [CrossRef]
  3. Krantz-Rülcker, C.; Stenberg, M.; Winquist, F.; Lundström, I. Electronic Tongues for Environmental Monitoring Based on Sensor Arrays and Pattern Recognition: A Review. Anal. Chim. Acta 2001, 426, 217–226. [Google Scholar] [CrossRef]
  4. Leon-Medina, J.X.; Pineda-Muñoz, W.A.; Burgos, D.A.T. Joint Distribution Adaptation for Drift Correction in Electronic Nose Type Sensor Arrays. IEEE Access 2020, 8, 134413–134421. [Google Scholar] [CrossRef]
  5. Ziyatdinov, A.; Marco, S.; Chaudry, A.; Persaud, K.; Caminal, P.; Perera, A. Drift Compensation of Gas Sensor Array Data by Common Principal Component Analysis. Sens. Actuators B Chem. 2010, 146, 460–465. [Google Scholar] [CrossRef]
  6. Vergara, A.; Vembu, S.; Ayhan, T.; Ryan, M.A.; Homer, M.L.; Huerta, R. Chemical Gas Sensor Drift Compensation Using Classifier Ensembles. Sens. Actuators B Chem. 2012, 166–167, 320–329. [Google Scholar] [CrossRef]
  7. Dong, X.; Han, S.; Wang, A.; Shang, K. Online Inertial Machine Learning for Sensor Array Long-Term Drift Compensation. Chemosensors 2021, 9, 353. [Google Scholar] [CrossRef]
  8. Pereira, M.; Glisic, B. Detection and Quantification of Temperature Sensor Drift Using Probabilistic Neural Networks. Expert Syst. Appl. 2023, 213, 118884. [Google Scholar] [CrossRef]
  9. Grover, A.; Lall, B. A Novel Method for Removing Baseline Drifts in Multivariate Chemical Sensor. IEEE Trans. Instrum. Meas. 2020, 69, 7306–7316. [Google Scholar] [CrossRef]
  10. Ahmad, R. Enhanced Drift Self-Calibration of Low-Cost Sensor Networks Based on Cluster and Advanced Statistical Tools. Measurement 2024, 236, 115158. [Google Scholar] [CrossRef]
  11. Rudnitskaya, A. Calibration Update and Drift Correction for Electronic Noses and Tongues. Front. Chem. 2018, 6, 433. [Google Scholar] [CrossRef]
  12. Wang, S.; Wu, Z.; Lim, A. Denoising, Outlier/Dropout Correction, and Sensor Selection in Range-Based Positioning. IEEE Trans. Instrum. Meas. 2021, 70, 1007613. [Google Scholar] [CrossRef]
  13. Paul, S.; Sharma, R.; Tathireddy, P.; Gutierrez-Osuna, R. On-Line Drift Compensation for Continuous Monitoring with Arrays of Cross-Sensitive Chemical Sensors. Sens. Actuators B Chem. 2022, 368, 132080. [Google Scholar] [CrossRef]
  14. Padilla, M.; Perera, A.; Montoliu, I.; Chaudry, A.; Persaud, K.; Marco, S. Drift Compensation of Gas Sensor Array Data by Orthogonal Signal Correction. Chemom. Intell. Lab. Syst. 2010, 100, 28–35. [Google Scholar] [CrossRef]
  15. Peng, Y.; Yang, X.; Li, D.; Ma, Z.; Liu, Z.; Bai, X.; Mao, Z. Predicting Flow Status of a Flexible Rectifier Using Cognitive Computing. Expert Syst. Appl. 2025, 264, 125878. [Google Scholar] [CrossRef]
  16. Mao, Z.; Suzuki, S.; Wiranata, A.; Zheng, Y.; Miyagawa, S. Bio-Inspired Circular Soft Actuators for Simulating Defecation Process of Human Rectum. J. Artif. Organs 2025, 28, 252–261. [Google Scholar] [CrossRef] [PubMed]
  17. Liu, Q.; Dai, Y.; Li, M.; Yao, B.; Xin, Y.; Zhang, J. Real-Time Processing of Force Sensor Signals Based on LSTM-RNN. In Proceedings of the 2022 IEEE International Conference on Robotics and Biomimetics (ROBIO), Jinghong, China, 5–9 December 2022; pp. 167–171. [Google Scholar]
  18. Liang, Z.; Chen, D.; Yang, L.; Chen, Y. A Multibranch LSTM-Attention Ensemble Classification Network for Sensor Drift Compensation. IEEE Sens. J. 2024, 24, 25830–25841. [Google Scholar] [CrossRef]
  19. Bhadola, P.; Khosla, A. Multiplex Network-Based Approach to Gas Sensing. IEEE Access 2025, 13, 67588–67598. [Google Scholar] [CrossRef]
  20. Gao, L.; Tian, Y.; Hussain, A.; Guan, Y.; Xu, G. Recent Developments and Challenges in Resistance-Based Hydrogen Gas Sensors Based on Metal Oxide Semiconductors. Anal. Bioanal. Chem. 2024, 416, 3697–3715. [Google Scholar] [CrossRef]
  21. Dennler, N.; Rastogi, S.; Fonollosa, J.; van Schaik, A.; Schmuker, M. Drift in a Popular Metal Oxide Sensor Dataset Reveals Limitations for Gas Classification Benchmarks. Sens. Actuators B Chem. 2022, 361, 131668. [Google Scholar] [CrossRef]
  22. Salman, H.A.; Kalakech, A.; Steiti, A. Random Forest Algorithm Overview. Babylon. J. Mach. Learn. 2024, 2024, 69–79. [Google Scholar] [CrossRef]
  23. Mayer, M.J.; Yang, D. Potential Root Mean Square Error Skill Score. J. Renew. Sustain. Energy 2024, 16, 016501. [Google Scholar] [CrossRef]
  24. Roustaei, N. Application and Interpretation of Linear-Regression Analysis. Med. Hypothesis Discov. Innov. Ophthalmol. 2024, 13, 151–159. [Google Scholar] [CrossRef]
  25. Shantal, M.; Othman, Z.; Abu Bakar, A. Missing Data Imputation Using Correlation Coefficient and Min-Max Normalization Weighting. Intell. Data Anal. 2025, 29, 372–384. [Google Scholar] [CrossRef]
  26. Zhang, Z.; Wang, Y.; Ruan, X.; Zhang, X. Lithium-Ion Batteries Lifetime Early Prediction Using Domain Adversarial Learning. Renew. Sustain. Energy Rev. 2025, 208, 115035. [Google Scholar] [CrossRef]
  27. Lee, D.; Yoo, M.; Kim, W.K.; Choi, W.; Woo, H. Incremental Learning of Retrievable Skills For Efficient Continual Task Adaptation. Adv. Neural Inf. Process. Syst. 2024, 37, 17286–17312. [Google Scholar]
  28. Zhao, X.; Wang, L.; Zhang, Y.; Han, X.; Deveci, M.; Parmar, M. A Review of Convolutional Neural Networks in Computer Vision. Artif. Intell. Rev. 2024, 57, 99. [Google Scholar] [CrossRef]
  29. Ige, A.O.; Sibiya, M. State-of-the-Art in 1D Convolutional Neural Networks: A Survey. IEEE Access 2024, 12, 144082–144105. [Google Scholar] [CrossRef]
  30. Nagahama, H.; Yoshioka, M.; Inoue, K.; Todorokihara, M.; Omori, K. Improvement of Anomaly Detection Through Enhanced Feature Extraction in TimesNet. In Distributed Computing and Artificial Intelligence, 21st International Conference; Chinthaginjala, R., Sitek, P., Min-Allah, N., Matsui, K., Ossowski, S., Rodríguez, S., Eds.; Springer Nature: Cham, Switzerland, 2025; pp. 11–20. [Google Scholar]
  31. Al-Selwi, S.M.; Hassan, M.F.; Abdulkadir, S.J.; Muneer, A.; Sumiea, E.H.; Alqushaibi, A.; Ragab, M.G. RNN-LSTM: From Applications to Modeling Techniques and beyond—Systematic Review. J. King Saud Univ. Comput. Inf. Sci. 2024, 36, 102068. [Google Scholar] [CrossRef]
Figure 1. Results of the different processing methods on S5 sensor data, with different colors representing different gases: (a) raw data; (b) data distribution after processing with the method from ref. [12]; (c) data distribution after model correction using batch 1 data as the training set; (d) data distribution after model correction using batch 1–2 data as the training set; (e) data distribution after model correction using batch 1–3 data as the training set.
Figure 1. Results of the different processing methods on S5 sensor data, with different colors representing different gases: (a) raw data; (b) data distribution after processing with the method from ref. [12]; (c) data distribution after model correction using batch 1 data as the training set; (d) data distribution after model correction using batch 1–2 data as the training set; (e) data distribution after model correction using batch 1–3 data as the training set.
Micromachines 16 00991 g001aMicromachines 16 00991 g001b
Figure 2. The network architecture and incremental adaptation mechanism of IDAN.
Figure 2. The network architecture and incremental adaptation mechanism of IDAN.
Micromachines 16 00991 g002
Figure 3. Experimental results for different methods on Dataset1.
Figure 3. Experimental results for different methods on Dataset1.
Micromachines 16 00991 g003
Figure 4. Experimental results for different methods on Dataset2.
Figure 4. Experimental results for different methods on Dataset2.
Micromachines 16 00991 g004
Table 1. The 128-column features of the GSAD dataset.
Table 1. The 128-column features of the GSAD dataset.
Feature
Type Abbr.
Features (S1)Features (S2)Features (S3)Features (S16)
F11.9.17.121.
Δ R _ S 1 Δ R _ S 2 Δ R _ S 3 Δ R _ S 16
F22.10.18.122.
Δ R _ S 1 Δ R _ S 2 Δ R _ S 3 Δ R _ S 16
F33.11.19.123.
e m a 0.001 I _ S 1 e m a 0.001 I _ S 2 e m a 0.001 I _ S 3 e m a 0.001 I _ S 16
F44.12.20.124.
e m a 0.01 I _ S 1 e m a 0.01 I _ S 2 e m a 0.01 I _ S 3 e m a 0.01 I _ S 16
F55.13.21.125.
e m a 0.1 I _ S 1 e m a 0.1 I _ S 2 e m a 0.1 I _ S 3 e m a 0.1 I _ S 16
F66.14.22.126.
e m a 0.001 D _ S 1 e m a 0.001 D _ S 2 e m a 0.001 D _ S 3 e m a 0.001 D _ S 16
F77.15.23.127.
e m a 0.01 D _ S 1 e m a 0.01 D _ S 2 e m a 0.01 D _ S 3 e m a 0.01 D _ S 16
F88.16.24.128.
e m a 0.1 D _ S 1 e m a 0.1 D _ S 2 e m a 0.1 D _ S 3 e m a 0.1 D _ S 16
Table 2. Record distribution for each gas in different batches.
Table 2. Record distribution for each gas in different batches.
Batch IdMonth IdsTotal
Number
The Number of Records for Each Gas in Each Batch
EthanolEthyleneAmmoniaAcetaldehydeAcetoneToluene
11, 2445909883307074
23, 4, 8, 9, 1012441643341001095325
311, 12, 1315863654902162402750
414, 1516164431230120
51619728402046630
617, 18, 19, 20230051457411029606467
7213613649662360744630568
822, 232943030403314318
924, 3047061551007578101
10363600600600600600600600
Total Number13,910256529261641193630091833
Table 3. The performance of the models trained on different batches of data across eight dimensions.
Table 3. The performance of the models trained on different batches of data across eight dimensions.
Train SetDimensionMin_MSEMax_MSEMSER2
batch 1F13,863,6541824685,5230.9995
F20.07280.00260.01510.9990
F30.59130.00040.09040.9990
F40.76960.00390.17380.9984
F522.69280.04003.83870.9863
F60.20830.00030.03540.9994
F72.67540.00120.33080.9989
F848.43640.06567.61140.9816
batch 1–2F11,652,7961580324,7210.9995
F2117,0570.000798810.9278
F30.95930.00040.12410.9926
F420.62080.03463.67560.9787
F54970.10921.4989417.56560.9532
F60.14260.00010.03370.9969
F77.44040.02401.20410.9736
F8403.41417.426845.28150.9253
batch 1–3F1804,0491257174,6020.9997
F256,1620.00064700.37030.9383
F30.67750.00020.07990.9945
F411.48530.02122.04810.9865
F53608.47000.7870297.39780.9580
F60.09310.00010.02100.9985
F72.77950.01130.46740.9801
F8142.37042.498119.10850.9470
Table 4. Experimental results for all methods across the different test batches on Dataset1.
Table 4. Experimental results for all methods across the different test batches on Dataset1.
ModelsOverall ACC(%)ACC (%) of Test Set Batch
2345678910
SVM41.1376.2149.4333.5423.8533.7333.2925.5134.2541.41
ISVM52.0787.9483.3580.7573.1070.8356.8843.2045.118.31
CNN37.4667.2860.2832.3022.3424.1321.2622.1145.1143.19
1DCNN42.7248.4761.7354.0442.6432.3539.0323.4743.4043.67
TimesNet25.0247.0337.3324.2210.159.2620.5119.3934.2626.69
LSTM25.0550.0829.5732.305.0810.7418.744.082.9835.25
IDAN71.3482.7280.6461.4969.0470.8370.3381.6363.4065.42
Table 5. Experimental results for all methods across the different test batches on Dataset2.
Table 5. Experimental results for all methods across the different test batches on Dataset2.
ModelsOverall ACC(%)ACC (%) of Test Set Batch
345678910
SVM47.2888.2786.9587.332.5243.7529.2553.438.91
ISVM61.2298.890.6895.4372.2267.3747.9651.2830.67
CNN57.5897.4890.0671.5760.059.7657.8272.1332.14
1DCNN56.1793.1390.0678.1762.6156.8550.6866.3831.50
TimesNet43.9070.1847.2077.6634.2245.7840.8258.9432.92
LSTM32.6869.2340.9949.7530.0435.8215.9937.0214.61
IDAN76.2992.5093.7991.3783.3075.3769.7376.6064.47
Table 6. The performance of IDAN on Dataset1.
Table 6. The performance of IDAN on Dataset1.
Test Set BatchACC (%)PR (%)RC (%)F1 (%)
282.7284.9582.7281.97
380.6484.5580.6480.27
461.4970.0361.4958.62
569.0461.9969.0462.56
670.8394.7970.8373.11
770.3381.0470.3366.07
881.6383.9881.6381.00
963.4054.4863.4056.24
1065.4275.5465.4265.91
Table 7. The performance of IDAN on Dataset2.
Table 7. The performance of IDAN on Dataset2.
Test Set BatchACC (%)PR (%)RC (%)F1 (%)
392.5093.9592.5092.50
493.7993.8793.7993.73
591.3792.5791.3791.28
683.397.4183.3087.40
775.3786.2775.3771.63
869.3980.6769.3971.39
976.6068.8476.6070.70
1064.4767.7764.4764.88
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dong, X.; Han, S. Real-Time Correction and Long-Term Drift Compensation in MOS Gas Sensor Arrays Using Iterative Random Forests and Incremental Domain-Adversarial Networks. Micromachines 2025, 16, 991. https://doi.org/10.3390/mi16090991

AMA Style

Dong X, Han S. Real-Time Correction and Long-Term Drift Compensation in MOS Gas Sensor Arrays Using Iterative Random Forests and Incremental Domain-Adversarial Networks. Micromachines. 2025; 16(9):991. https://doi.org/10.3390/mi16090991

Chicago/Turabian Style

Dong, Xiaorui, and Shijing Han. 2025. "Real-Time Correction and Long-Term Drift Compensation in MOS Gas Sensor Arrays Using Iterative Random Forests and Incremental Domain-Adversarial Networks" Micromachines 16, no. 9: 991. https://doi.org/10.3390/mi16090991

APA Style

Dong, X., & Han, S. (2025). Real-Time Correction and Long-Term Drift Compensation in MOS Gas Sensor Arrays Using Iterative Random Forests and Incremental Domain-Adversarial Networks. Micromachines, 16(9), 991. https://doi.org/10.3390/mi16090991

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop