Next Article in Journal
Beyond Significance: Promoting Effect Size Measures in Comparing Metaheuristic Algorithms
Previous Article in Journal
Sideslip Angle Estimation for Electric Vehicles Based on Adaptive Weight Fusion: Collaborative Optimization of Robust Observer and Kalman Filter
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence Promotes Rapid Detection of Humidity in Semiconductor Manufacturing Environments

1
Department of Computer Science, Beijing Jiaotong University, Weihai 264402, China
2
School of Computing and Artificial Intelligence, Southwest Jiaotong University, Chengdu 611756, China
*
Author to whom correspondence should be addressed.
Algorithms 2026, 19(3), 190; https://doi.org/10.3390/a19030190
Submission received: 14 February 2026 / Revised: 27 February 2026 / Accepted: 28 February 2026 / Published: 4 March 2026

Abstract

Precise control of humidity levels within semiconductor manufacturing environments is paramount to ensuring product quality and yield. Unsuitable conditions can induce various wafer-related defects, including corrosion, oxidation, and poor film adhesion, thereby increasing production costs and compromising equipment reliability. This paper presents an innovative artificial intelligence-based framework, Lifelong Boosting Learning ( L 2 Boost), for rapid and accurate environmental detection within manufacturing facilities. By utilising datasets correlating sensor data with wafer defect labels, we establish links between environmental conditions and defects. Our approach employs an L 2 Boost strategy to analyse heterogeneous sensor data and identify patterns indicative of environment-induced anomalies. The proposed system enables near-real-time environmental monitoring by indirectly measuring process characteristics correlated with defects, providing an early warning mechanism for environmental control systems. Experimental results demonstrate that L 2 Boost accurately identifies environment-related defects from sensor data, achieving a macro-averaged Precision of 0.9912, Recall of 0.9804, F1 of 0.9858, and an ROC-AUC of 0.9945. This research contributes to the development of intelligent environmental monitoring systems for semiconductor manufacturing, offering a cost-effective solution for maintaining optimal production conditions. Ultimately, this framework provides actionable insights and automated diagnostic capabilities that are highly useful for process engineers, facility managers, and quality control teams striving to optimize yield in smart manufacturing environments.

1. Introduction

Semiconductor manufacturing, a cornerstone of the modern digital economy, is characterized by its exceptional technological complexity and capital intensity [1]. The fabrication of integrated circuits (ICs) entails hundreds of meticulously controlled steps within highly specialized cleanrooms, where precise regulation of environmental parameters—including temperature, particle concentration, air pressure, vibration, and humidity—is paramount for achieving high yields and product reliability [2]. Among these, humidity control is particularly critical due to the pervasive and multifaceted interactions of water molecules with semiconductor materials, chemicals, and process equipment [3].
Even trace amounts of moisture can induce significant detrimental effects throughout the manufacturing process. During front-end-of-line (FEOL) stages, adsorbed moisture can accelerate native oxide growth on silicon wafers, compromising surface integrity and disrupting subsequent epitaxial or gate oxide formation [4]. In photolithography, humidity fluctuations alter photoresist properties, leading to critical dimension variations and patterning errors [5]. During etching and deposition, moisture acts as an unintended contaminant, resulting in incomplete reactions, poor film coverage, and incorporated impurities [6]. In back-end-of-line (BEOL) processes, moisture, especially in the presence of ionic contaminants and electrical bias, accelerates electrochemical migration and corrosion in metallic interconnects, causing dendrite formation, voids, and circuit failures that severely impact device lifetime and reliability [7,8]. The economic stakes are substantial: a single humidity excursion can lead to the loss of entire wafer batches, each valued at tens to hundreds of thousands of dollars, alongside costly production downtime [9].
Conventional humidity monitoring in fabrication facilities relies on networks of physical sensors (e.g., capacitive, resistive) deployed at discrete locations [10]. While essential, this approach suffers from inherent limitations: limited spatial resolution; a temporal disconnect between a humidity event and its observable impact on product quality; and an indirect, probabilistic correlation between sensor readings and wafer yield [11]. These challenges underscore the need for a paradigm shift from merely monitoring ambient conditions to establishing a direct, near-real-time link between environmental factors and product outcomes.
Artificial Intelligence (AI) offers transformative potential in this regard. While prior research has extensively applied machine learning (ML)—particularly deep learning-based computer vision—to automated wafer defect inspection and classification [12,13], the focus has largely remained on post facto quality control. A promising yet underexplored direction is leveraging AI not just to classify defects, but to use product-related data (e.g., defect labels coupled with in situ process sensor logs) as a high-resolution, distributed sensor network for proactive environmental diagnostics. Specifically, if AI models can learn the subtle signatures of humidity-induced anomalies from multi-source process and sensor data, they could enable rapid, indirect detection of humidity issues, effectively using the process outcome and its associated data as a sensitive environmental probe [14].
However, most existing AI models for semiconductor monitoring are static, trained on historical data and deployed with fixed parameters. This poses a significant limitation in dynamic manufacturing environments where operational conditions, process recipes, and tool states continuously evolve, causing data distributions to shift over time. A model that cannot adapt may experience performance degradation, failing to detect novel or gradually emerging humidity-related patterns.
To address this gap, this paper proposes an AI-driven framework centered on Lifelong Learning for rapid humidity detection in semiconductor manufacturing [15]. Our core premise is that by integrating an adaptive, continuously learning mechanism, an AI system can maintain high detection accuracy over extended periods, adapting to new data batches and evolving environmental patterns without requiring complete retraining. The proposed framework moves beyond static data analysis. It employs a Lifelong Boosting Learning strategy, which combines ensemble methods with a knowledge retention and reuse mechanism to continuously adapt to streaming sensor data. This allows the model to incrementally learn from sequential data arrivals, preserving valuable historical knowledge while efficiently incorporating new information about humidity-related process deviations.
Based on the current research situation, the following research questions are proposed.
  • How can artificial intelligence leverage product-related data to establish a direct, near-real-time link between environmental factors and product outcomes for proactive environmental diagnostics?
  • How can an AI system continuously adapt to evolving operational conditions and data distribution shifts in dynamic manufacturing environments without experiencing performance degradation or requiring complete retraining?
The main contributions of this study are as follows.
  • A novel “environmental probe” paradigm is proposed, wherein heterogeneous process and wafer defect data are leveraged to establish a direct, near-real-time diagnostic link between ambient humidity and product quality outcomes.
  • The Lifelong Boosting Learning ( L 2 Boost) framework is developed, which integrates an adaptive ensemble mechanism with time-weighted knowledge retention to effectively mitigate catastrophic forgetting and adapt to non-stationary manufacturing patterns.
  • A robust evaluation methodology for environmental diagnostics under extreme class imbalance is established; it is demonstrated through macro-averaged metrics that the proposed system achieves superior detection sensitivity even when subject to stochastic sensor perturbations.
The remainder of this paper is organized as follows: Section 2 presents a comprehensive review of related works in environmental monitoring, wafer defect classification, and smart manufacturing. Section 3 describes the proposed L 2 Boost framework, including problem definition and algorithm design. Section 4 details the experimental evaluation, baseline comparisons, and result analysis. Section 5 provides a discussion of the findings, implications, and current limitations. Finally, Section 6 concludes the paper and outlines future research directions.

2. Related Works

In this section, we provide a comprehensive review of research streams closely related to the AI-driven framework proposed in this study.
The pursuit of robust humidity control and its intricate relationship with product quality in semiconductor manufacturing has spawned extensive research across multiple disciplines, including materials science, process engineering, environmental control, and, more recently, data science. This section reviews the foundational and contemporary work in three interconnected domains: (1) the impact of humidity and established monitoring techniques in fabs, (2) the evolution of wafer defect inspection and classification methodologies, and (3) the proliferation of AI and data-driven approaches in smart manufacturing. Critically, we examine the current state of the art at the intersection of these domains, identifying a conspicuous gap that the present study aims to address—specifically, the lack of an adaptive, continuously learning system that can perform real-time humidity diagnostics from product-inspection data.
The criticality of atmospheric moisture control in IC production has been understood since the industry’s early days. Foundational research meticulously documented the mechanisms of humidity-induced failure. Studies on surface chemistry elucidated how water molecules physisorb and chemisorb on silicon and other semiconductor surfaces, leading to uncontrolled oxide growth and surface state changes that degrade device performance [16]. In the realm of metallization, the electrochemistry of corrosion processes—particularly for aluminum and, later, copper interconnects—was detailed, establishing models for failure acceleration under conditions of elevated humidity and bias [17]. Furthermore, process-specific investigations revealed how humidity affects photoresist kinetics, the stability of etching plasmas, and the nucleation kinetics in chemical and physical vapor deposition chambers [18]. This body of work firmly established the quantitative relationships between Relative Humidity (RH) levels, often in the parts-per-million (ppm) range for specific process steps, and the propensity for various defect modes.
To manage this risk, the industry adopted a multi-faceted engineering approach centered on physical sensor networks. The cleanroom environmental control system (ECS) is typically governed by a vast array of hygrometers. Capacitive polymer sensors are most common for general ambient monitoring due to their good balance of accuracy, response time, and cost [19]. For more demanding applications, such as inside lithography scanners or deposition tools, chilled mirror hygrometers, which provide a fundamental dew-point measurement, are often employed for their high precision. These sensors feed data into a supervisory control and data acquisition (SCADA) or distributed control system (DCS), which orchestrates heating, ventilation, and air conditioning (HVAC) units, desiccant dryers, and humidifiers to maintain setpoints.
However, the limitations of this traditional paradigm are well-recognized in both academic and industrial literature. The challenge of spatial resolution is frequently cited; a sensor provides a point measurement that may not represent conditions a few meters away, especially near equipment with local heat loads or chemical emissions. The concept of “micro-environments” within the broader cleanroom is a key area of study. Researchers have used computational fluid dynamics (CFD) modeling to predict airflow and humidity gradients, and some have proposed dense grids of low-cost wireless sensor nodes to empirically map these variations [20]. Another significant limitation is the latency between a humidity excursion and its confirmation via product metrology. This delay hinders rapid corrective action and complicates root cause analysis, as engineers must retrospectively correlate sensor logs with electrical test results from wafers processed days or weeks earlier [21]. While advanced sensors and tighter control loops have improved, the fundamental disconnection between the environmental parameter (humidity) and the quality metric (wafer defects) remains a persistent challenge.
Parallel to environmental control, the field of wafer defect inspection has undergone its own revolution, evolving from manual microscope examination to fully automated, high-speed image acquisition and analysis. Automated Optical Inspection (AOI) and Scanning Electron Microscope (SEM) review tools are now ubiquitous in fabs. These systems generate terabytes of image data daily, depicting the patterned wafer surface at high resolution. The initial task of defect detection—finding anomalous pixels against a complex background—is largely solved through sophisticated image subtraction techniques, comparing a die image to a reference die or a golden template [22].
The subsequent and more complex task is defect classification: categorizing each detected anomaly into a predefined type (e.g., particle, scratch, pattern bridge, residue, pit). The classification is crucial for determining the defect’s root cause and initiating the correct corrective action. Historically, this was accomplished using rule-based algorithms and handcrafted feature descriptors. Engineers would define thresholds for geometric features (size, aspect ratio, perimeter), texture, and intensity profiles [23]. While effective for simple, high-contrast defects, these methods struggled with subtle or complex defect morphologies and required constant tuning as new process layers were introduced.
The advent of ML, and later deep learning, fundamentally transformed this field. Early ML approaches utilized classical algorithms like Support Vector Machines (SVMs), decision trees, and K-Nearest Neighbors (KNN), but they still relied on human experts to define and extract the relevant feature vectors from defect images [24]. The breakthrough came with the application of deep Convolutional Neural Networks (CNNs). CNNs automate the feature extraction process, learning hierarchical representations of visual patterns directly from raw pixel data. Models based on architectures like ResNet, VGG, and Inception, often pre-trained on large natural image datasets and fine-tuned on proprietary wafer image datasets, have demonstrated classification accuracies exceeding 100% in research settings [25]. These models can distinguish between dozens of defect types with a consistency and speed unattainable by human inspectors or older algorithms [26].
This powerful capability has turned defect classification from a purely quality control activity into a rich source of data for fab intelligence. Modern systems not only classify defects but also cluster them, track their spatial signatures (e.g., wafer-edge rings, random distributions, reticle fingerprints), and correlate them with process tool and chamber histories [27]. The defect map has become a diagnostic language. However, the translation of this language to pinpoint specific environmental root causes, like a precise humidity condition, remains largely interpretive and reliant on the experience of process integration engineers. The link is associative and manual, not yet encoded into the AI models themselves.
The application of AI in semiconductor manufacturing extends far beyond defect inspection, forming the core of the Industry 4.0 or “Smart Manufacturing” transformation. This movement seeks to leverage the massive volumes of data generated by equipment sensors (equipment engineering systems, or EES), process state logs, metrology tools, and testers to optimize every aspect of production [28].
A prominent application is predictive maintenance (PdM), where ML models analyze multivariate time-series data from pumps, robots, and plasma sources to predict impending failures before they cause unscheduled downtime. Virtual metrology (VM) uses sensor data from a process tool to predict the outcome of a physical measurement (like film thickness or critical dimension), reducing the need for slow, destructive sampling [29]. Run-to-run (R2R) control employs algorithms to adjust recipe parameters for each subsequent wafer lot based on measured results from the previous lot, compensating for process drift. Yield prediction models attempt to forecast the final test yield of a wafer or lot based on early metrology and inspection data, enabling early disposition of potentially bad material [30].
A subset of this research focuses specifically on the impact of the fab environment on yield and quality. Studies have applied statistical methods and basic ML models (like linear regression or random forests) to find correlations between aggregated environmental data—average temperature, humidity, particle counts—and final yield metrics. Some have explored more sophisticated time-series analysis to link environmental excursions recorded by the facility system with yield dips observed later in the production timeline [31]. These studies affirm that environmental factors are significant variables in yield models.
Despite the mature state of research in the three domains outlined above, a critical synthesis is missing. On one side, we have highly advanced AI models that excel at classifying wafer defects from images. On the other, we have a well-understood physical problem: humidity causes specific, identifiable types of defects. Yet, the current state of the art keeps these threads separate. Environmental monitoring systems produce humidity data streams. Defect inspection systems produce classified defect maps and images. Engineers and yield analysis software attempt to correlate them post facto.
The gap is the absence of an AI framework that directly internalizes the physics of humidity-induced failure by learning to recognize its visual signature as a primary classification objective. Existing defect classifiers may identify “corrosion” or “haze,” but they do not explicitly attribute it to an environmental root cause during the classification act. Furthermore, the potential to use the spatial and temporal distribution of these classified defects as a high-resolution, product-integrated sensor network for humidity is largely unexplored. No published work, to our knowledge, has proposed and demonstrated an end-to-end AI model trained specifically to detect and diagnose humidity issues by analyzing wafer defect imagery as the primary input signal, thereby creating a closed-loop link between product quality and environmental health.
Importantly, while AI methods for defect classification and environmental correlation have advanced, most existing models are static—trained on fixed datasets and deployed without ongoing adaptation. Semiconductor manufacturing, however, is a dynamic environment where process conditions, tool states, and ambient parameters continuously evolve. This leads to data distribution shifts over time, which can degrade the performance of static models. There is a pressing need for AI systems that can learn continuously, adapting to new patterns while retaining previously acquired knowledge—a capability known as lifelong or continual learning. Yet, the integration of lifelong learning mechanisms into environmental diagnostics for semiconductor manufacturing remains notably absent in the literature.
This paper positions itself directly within this gap. Our work is distinct from (a) pure environmental monitoring research, as we use the product (wafer) as the sensor; (b) pure defect classification research, as we specifically curate and model a subset of defects tied to a single environmental factor; and (c) broad yield–environment correlation studies, as we propose a real-time, image-based diagnostic tool rather than a retrospective statistical model. Furthermore, we introduce a lifelong boosting learning framework that enables continuous adaptation and knowledge retention, addressing the dynamic nature of semiconductor production environments. By bridging these fields, we aim to demonstrate a novel paradigm for intelligent environmental assurance in semiconductor manufacturing.

3. Lifelong Boosting Learning for Environment Humidity Detection

This section provides a detailed elaboration of the proposed L 2 Boost framework, which is designed to enable rapid and adaptive detection of humidity anomalies in semiconductor manufacturing environments by continuously learning from sequential data batches.

3.1. Motivation

Humidity control is critical in semiconductor manufacturing, as slight deviations can cause electrostatic discharge, material degradation, lithography defects, and yield loss. Although modern facilities deploy dense physical sensor networks, most existing detection systems rely on fixed thresholds or static ML models, which face significant limitations in dynamic production settings. Environmental humidity is highly non-stationary due to fluctuations in production load, equipment states, airflow, and seasonal variations, leading to performance degradation in models trained offline. Rapid detection is particularly challenging, as humidity anomalies can emerge and propagate quickly, while conventional approaches lack the agility to adapt in real time. Furthermore, long-term knowledge reuse is seldom implemented; recurring humidity patterns across production batches are rarely systematically captured or leveraged. To address these gaps, we propose a L 2 Boost framework that integrates adaptive ensemble learning with incremental knowledge retention and reuse. This design enables fast, robust, and continuously improving humidity detection by learning from data streams over time, effectively handling the evolving nature of semiconductor production environments [32,33].

3.2. Problem Definition

In semiconductor fabrication, environmental humidity must be maintained within a strict operational range to prevent defects. Let H t denote the observed humidity level at time t. The safe operational range is defined as [ H low , H high ] , where H low and H high are process-dependent thresholds. A humidity anomaly is declared if H t deviates beyond this range, i.e., H t < H low or H t > H high .
Given a multi-dimensional feature vector x t R d extracted from process sensors and contextual data at time t, the goal is to learn a detection function f : R d 0 , 1 that predicts whether the corresponding humidity level is abnormal ( y t = 1 ) or normal ( y t = 0 ).
The problem is formulated as a binary classification task over sequential data batches D 1 , D 2 , , D t , where each batch D s = ( x i , y i ) i = 1 N s corresponds to a production period. The model must adapt to potential distribution shifts across batches while retaining useful knowledge from previous batches, without catastrophic forgetting.
To quantify detection performance, we define a weighted loss function that accounts for both false alarms and missed detections. Let ( f ( x i ) , y i ) be a differentiable loss (e.g., binary cross-entropy). The objective for batch D s is to minimize the empirical risk, as defined in Equation (1):
min f F i = 1 N s w i ( s ) · f ( x i ) , y i + λ r e g · Ω ( f ) ,
where w i ( s ) is a sample weight that can be adjusted to emphasize hard or critical examples, λ r e g controls the regularization strength and Ω ( f ) is a knowledge-driven regularization term. To encourage knowledge reuse and mitigate catastrophic forgetting, we define Ω ( f ) as the functional distance between the current candidate model f and the ensemble of previously stored experts M t 1 , as presented in Equation (2):
Ω ( f ) = d i s t ( f , M t 1 )
where d i s t ( · ) can be implemented as the Kullback–Leibler (KL) divergence or a parameter-based Euclidean distance. This constraint ensures that the updated model incorporates new humidity patterns while remaining consistent with previously acquired environmental knowledge.
An activation (alert) function A ( · ) is defined to map the model’s continuous output (e.g., probability of anomaly) to a discrete alert decision. Let p t = f ( x t ) [ 0 , 1 ] be the predicted anomaly probability. The alert is triggered according to Equation (3):
A ( p t ; τ ) = I ( p t τ ) ,
where τ ( 0 , 1 ) is a sensitivity threshold that can be tuned to balance precision and recall, and I ( · ) is the indicator function. In practice, τ can be adapted online based on the observed false alarm rate or operational requirements.

3.3. Lifelong Boosting Model

The proposed lifelong learning framework is designed to handle environmental data in semiconductor manufacturing. Unlike traditional methods that either retrain the model from scratch or selectively reuse weak learners, our framework adopts a “time-weighted ensemble of ensembles” strategy. This method generates a dedicated strong classifier for each incoming data batch and dynamically integrates them into a global model through a time decay mechanism, ensuring that the system prioritizes the latest data patterns while retaining historical knowledge to prevent catastrophic forgetting.
The architecture of this framework is shown in Figure 1. Its workflow is carried out in a linearly cumulative manner over time steps ( t = 1 , 2 , , t current ):
1.
Data Input: The system receives sensor data streams in discrete batches (e.g., D 1 , D 2 , , D t ).
2.
Specific Batch Model Training: For each batch D t , a robust ensemble model C ( t ) ( x ) is trained. This step builds a powerful classifier by aggregating multiple weak learners (such as decision trees) for the difficult samples in that specific batch.
3.
Model Storage: Unlike discarding historical models, each trained batch model C ( t ) ( x ) is serialized and stored in the historical knowledge base. This forms an expert sequence, with each expert focusing on the environmental distribution of their corresponding time period.
4.
Model Fusion: The core of the lifelong learning mechanism is the fusion layer. This framework aggregates the decisions of all stored models ( C ( 1 ) , , C ( t ) ) to construct the final global model M t ( x ) .
5.
Real-time Detection: The fused global model is used for real-time monitoring, and abnormal alerts are output based on the consensus of historical and current experts.
Let D t represent the data batch arriving at time t. For this batch, we train a powerful classifier C ( t ) ( x ) , which aggregates T weak learners with weights α using an algorithm expressed in Equation (4):
C ( t ) ( x ) = arg max y j = 1 T α j δ ( C j ( x ) = y )
where C j represents the t-th weak learner trained on D t , and α j is its corresponding weight based on classification error.
To construct the global lifelong model M t ( x ) at the current time t, we fuse all historical models C ( i ) ( x ) (where i { 1 , , t } ). We introduce a Time-Decay Weighting Function W ( i , t ) to address concept drift. The hypothesis is that recent models are more reflective of the current manufacturing environment than older models. The weighting function is defined as shown in Equation (5):
W ( i , t ) = λ t i
where λ ( 0 , 1 ] is the forgetting factor. When i = t (current batch), the weight is λ 0 = 1 (maximum influence). As the time gap ( t i ) increases, the influence of older models decays exponentially.
The final global decision function is given by the weighted majority vote of all batch models, computed via Equation (6):
M t ( x ) = arg max y i = 1 t W ( i , t ) · C ( i ) ( x ) = arg max y i = 1 t λ t i · C ( i ) ( x )
This form enables the framework to continuously adjust. The new pattern is captured by the model with the highest weight C ( t ) , while recurring historical anomalies are identified by a series of models from C ( 1 ) to C ( t 1 ) , thereby forming a stable and reliable lifelong learning detection mechanism.

3.4. Algorithm Design

The program implementation details of the lifelong enhanced learning framework are elaborated in Algorithm 1. This algorithm performs “Training—Storage—Fusion— Detection” for each new received data batch to ensure continuous adaptation to the semiconductor manufacturing environment.
Algorithm 1: Lifelong Boosting Learning for Environment Humidity Detection
  • Require: New data batch D t at time t, Historical Model Repository R t 1 = { C ( 1 ) , , C ( t 1 ) } , Decay factor λ , Boosting rounds T, Alert Threshold τ .
  • Ensure: Anomaly alerts for D t , Updated Repository R t .
      1:
Step 1: Data Input
      2:
Input sensor data stream D t .
      3:
Step 2: Batch Model Training
      4:
Initialize sample weights w i ( s ) = 1 / N for training samples in D t .
      5:
Initialize batch model ensemble E t = .
      6:
for  j = 1 to T do
      7:
   Train weak learner C j on D t using weights w i ( s ) .
      8:
   Calculate error ϵ j and learner weight α j .
      9:
  Update sample weights w i ( s ) (increase weight for misclassified samples).
    10:
   Add to ensemble: E t E t { ( C j , α j ) } .
    11:
end for
    12:
Formulate batch strong classifier: C ( t ) ( x ) = arg max y j α j δ ( C j ( x ) = y ) .
    13:
Step 3: Model Storage
    14:
Store the newly trained model: R t R t 1 { C ( t ) } .
    15:
Step 4: Global Model Fusion
    16:
for  i = 1 to t do
    17:
   Calculate time weight: W ( i , t ) = λ t i .
    18:
end for
    19:
Construct global model M t ( x ) using Equation (6):
    20:
    M t ( x ) = arg max y i = 1 t W ( i , t ) · C ( i ) ( x )
    21:
Step 5: Inference Calculation
    22:
Apply global model M t ( x ) to the testing stream of D t .
    23:
Compute the anomaly probability score P ( x ) for current sensor readings.
    24:
Step 6: Real-time Semiconductor Environment Detection
    25:
return Result R t
The computational complexity of Algorithm 1 consists of a constant training cost and a linearly growing inference cost. Specifically, training a new batch model C ( t ) (Step 2) depends only on the boosting rounds T and batch size N, typically O ( T · N log N ) for decision trees, ensuring efficient updates independent of the accumulated history. In contrast, the Global Model Fusion and Inference phases (Steps 4–5) require aggregating predictions from all t historical models stored in R t , leading to a time and space complexity of O ( t · T ) . This linear growth reflects the trade-off between preventing catastrophic forgetting and resource consumption, although the inference latency can be effectively mitigated by parallelizing the evaluation of independent historical estimators.

4. Experiment

This section presents a comprehensive experimental evaluation of the proposed Lifelong Boosting Learning framework for environment humidity detection. The experiments are designed to: (1) verify the superior classification capability of boosting methods on the complex, non-stationary patterns typical of semiconductor manufacturing data; and (2) provide the empirical foundation that justifies the necessity and design of the lifelong learning mechanism for long-term, multi-batch deployment. The following subsections detail the dataset, baseline algorithms, evaluation metrics, and a critical analysis of the results.

4.1. Datasets

To validate the proposed framework, we conducted experiments on the publicly available semiconductor wafer defect classification dataset sourced from Kaggle [34]. A statistical summary is provided in Table 1. Regarding data integrity, all 5050 initial samples were complete and fully valid. An exhaustive inspection confirmed that there were no missing values across any of the feature dimensions; therefore, no data cleaning or removal of incomplete records was necessary prior to our experiments. To address the extreme class imbalance observed in the initial dataset, a data augmentation strategy was employed to ensure the model’s robustness and generalization. The single minority sample was replicated 50 times to construct a synthetic anomaly pool. To prevent the models from simply memorizing a single data point and to simulate real-world sensor fluctuations, additive Gaussian noise was applied to each replicated sample:
x i = x a n o m a l y + ϵ , ϵ N ( 0 , σ 2 )
where x a n o m a l y is the original feature vector, and ϵ was set to 5% of the standard deviation of each respective feature. This resulted in an augmented experimental set comprising 4999 normal samples and 51 anomaly samples, providing a more reliable basis for evaluating detection sensitivity under non-stationary conditions.

4.2. Baseline Algorithms and Evaluation Metrics

This section presents a comprehensive description of the experimental configuration and evaluation protocol for the proposed method. We first introduce the baseline algorithms used for comparison, followed by the evaluation metrics adopted to assess model performance. Subsequently, the experimental results are reported and analyzed, with particular emphasis on the comparative performance between the proposed approach and several representative classification methods.
  • K-Nearest Neighbors (KNN) [35]: a non-parametric instance-based learning algorithm that classifies a sample according to the majority label among its k nearest neighbors in the feature space.
  • Random Forest (RF) [36]: an ensemble learning method that constructs multiple decision trees using bootstrap sampling and random feature selection, and outputs predictions via majority voting.
  • Logistic Regression (LR) [37]: a probabilistic linear classifier that models the posterior probability of class membership using the logistic (sigmoid) function.
  • Linear Discriminant Analysis (LDA) [38]: a statistical classification method that projects data onto a lower-dimensional space by maximizing between-class variance while minimizing within-class variance.
  • Multi-Layer Perceptron (MLP) [39]: a feedforward neural network composed of multiple fully connected layers with nonlinear activation functions, capable of modeling complex decision boundaries.

4.3. Evaluation Metrics

To quantitatively evaluate classification performance, four commonly used evaluation metrics are employed: Precision, Recall, F1 [40], and ROC-AUC [41]. Given the high degree of class imbalance in the semiconductor manufacturing dataset, these metrics (excluding ROC-AUC) are calculated using macro-averaging to ensure that the detection performance of the rare anomaly class is accurately reflected.
Precision (Macro-averaged) measures the average proportion of correctly predicted samples for each class among all samples assigned to that class. It is defined in Equation (8):
Precision = 1 2 i = 0 1 T P i T P i + F P i
where i { 0 , 1 } represents the normal and anomaly classes respectively, T P i denotes the true positives for class i, and F P i denotes the false positives for class i.
Recall (Macro-averaged) (also known as sensitivity) evaluates the ability of the classifier to correctly identify samples belonging to each specific class. It is calculated in Equation (9):
Recall = 1 2 i = 0 1 T P i T P i + F N i
where F N i represents the number of false negatives for class i.
F1 (Macro-averaged) provides a balanced measure of Precision and Recall by computing their harmonic mean across both classes. It is defined in Equation (10):
F 1 = 2 · Precision · Recall Precision + Recall
This metric is particularly suitable for imbalanced datasets, as it accounts for both false positives and false negatives, preventing the results from being biased toward the majority distribution.
ROC-AUC (Area Under the Receiver Operating Characteristic Curve) assesses the overall discriminative capability of a classifier across different decision thresholds. It reflects the trade-off between the true positive rate ( T P R ) and the false positive rate ( F P R ), and is defined in Equation (11):
AUC = 0 1 T P R ( F P R 1 ( x ) ) d x
where T P R = T P T P + F N and F P R = F P F P + T N . A higher AUC value indicates stronger class separability and better overall classification performance.

4.4. Result Analysis

The performance comparison between the proposed L 2 Boost framework and several baseline classifiers on the augmented semiconductor wafer defect dataset is summarized in Table 2. To ensure a rigorous evaluation under the extreme class imbalance (4999 normal vs. 51 anomaly samples), macro-averaged metrics are employed. This approach ensures that the detection sensitivity for rare humidity-induced anomalies is given equal weight to the majority class.
As demonstrated in Table 2, the baseline models exhibit varying degrees of effectiveness in identifying the augmented anomaly signatures. Traditional algorithms, such as KNN and LDA, yield F1 of 0.6957 and 0.7431, respectively. These results indicate that while these models can capture some basic statistical distributions, they are highly sensitive to the stochastic Gaussian perturbations introduced during data augmentation. The higher Recall compared to Precision in several baselines suggests a tendency to over-generalize the anomaly boundaries, potentially leading to increased false alarm rates in a production environment. Logistic Regression (LR) and Multi-Layer Perceptron (MLP) achieve moderate performance with F1 of 0.8225 and 0.8154; however, their discriminative power remains insufficient for the high-precision requirements of semiconductor cleanroom monitoring.
In contrast, the proposed L 2 Boost framework achieves state-of-the-art performance across all metrics, with a macro-averaged F1 of 0.9858 and an ROC-AUC of 0.9945. The slight deviation from a perfect score (Recall of 0.9804) provides empirical evidence of the framework’s realistic response to sensor noise and data variance, indicating that a robust and generalized decision boundary has been successfully learned. This superior resilience is primarily attributed to the “ensemble of ensembles” architecture and the time-decay weighting mechanism W ( i , t ) . By prioritizing the most relevant batch-specific experts and effectively managing non-stationary patterns, L 2 Boost demonstrates the capacity to distinguish subtle humidity signatures from complex background noise. These findings validate that the proposed method can effectively transform heterogeneous sensor data into a high-resolution “environmental probe,” enabling precise and reliable humidity diagnostics in dynamic manufacturing settings.

5. Discussion

This study presents the Lifelong Boosting Learning framework for monitoring ambient humidity in semiconductor manufacturing through the analysis of process sensor data correlated with wafer defects. Experimental results demonstrate that L 2 Boost delivers superior classification performance, validating the efficacy of the proposed method in capturing subtle patterns associated with humidity-induced anomalies. This finding is significant as it provides an indirect, data-driven method for humidity surveillance and directly links process parameters to environmental conditions, facilitating a paradigm shift from passive environmental monitoring to active, product-informed diagnostics.
The strength of L 2 Boost lies in its innovative integration of lifelong learning principles with boosting ensemble methods. By using process and defect data as a distributed “sensor,” the framework overcomes spatial resolution limitations and temporal delays inherent in traditional point-based hygrometers. The lifelong learning mechanism allows L 2 Boost to continuously adapt to data distribution shifts during production, enabling systematic knowledge accumulation and reuse—a crucial capability for long-term, multi-batch semiconductor manufacturing.
However, several limitations should be acknowledged. Crucially, while the L 2 Boost framework is mathematically designed to handle sequential data batches and continuous adaptation, our current experimental validation was conducted on a static dataset. This initial evaluation primarily served to establish baseline classification metrics and validate the theoretical mechanics of the ensemble structure. We acknowledge that a true validation of its continual learning capabilities requires sequential, time series data streams that exhibit concept drift. Second, the model’s performance heavily relies on the quality of labeled humidity-related defect data, and in practice, defects often result from multiple interacting environmental and process factors. Therefore, testing the framework on dynamic data streams from an operational environment is a primary focus for our future work, alongside extending the framework to multi-parameter diagnostics.
Future research directions include: (1) incorporating semi-supervised or self-supervised learning to reduce dependency on large labeled datasets; (2) combining physics-based models with data-driven approaches to enhance interpretability and generalization; and (3) extending L 2 Boost to monitor other critical environmental parameters (e.g., temperature, airborne particles), thereby constructing a multi-dimensional environmental sensing system for semiconductor manufacturing.

6. Conclusions

In this study, the L 2 Boost framework is proposed as a potential data-driven approach for indirect humidity detection in semiconductor manufacturing. By integrating an ensemble-based lifelong learning mechanism with a time decay weighting function, the system demonstrates the capability to mitigate catastrophic forgetting while adapting to non-stationary production conditions. Evaluated on an augmented dataset to address extreme class imbalance, the framework achieved a macro-averaged F1 of 0.9858 and an ROC-AUC of 0.9945. However, we explicitly acknowledge that because these high metrics are derived from a synthetically augmented minority class, they primarily serve as a proof-of-concept rather than a definitive guarantee of field performance. Despite this limitation, these initial results suggest that the proposed methodology can effectively transform wafer defect data into a conceptual “environmental probe”, offering a valuable foundation for the development of resilient and intelligent manufacturing systems, pending further empirical validation on dynamic, real-world industrial data streams.

Author Contributions

Conceptualization, F.Y.; Methodology, F.Y.; Software, F.Y.; Validation, F.Y.; Formal analysis, F.Y.; Investigation, T.L.; Resources, F.Y.; Data curation, F.Y.; Writing—original draft, F.Y.; Writing—review and editing, F.Y.; Visualization, F.Y.; Supervision, H.W.; Project administration, H.W.; Funding acquisition, H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work are supported by the National Natural Science Foundation of China under Grant (No. 62276216) and the Natural Science Foundation of Sichuan Province under Grant (No. 2024NSFSC0501).

Data Availability Statement

The data presented in this study are available on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
L 2 BoostLifelong Boosting Learning
ICsIntegrated Circuits
FEOLFront-End-Of-Line
BEOLBack-End-Of-Line
AIArtificial Intelligence
MLMachine Learning
RHRelative Humidity
ECSEnvironmental Control System
SCADASupervisory Control And Data Acquisition
DCSDistributed Control System
HVACHeating, Ventilation, And Air Conditioning
CFDComputational Fluid Dynamics
SEMScanning Electron Microscope
AOIAutomated Optical Inspection
SVMsSupport Vector Machines
KNNK-Nearest Neighbors
CNNsConvolutional Neural Networks
VMVirtual metrology
RFRandom Forest
LRLogistic Regression
LDALinear Discriminant Analysis
MLPMulti-Layer Perceptron
ROC-AUCArea Under the Receiver Operating Characteristic Curve
MESManufacturing Execution Systems

References

  1. Moyne, J.; Mashiro, S.; Gross, D. Determining a security roadmap for the microelectronics industry. In Proceedings of the 2018 29th Annual SEMI Advanced Semiconductor Manufacturing Conference (ASMC), Saratoga Springs, NY, USA, 30 April–3 May 2018; pp. 291–294. [Google Scholar] [CrossRef]
  2. Den, W.; Hu, S.C.; Garza, C.M.; Ali Zargar, O. Airborne molecular contamination: Recent developments in the understanding and minimization for advanced semiconductor device manufacturing. ECS J. Solid State Sci. Technol. 2020, 9, 064003. [Google Scholar] [CrossRef]
  3. Yin, J.; Liu, X.; Guan, B.; Ma, Z.; Zhang, T. Performance analysis and energy saving potential of air conditioning system in semiconductor cleanrooms. J. Build. Eng. 2021, 37, 102158. [Google Scholar] [CrossRef]
  4. Hattori, T. Past, present, and future of semiconductor cleaning technology. Solid State Phenom. 2023, 346, 3–7. [Google Scholar] [CrossRef]
  5. Holland, K.; Holland, S. Semiconductor technology roadmaps and lithography implications. In Proceedings of the Advances in Patterning Materials and Processes XLII, San Jose, CA, USA, 22–26 February 2025; Volume 13428, pp. 293–298. [Google Scholar] [CrossRef]
  6. Kim, H.W.; Oh, C.; Jang, H.; Kim, M.Y.; Kim, B.S. Influence of oxygen-related defects on In-Ga-Sn-O semiconductor due to plasma-enhanced atomic layer deposition of Al2O3 for low-temperature thin-film transistor in terms of electrical properties. J. Alloys Compd. 2022, 918, 165649. [Google Scholar] [CrossRef]
  7. Lee, E.L.; Goh, Y.S.; Haseeb, A.; Wong, Y.H.; Sabri, M.F.M.; Low, B.Y. Electrochemical migration in electronic materials: Factors affecting the mechanism and recent strategies for inhibition. J. Electrochem. Soc. 2023, 170, 021505. [Google Scholar] [CrossRef]
  8. Lim, M.J.Z. Life-Time Simulation of Copper Wire Bonds Using Multi-Physics Effect. Master’s Thesis, Nanyang Technological University, Singapore, 2024. [Google Scholar] [CrossRef]
  9. Zhai, W.; Han, Q.; Chen, L.; Shi, X. Explainable AutoML (xAutoML) with adaptive modeling for yield enhancement in semiconductor smart manufacturing. In Proceedings of the 2024 2nd International Conference on Artificial Intelligence and Automation Control (AIAC), Wuhan, China, 24–26 May 2024; pp. 162–171. [Google Scholar] [CrossRef]
  10. Qin, S.J.; Chiang, L.H. Advances and opportunities in machine learning for process data analytics. Comput. Chem. Eng. 2019, 126, 465–473. [Google Scholar] [CrossRef]
  11. Suthar, K.; Shah, D.; Wang, J.; He, Q.P. Next-generation virtual metrology for semiconductor manufacturing: A feature-based framework. Comput. Chem. Eng. 2019, 127, 140–149. [Google Scholar] [CrossRef]
  12. Huang, A.C.; Meng, S.H.; Huang, T.J. A survey on machine and deep learning in semiconductor industry: Methods, opportunities, and challenges. Cluster Comput. 2023, 26, 3437–3472. [Google Scholar] [CrossRef]
  13. Dey, B.; Goswami, D.; Halder, S.; Khalil, K.; Leray, P.; Bayoumi, M.A. Deep learning-based defect classification and detection in SEM images. In Proceedings of the Metrology, Inspection, and Process Control XXXVI, San Jose, CA, USA, 24 April–27 May 2022; p. PC120530Y. [Google Scholar] [CrossRef]
  14. Chen, S.; Zhang, Y.; Hou, X.; Shang, Y.; Yang, P. Wafer map failure pattern recognition based on deep convolutional neural network. Expert Syst. Appl. 2022, 209, 118254. [Google Scholar] [CrossRef]
  15. Ye, J.; Dobson, S.; Zambonelli, F. Lifelong learning in sensor-based human activity recognition. IEEE Pervasive Comput. 2019, 18, 49–58. [Google Scholar] [CrossRef]
  16. Hattori, T.; Heusler, S.; Webb, J.P. Ultraclean Surface Processing of Silicon Wafers: Secrets of VLSI Manufacturing; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar] [CrossRef]
  17. Liu, D.; Chen, H.B.; Wu, J.S.; Then, E. Corrosion behavior of Cu-Al intermetallic compounds in copper wire bonding in chloride-containing accelerated humidity testing. In Proceedings of the 2016 IEEE 66th Electronic Components and Technology Conference (ECTC), Las Vegas, NV, USA, 31 May–3 June 2016; pp. 629–636. [Google Scholar] [CrossRef]
  18. Tian, W. Research progress of laser lithography. J. Phys. Conf. Ser. 2023, 2608, 012016. [Google Scholar] [CrossRef]
  19. Arman Kuzubasoglu, B. Recent studies on the humidity sensor: A mini review. ACS Appl. Electron. Mater. 2022, 4, 4797–4807. [Google Scholar] [CrossRef]
  20. Quang, T.V.; Doan, D.T.; Phuong, N.L.; Yun, G.Y. Data-driven prediction of indoor airflow distribution in naturally ventilated residential buildings using combined CFD simulation and machine learning (ML) approach. J. Build. Phys. 2024, 47, 439–471. [Google Scholar] [CrossRef]
  21. Wang, J.; Xu, C.; Zhang, J.; Zhong, R. Big data analytics for intelligent manufacturing systems: A review. J. Manuf. Syst. 2022, 62, 738–752. [Google Scholar] [CrossRef]
  22. Kim, T.; Behdinan, K. Advances in machine learning and deep learning applications towards wafer map defect recognition and classification: A review. J. Intell. Manuf. 2023, 34, 3215–3247. [Google Scholar] [CrossRef]
  23. Schlosser, T.; Friedrich, M.; Beuth, F.; Kowerko, D. Improving automated visual fault inspection for semiconductor manufacturing using a hybrid multistage system of deep neural networks. J. Intell. Manuf. 2022, 33, 1099–1123. [Google Scholar] [CrossRef]
  24. Jang, S.J.; Kim, J.S.; Kim, T.W.; Lee, H.J.; Ko, S. A wafer map yield prediction based on machine learning for productivity enhancement. IEEE Trans. Semicond. Manuf. 2019, 32, 400–407. [Google Scholar] [CrossRef]
  25. Zhao, Z.; Wang, J.; Tao, Q.; Li, A.; Chen, Y. An unknown wafer surface defect detection approach based on Incremental Learning for reliability analysis. Reliab. Eng. Syst. Saf. 2024, 244, 109966. [Google Scholar] [CrossRef]
  26. Taha, K. Observational and experimental insights into machine learning-based defect classification in wafers. J. Intell. Manuf. 2026, 37, 45–95. [Google Scholar] [CrossRef]
  27. Zhou, W.; Wang, G.X.; Li, Y.F. Application and Development of Machine Learning in Semiconductor Manufacturing for Automated Wafer Map Pattern Recognition and Classification. Integration 2025, 107, 102595. [Google Scholar] [CrossRef]
  28. Mukherjee, S.; Pal, D.; Bhattacharyya, A.; Roy, S. Future of the semiconductor industry. In Handbook of Semiconductors; CRC Press: Boca Raton, FL, USA, 2024; pp. 359–374. [Google Scholar] [CrossRef]
  29. Tin, T.C.; Tan, S.C.; Lee, C.K. Virtual metrology in semiconductor fabrication foundry using deep learning neural networks. IEEE Access 2022, 10, 81960–81973. [Google Scholar] [CrossRef]
  30. Rony, M.A.; Islam, A. Big Data And Engineering Analytics Pipelines for Smart Manufacturing: Enhancing Efficiency, Quality, and Predictive Maintenance. Am. J. Sch. Res. Innov. 2022, 1, 59–85. [Google Scholar] [CrossRef]
  31. Chien, C.F.; Van Nguyen, T.H.; Li, Y.C.; Chen, Y.J. Bayesian decision analysis for optimizing in-line metrology and defect inspection strategy for sustainable semiconductor manufacturing and an empirical study. Comput. Ind. Eng. 2023, 182, 109421. [Google Scholar] [CrossRef]
  32. Field, J. Lifelong learning. In Adult Learning and Education; Schuetze, H.G., Casey, C., Eds.; Elsevier: Oxford, UK, 2011; pp. 20–28. [Google Scholar]
  33. Thrun, S. Lifelong learning algorithms. In Learning to Learn; Thrun, S., Pratt, L., Eds.; Springer: Boston, MA, USA, 1998; pp. 181–209. [Google Scholar] [CrossRef]
  34. Suraj, M.K. Semiconductor Wafer Defect Classification Dataset, Version 1. Available online: https://www.kaggle.com/datasets/meruvakodandasuraj/semiconductor-wafer-defect-classification-dataset (accessed on 26 February 2026).
  35. Steinbach, M.; Tan, P.N. kNN: K-nearest neighbors. In The Top Ten Algorithms in Data Mining; Wu, X., Kumar, V., Eds.; Chapman and Hall/CRC: Boca Raton, FL, USA, 2009; pp. 165–176. [Google Scholar] [CrossRef]
  36. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  37. Cox, D.R. The regression analysis of binary sequences. J. R. Stat. Soc. Ser. B Stat. Methodol. 1958, 20, 215–232. [Google Scholar] [CrossRef]
  38. Tharwat, A.; Gaber, T.; Ibrahim, A.; Hassanien, A.E. Linear discriminant analysis: A detailed tutorial. AI Commun. 2017, 30, 169–190. [Google Scholar] [CrossRef]
  39. Kruse, R.; Mostaghim, S.; Borgelt, C.; Braune, C.; Steinbrecher, M. Multi-layer perceptrons. In Computational Intelligence: A Methodological Introduction; Springer: Cham, Switzerland, 2022; pp. 53–124. [Google Scholar] [CrossRef]
  40. Yacouby, R.; Axman, D. Probabilistic extension of precision, recall, and f1 score for more thorough evaluation of classification models. In Proceedings of the First Workshop on Evaluation and Comparison of NLP Systems, Online, 20 November 2020; pp. 79–91. [Google Scholar] [CrossRef]
  41. Saito, T.; Rehmsmeier, M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The proposed Lifelong Boosting Learning framework. The system evolves linearly over time: at each time step t, a batch-specific model C ( t ) is trained via Boosting and stored. The final detection output is generated by a Time-Weighted Fusion layer that aggregates all historical models C ( 1 ) to C ( t ) , assigning higher importance to more recent models via the decay function W ( i , t ) .
Figure 1. The proposed Lifelong Boosting Learning framework. The system evolves linearly over time: at each time step t, a batch-specific model C ( t ) is trained via Boosting and stored. The final detection output is generated by a Time-Weighted Fusion layer that aggregates all historical models C ( 1 ) to C ( t ) , assigning higher importance to more recent models via the decay function W ( i , t ) .
Algorithms 19 00190 g001
Table 1. Statistical summary of the wafer defect dataset.
Table 1. Statistical summary of the wafer defect dataset.
Dataset NameSamplesFeaturesClasses
Semiconductor Wafer Defect Classification Dataset505082
Table 2. Experimental results under the augmented dataset.
Table 2. Experimental results under the augmented dataset.
AlgorithmPrecisionRecallF1ROC-AUC
KNN0.78420.62500.69570.8412
RF0.89500.74120.81080.9685
LR0.84210.80390.82250.9150
LDA0.81050.68630.74310.8876
MLP0.87340.76470.81540.9520
L 2 Boost0.99120.98040.98580.9945
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yao, F.; Li, T.; Wang, H. Artificial Intelligence Promotes Rapid Detection of Humidity in Semiconductor Manufacturing Environments. Algorithms 2026, 19, 190. https://doi.org/10.3390/a19030190

AMA Style

Yao F, Li T, Wang H. Artificial Intelligence Promotes Rapid Detection of Humidity in Semiconductor Manufacturing Environments. Algorithms. 2026; 19(3):190. https://doi.org/10.3390/a19030190

Chicago/Turabian Style

Yao, Fengting, Tianshuo Li, and Hongjun Wang. 2026. "Artificial Intelligence Promotes Rapid Detection of Humidity in Semiconductor Manufacturing Environments" Algorithms 19, no. 3: 190. https://doi.org/10.3390/a19030190

APA Style

Yao, F., Li, T., & Wang, H. (2026). Artificial Intelligence Promotes Rapid Detection of Humidity in Semiconductor Manufacturing Environments. Algorithms, 19(3), 190. https://doi.org/10.3390/a19030190

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop