Next Article in Journal
Word-Level Motion Learning for Contactless QWERTY Typing with a Single Camera
Previous Article in Journal
A Generative AI–Based Technical Data Extraction Tool for IoT Application Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Explainable Machine Learning for Tower-Radar Monitoring of Wind Turbine Blades: Fine-Grained Blade Recognition Under Changing Operational Conditions

1
Department of Mechanical Engineering, University of Siegen, Paul-Bonatz-Straße 9-11, 57076 Siegen, Germany
2
Verein für Datenwissenschaften (VEFDAWI), 60326 Frankfurt am Main, Germany
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(4), 1083; https://doi.org/10.3390/s26041083
Submission received: 23 December 2025 / Revised: 30 January 2026 / Accepted: 3 February 2026 / Published: 7 February 2026
(This article belongs to the Section Industrial Sensors)

Abstract

This paper evaluates a data-driven classification approach of operational wind turbine blades based on consecutive tower-radar measurements that are each compressed in a two-dimensional slow-time to range representation (radargram). Like many real-world machine learning systems, installed tower-radar systems face some key challenges: (i) transferability to new operational contexts, (ii) impediments due to evolving environmental and operational conditions (EOCs), and (iii) limited explainability of their deep neural decisions. These challenges are addressed here with a set of structured machine learning studies. The unique field data comes from a sensor box equipped with a frequency-modulated continuous wave (FMCW) radar (33.4–36 GHz frequency range). Relevant parts of the radargram that contribute to a decision of the used convolutional neural networks were identified by a class-sensitive visualization technique named GuidedGradCAM (Guided Gradient-weighted Class Activation Mapping). The following main contributions are provided to the field of tower-radar monitoring (TRM) in the context of wind energy applications: (i) every individual rotor blade holds a number of characteristic structural features revealed by the radar sensor, which can be used to discriminate rotor blades from the same turbine via neural networks; (ii) those unique features are not agnostic to changing EOCs; and (iii) pixel-level distortions reveal the necessity of low-level information for a precise rotor blade classification.

Graphical Abstract

1. Introduction

1.1. Motivation for Explainable Machine Learning

Technologies in the realm of machine learning (ML) and artificial intelligence (AI) continue to evolve rapidly and will contribute to progress in various industries. Enhancing predictions and thereby optimizing operations can provide economic advantages as well as socially or environmentally beneficial outcomes, e.g., safety or resilience of energy generation and respective networks, the conservation of biodiversity and ecosystems, as well as climate change mitigation or adaptation. ML, therefore, can substantially serve the public interest. Tower-radar monitoring (TRM) for smart wind turbines represents such an envisioned application portfolio [1,2,3].
In general, power plants supply societies around the world with much-needed energy. High-risk ML applications with respect to such critical infrastructure encompass ML systems intended to be used as safety components in the management and supply of electricity. Here, ML deployment should meet advanced requirements that include transparency of the results or verified reliability against foreseeable deviations from their static training setting. Transparency means that ML gets developed and used in a way that allows appropriate traceability and explainability, as well as laying open the limitations and strengths. The application of those principles should be translated into the design and deployment of ML models, e.g., according to the EU AI Act [4].
In addition to regulatory compliance, explainability nourishes the trust of users and stakeholders in the ML components of intelligent systems. In particular, novel products and services are challenged with gaining acceptance. Explainability enforces productivity, because it remains difficult to identify issues and solutions when models fail, when an understanding about their internal decision-making is lacking [5]. Explainable ML is maturing, and ML gets more and more deployed in real-world applications, and hence a topical issue has become the combination of explainability and robustness analysis.
The concept of systematically perturbing a system or process and observing how sensitively outcomes change is a very general idea in science and technology [6]. In particular, an early contribution to explainability in ML has been the introduction of Local Interpretable Model-Agnostic Explanations (LIME). Here, the fundamental insight [7] lies in the use of local perturbations to assess the reliability of the prediction. In this paper, we will use perturbations, e.g., to study visual explanations, among other things. In our case of computer vision, where only the radar-derived image and no context are available at test time, many distinct concepts and image transformations become closely related or cannot be clearly differentiated anymore. Several of those concepts are shown in Figure 1. They include adversarial examples without visible difference to the original image, common corruption (including blurriness), data augmentation techniques (e.g., contrast or brightness changes), sim2real gap [8] between original and synthetic images, geometric image transformation through environmental and operational conditions (EOCs), and style transfer [9,10] where the content of an image is kept but the appearance changes.

1.2. Perturbations, Model Robustness and Generalization

Adversarial images [13,14] are specifically designed to mislead models while remaining visually indistinguishable from original images to human observers. These images are created by adding carefully calculated perturbations, such as noise patterns, to exploit vulnerabilities in a model’s decision boundaries. In contrast, data augmentations [15] represent intentional modifications of training data to improve model robustness and generalization. They create plausible variations to learn logical invariances. In particular, data augmentation techniques can be divided into two technical categories [16]: One category comprises collective techniques that coherently alter the geometry, shape, or size of the given image (such as stretching); the other category contains pixel-level transforms, which only manipulate the pixels of an image without changing the overall geometry or size of the image (i.e., such transforms can imitate certain alterations in the measurement process of the radar system).
Moreover, common corruptions represent natural-caused image degradations that occur in real-world settings. They include phenomena like motion blur, noise or compression artifacts. Lastly, transformations of two-dimensional images due to environmental or operational conditions (EOCs) occur when a similar physical scene is captured under different conditions or viewing parameters (geometric angles of the wind turbine tower and rotor blade). These transformations again represent natural variations that occur during field deployments.
This theoretical and motivating perspective connects seemingly different concepts. Understanding these connections helps in holistically developing more robust vision systems that can handle the full spectrum of image variations encountered in the real world.

1.3. Related Works and Novelty

The inspection or even monitoring of wind turbine blades can be deployed in many different ways. Typical inspection methods include unmanned aerial vehicles (UAVs) that are remotely controlled to scan the surface of a rotor blade for defects [17]. This approach is accompanied by several challenges and drawbacks compared to ours. First, operating a drone still requires human guidance, which is a costly endeavor when used frequently. Second, this kind of inspection is only possible during downtimes outside the normal operation window of a wind turbine. It is therefore not designed for the continuous monitoring of wind turbines, which requires rapid reactions during service. In addition, it is affected by weather conditions, which further limits the accessibility of a wind turbine and impacts the quality of the measurements.
A recent study on blade surface defect detection has used optical images from a UAV to train a YOLOv10s network for the respective object detection task [18]. Another work applied data fusion on operational wind turbine measurements to improve the segmentation of rotor blades in video data with the help of thermal videos [19]. A general drawback of optical imaging methods is the strong dependency on sunlight exposure. Such monitoring and inspection systems rely on constant ideal lighting because the direction and intensity of the sunlight influence general image quality and can cause distracting shadows or glare.
Radar sensor systems instead offer continuous monitoring of wind turbines much less affected by weather conditions. A recent work performed a fatigue loading experiment on an isolated wind turbine blade with several frequency-modulated continuous wave radar sensors (FMCW) embedded in its fiber mantle [20]. Although directly embedded sensors offer greater potential to detect damage locations with higher accuracy, transferring this setup from the laboratory to real operational wind turbines comes with many technical challenges that have not yet been sufficiently solved. With respect to tower-radar monitoring (TRM), a few previous studies exist. In Moll et al. [21] the feasibility of damage detection with a tower radar in the laboratory was investigated. Moreover, an ML analysis of radar data in the time domain and in the frequency domain was presented separately in another laboratory setting for wind turbine blade materials [22]. Tang et al. [22] successfully distinguished various defect types within a fixed grid of composite cubes using measurements from an FMCW radar with a 24–25.5 GHz band.
Our experimental setup of a tower-mounted radar system that monitors the operational wind turbine blades of the respective tower, to the best of our knowledge, is the first engineering application of its kind.
A detailed description of this setup can be found in Mälzer et al. [11]. In the time of that publication a dataset of 104,000 verified samples made of radargram, label and metadata was collected between 21 October 2022 and 31 March 2023 for one turbine and 59,000 samples for another turbine, collected from 15 December 2022 through 25 February 2023. Mälzer et al. [11] only used one small subset example with 11,500 random samples from the first turbine to validate the general classification potential of the data via a simple convolutional neural network (CNN). This work in comparison includes the complete dataset from the second turbine in Mälzer et al. [11] together with an extended dataset of the first turbine with 276,000 verified samples collected between 21 October 2022 and 23 March 2024. This large dataset was extensively investigated in this paper in terms of transferability between turbines, robustness under unseen environmental and operational conditions and model sensitivity towards physically expected radar signal perturbations accompanied by visual explanations.
Furthermore, in Alipek et al. [23] the ability of a CNN to classify moving rotor blades was discussed when encountering feature drift through changing environmental and operational conditions. It is important to mention that Alipek et al. [23] worked with measurements from a third turbine unrelated to this study, but from the same wind park. The technical setup was equivalent, while two out of three rotor blades of that turbine were differently marked with a microwave absorber foil either facing the radar or the opposite side away from the radar. Alipek et al. [23] analyzed the potential impact of markings on the rotor blade surface by performing a binary anomaly detection task for each individual rotor blade. This task is inherently different from this paper’s efforts that were mentioned before.
An additional context is given by a large-scale study of many common deep learning models on natural images that has investigated how accuracy changes when factors such as the pose of objects are changed [24].
This is analogous to studies focusing on accuracy under common corruptions [25]. Subsequently, the performance under corruptions was studied when various data augmentations are applied [26]. In another study [27], an incremental learning task on synthetic aperture radar data is improved by a LIME-based interpretability metric. Furthermore, the robustness of visual explanations in a medical setting was studied with a focus on the impact of network depth [28]. With respect to adversarial perturbation, the reliability of explainability methods was examined in Galli et al. [29].
Recent work has increasingly addressed interpretability and robustness aspects of learning-based radar signal analysis [30,31,32], as well as the challenges posed by distribution shifts in structural health monitoring scenarios [33,34]. In particular, those studies explored layer-wise relevance propagation on synthetic aperture radar (SAR) images [31], adversarial training of CNNs on micro-Doppler radar data [30] and the integration of explainable AI into CNNs to improve resilience of classifiers [32]. Another recent work includes adversarial training and domain-invariant feature learning with neural networks to manage distribution shifts in nonstationary time series data [33]. Shu et al. [34] trained a ResNet-34 with learning without forgetting, a continual learning method, to improve the damage classification on concrete structures.
Although our preliminary studies [11,23] only covered small datasets or aspects of the task at hand, they also assumed reliable learning behavior based on high classification f1-scores. Since neural networks are prone to learn semantically irrelevant patterns for certain classes, such as background noise, a critical examination of the learning process is required. Therefore, the novelty of this work is to investigate the robustness and decision-making process of CNNs that are trained with radar data to classify operational wind turbine blades.
With our unique set of radar measurements, we provide several novel contributions to the field of TRM in the context of wind energy applications: (i) Every individual rotor blade holds a number of characteristic structural features revealed by the radar sensor, which can be used to distinguish rotor blades from the same wind turbine with neural networks. (ii) Those unique features of a rotor blade are not invariant to changing environmental and operational conditions (EOCs), suggesting the need for robustly designed models. (iii) Pixel-level augmentation methods, such as blur, reveal the necessity of low-level information for a precise rotor blade classification. (iv) Differently structured neural networks have distinct visual explanations. (v) Higher-order reflections (echoes) that occur in some of the radar measurements can be exploited by the model to detect one of the three blades, which harms the robustness of the model if the other rotor blades are still recognized by the first-order reflection. (vi) Customized cropping helps the model focus on the first-order reflection, suggesting that adaptive cropping or background removal techniques are expected to improve robustness.
Section 2 covers the dataset acquisition and properties, the characteristics of the physical experiment and the theory of neural networks and respective visualization methods. Section 3 is a structured series of machine learning studies to investigate various properties of the radar data and the used neural networks. It begins with the transferability of data from one turbine to another turbine, and continues with the model’s robustness across various EOCs, followed by visual explanations for two neural network architectures revealing model decisions. Section 3 ends with a compact hypothesis that explains the physical and theoretical foundation to classify moving wind turbine blades with FMCW radar data. Finally, Section 4 provides the conclusion of this work.

2. Materials and Methods

2.1. Data Acquisition

For data collection, two equally constructed wind turbines of the same wind park were used (Vestas Wind Systems A/S, Aarhus, Denmark). A 33.4 to 36 GHz FMCW radar (IMST GmbH, Kamp-Lintfort, Germany) and a high-speed camera, placed inside a weatherproof box (Figure 2), were mounted outside of each wind turbine tower at 90 m height (Figure 3). Both boxes were positioned with radar and camera in the main wind direction of approximately 230°. Each rotor blade was equipped with a unique barcode foil on the side facing the sensor box (Figure 3). Every rotor blade pass triggered the radar sensor, which stored the measurements from 0.3 s before to 0.5 s after the trigger event, which forms a single radargram. The radar also simultaneously triggered the camera to store a video for the same interval. From each video, a barcode scanner extracted the ground-truth label for the present rotor blade. An additional setup to collect supervisory control and data acquisition (SCADA) data was also used to store wind parameters for monitoring purposes during each rotor blade pass.

2.2. Properties of the Dataset

The current article deals with radargrams collected from 2022 to 2024 on two wind turbines during regular operation. A volume of 276,000 radargrams was measured for the wind turbine “vb06” between 21 October 2022 and 23 March 2024 and 59,000 for the wind turbine “vb07” collected from 15 December 2022 to 25 February 2023. The lower number of examples for “vb07” was caused by an initial delay of the installation and a few technical interruptions that caused shortages in measurement times.
In general, a radargram X with pixel values X i j constitutes a space-time representation of physical events, i.e., a rotor blade passing the wind turbine mast (Figure 4). In such a radargram, the horizontal axis represents the spatial dimension (range), and the vertical axis indicates the temporal dimension (slow-time). More specifically, the horizontal axis constitutes the distance the reflected signal travels back to the sensor. The vertical axis is the moment in time after a ramp was emitted. Each point in the radargram represents the integrated reflection signal over a given range bin at a specific emission time. Each radargram therefore compresses a rotor-blade pass across the sensor field into a two-dimensional image.
The plots in Figure 4 represent the input format of the radargrams used in this work. The choice of measurement intervals was empirical with the intention of obtaining well-centered rotor blade passes. The time interval ranges from 0.3 s before the event trigger to 0.5 s after the event trigger. With a sweep rate of approximately 1000 frequency ramps/s; these 0.8 s equal 795 ramps in each radargram. For the range interval, the range bins 100 to 350 were selected. With a bandwidth of 33.4 to 36 GHz, a range bin has a resolution of 5.77 cm. This precisely corresponds to a coverage between 5.77 m and 20.18 m. The final size of a radargram is then 795 (ramps) × 251 (range bins).
In accordance with the preliminary works [11,23], the use of slow-time range plots was chosen over range-Doppler maps. While both forms of display are highly related [35], since the range-Doppler map is obtained by a fast Fourier transform (FFT) along the slow-time dimension, they serve different purposes. A slow-time range representation preserves time-domain features such as duration, slope and curvature of the object’s motion over multiple scatter points (Figure 4). These reflected signals over the full motion are a compressed form of blade-specific structural information and are crucial for rotor blade discrimination with neural networks (see Section 3.5). Such characteristic temporal information is averaged out in a range-Doppler map, making it less effective for blade classification.
Applying a further FFT along the range dimension would result in a 2D FFT representation that removed explicit spatial locality in both dimensions. Using either of the two discarded representations as input for a CNN makes the interpretation of visual attributions less intuitive. Another common representation of radar signals is the continuous wavelet transform. The continuous wavelet transform provides time-frequency localization by projecting the measured signal onto predefined wavelet functions at multiple scales. While this approach can be beneficial to reveal characteristic oscillatory components that may aid classification [36], it introduces additional design choices that require prior assumptions. Inappropriate adjustments may suppress or distort structural features relevant for classification. The intention was therefore to preserve the physically meaningful slow-time range representation and rely on the CNN’s ability to learn task-specific filters directly from the data. Prior work has shown that deep CNNs trained on time-domain signals can develop localized frequency-selective filters when respective features are relevant for the task [37].

2.2.1. Problem Geometry

When a wind turbine is operating, there are several environmental and operational conditions (EOCs) that can change during its deployment. Some of those directly impact others, which creates correlations between certain pairs of EOCs. With an automatic yaw control system, the wind direction and nacelle orientation are similar but not exactly matching. The combination of pitch angle, wind direction, and wind speed defines the resulting rotation speed for a fixed nacelle orientation. In case of fixed pitch angle, wind direction, and nacelle orientation, the only wind parameter that drives the rotation speed is the wind speed. Based on typical control systems applied in wind turbines [38], an illustration of dependencies is given in Figure 5.
Depending on the current EOCs, a multitude of reflection patterns emerge within the radargrams. A variation is shown in Figure 4. An increased rotor speed leads to a reduced time span of the blade in the illuminated volume, creating a shorter profile along the time axis. Different nacelle orientations imply an altered yaw angle that leads to a tilted reflection pattern. Additionally, a pitch angle introduces a slight shift in distance and a change in the width of the reflection pattern.

2.2.2. Fundamentals of Frequency-Modulated Continuous Waves (FMCW)

In this work, the signal transmitted from the FMCW radar is a continuous wave that varies in frequency over time,
s t ( t ) = cos ( 2 π · f ( t ) · t ) with f ( t ) = f 0 + θ · t and θ = f 1 f 0 τ
where zero phase, unit amplitude, and a linear frequency sweep are assumed, where the latter possesses a rate θ and ends after duration τ at frequency f 1 .
When the transmitted energy encounters an object, it is in part reflected back. The delay in traveling to the object and back reads t d = 2 D c where D designates the distance to the object and c the speed of light. The frequency of the received signal can be expressed as f r ( t ) = f ( t t d ) = f 0 + θ ( t t d ) with the so-called beat frequency given by f b = f ( t ) f r ( t ) = θ · t d . Substituting and rearranging yields the object distance D = c 2 θ f b . Contributions from noise and interference effects were omitted for simplicity. A more detailed mathematical form can be found in [35,39].

2.2.3. Scattering

A radiation of 35 GHz has roughly a wavelength λ 1 cm. The emitted electromagnetic waves of the given frequency partially penetrate the rotor blade, revealing internal structural information [40] compared to optical methods. The exact internal structure will vary between manufacturers and the type of blade [41], however, an illustration of millimeter-scale layering is depicted in Figure 6. The assumption behind the recognition of the three individual blades per turbine is the following: Manufacturing uncertainties (e.g., resin curing or fiber inconsistencies), as well as aging effects (due to transport, assembly, or environmental degradation), will lead to distinct structural fingerprints. They can be learned during training and identified at the test time.
The barcode foil attached to each rotor blade for the optical labeling has no measurable signature in the radar profile, i.e., it had no significant impact on the CNN’s decision. This statement is supported by the results from Alipek et al. [23] where an equally large microwave absorber foil on an identically built wind turbine was shown not to impact CNN classification performance. The microwave absorber foil was also thicker and had better reflective properties. Furthermore, Moll et al. [21] demonstrated the ability of a tower radar sensor with the same frequency level and range to detect small structural defects on rotating objects made with similar fiber composite.

2.3. Convolutional Classifiers

Computer vision represents the realm in which deep learning has historically excelled in numerous tasks. Deep learning triggered a revolution, marked by the 2012 ImageNet challenge [42] where a data-driven approach called convolutional neural network (CNN) outperformed conventional hand-made features. In particular, the introduction of ResNet’s skip connections [43] allowed users to train increasingly deeper networks with respective improved performance.
In this paper, rotor blade recognition is treated, which represents a typical computer vision classification problem. Here, at test time, an ML model ϕ is fed with a radargram X to predict the correct class among the K = 3 classes indexed by m. Its prediction consists of the class with the highest score ϕ ( X ) = argmax m z m . The layered, multi-channel CNN architecture specifically entails activation maps A i j k l . Here, k denotes the channel of layer l and pixels in the x- and y-direction are indicated by i , j .
In order to obtain the trained network ϕ via deep learning, a cross-entropy loss L q is minimized over ground-truth labels y m { 0 , 1 } for the input radargram X q with z p as its raw logits in the softmax function σ :
L q = p = 1 K y p · σ ( z p ) with σ ( z p ) = e z p m = 1 K e z m
In this work, a short custom CNN as well as a ResNet-18 are utilized. The custom CNN has a feature extractor containing three convolutional layers with an intermediate rectified linear activation function (ReLU) and one max pooling operation (MaxPool), followed by two fully connected layers for the classification head. All three convolutional layers have a kernel size of 3 × 3 and stride and padding of 1. The max pooling operation uses a 2 × 2 kernel with a stride of 2. The ResNet-18 architecture was implemented via the Huggingface timm library [44] (version 1.0.9). The utilized model is found under the name “resnet18.tv_in1k”. Its first module holds a convolutional layer with a kernel size of 7 × 7, a stride of 2, and padding of 3, followed by BatchNorm, ReLU, and MaxPool. The remaining modules have convolutions with a kernel size 3 × 3 and a stride and padding of 1 or 2. For training, the ADAM optimizer was deployed with a learning rate of 0.0001 and batch size 16. Smaller batch sizes such as this were shown to improve model generalization and to enable stable training across a wide range of learning rates [45]. The only architectural modification was to define a customized input configuration function inside of the timm library that changed the following parameters: number of classes from 1000 to 3, input size from [3, 224, 224] to [1, 795, 251], and the cropping percentage from 0.875 to 1, because the cropping was applied outside of this library. The ResNet-18 model was, with one exception, also trained from scratch and did not use the pretrained weights from the ImageNet-1k dataset. A larger ablation study with pretrained models is not included in order to keep the analysis focused on the intrinsic properties of the proposed dataset and to ensure interpretability of the presented results. Since variations in pretraining datasets may influence model behavior differently, the primary goal is to assess the direct potential and limitations of the radargram dataset for the given classification task. Secondarily, the model-related motivation is to compare a small, simple CNN with a second CNN that has a significantly larger receptive field and several more nonlinearities. While the 7 × 7 convolution in the first layer is inefficient by modern architectural standards, it does not affect the outcome of the conducted studies. The core architecture of the ResNet-18 is maintained for convenient usage of the timm library and comparability.

2.4. Visual Explanations

In an effort to make models more transparent by highlighting relevant characteristics of the input data that contribute to predictions, visual explanations encompass techniques (such as DeconvNets [46]) to provide such insights into the decision-making process of ML models in computer vision. Other visual explanation models followed that do not require architectural change compared to DeconvNets.
One similar method, called Guided Backpropagation [47], produces very sharp, pixel-level saliency maps that highlight detailed edges and textures. While this algorithm is simple and fast, it is not considered a true attribution method based on the axiomatic attribution theory [48], which may lead to partially incomplete attributions in theory. Another popular visual explanation model in computer vision, named GradCAM [49], satisfies the expectations of such a true attribution method. GradCAM by itself provides relevant class-discriminative spatial maps, but which are only coarse, and of low resolution.
The element-wise multiplication of both methods, Guided Backpropagation and GradCAM, leads to high-resolution, class-specific saliency maps and is known as GuidedGradCAM [49,50]. The formula X GuidedGradCAM = X GuidedBackprop X GradCAM combines the advantageous properties of fast and simple high-resolution activations from Guided Backpropagation with the class-discriminative attributions from GradCAM, while mitigating the potential inconsistency of Guided Backpropagation. GradCAM works as follows:
  • Identify the target layer (usually the final convolutions l = L ), which should both capture high-level features while retaining spatial information,
  • compute the gradients of the score z for the target class with respect to the k-th feature map of layer L,
  • average gradients to obtain importance of each feature map α k i j z A i j k L (where the normalization constant is omitted for simplicity),
  • weight feature maps to create a coarse localization highlighting influential regions X i j GradCAM = ReLU ( k ( α k · A i j k L ) ) = max k ( α k · A i j k L ) , 0
  • upsample heatmap for overlay to original radargram or combination with results from Guided Backpropagation.
Here, Guided Backpropagation modifies the standard backpropagation algorithm in neural networks to allow only a positive forward flow and backward flow of gradients, which helps to highlight the features that are most relevant to the model’s predictions. This is performed by setting the negative contributions to zero [47]:
max A L A l , 0 · max A l , 0
Although the traditional variant of GradCAM uses a ReLU to remove negative contributions, turning off the final ReLU also enables visualization of negative attributions. In this way, GuidedGradCAM can highlight both the positive and negative impact of features in the image that only contribute to the decision of a chosen class. GuidedGradCAM reveals which patterns in an input image have contributed to the models’ decision in favor of or against each specific class. Hence, each class (i.e., individual rotor blade) can be evaluated separately.
There is a notable limitation to the combination of Guided Backpropagation and GradCAM. Guided Backpropagation depends on the architecture using ReLU as its nonlinear activation function. Other activation functions, such as sigmoid or tanh, are not compatible with this algorithm [47]. Nevertheless, this limitation does not significantly harm the practical applicability of Guided Backpropagation and therefore GuidedGradCAM, because the ReLU is theoretically well-suited and empirically the most widely used activation function for CNNs [51,52].

3. Results and Discussion

In principle, the aforementioned field of tower-radar monitoring (TRM) encompasses several diverse machine learning tasks. In particular, the following section focuses on the fine-grained task of recognizing the specifics of individual rotor blades.
Hereafter, radargram data for all three rotor blades (ROT1 to ROT3) of wind turbines will be considered where studies are conducted by varying important influencing factors. Different power plant sites (i.e., comparable turbines but at adjacent locations dubbed “vb06” and “vb07”) are treated as well as distinct EOCs (rotor speed, nacelle rotation, and pitch angles) and distinct data augmentations of the radargrams. In total, two different machine learning classifiers (custom CNN and ResNet-18) are studied, and two different kinds of results are discussed: quantitative classification performance (as measured by F1 scores) as well as qualitative evaluation of visual explanations (via GuidedGradCAM).

3.1. Classifier Robustness Across Turbine Locations

In this subsection, the quantitative classification performance is studied for three scenarios (i.e., static, ad hoc transfer and fine-tuning) where data is available from two equally built turbines, which are located only 300 m apart. Such population-based settings represent a common case in modern structural monitoring applications [53]. To fairly compare the varying performance of the model for the three scenarios, the same partitions were used for each task. A stable, simultaneous measurement period between 15 December 2022 and 25 February 2023 provided either 54,000 or 59,000 rotor blade passes for a turbine and either 18,000 or 19,600 measurements per rotor blade.

3.1.1. Identically Distributed Train and Test Data

A first step in the behavioral and performance analysis of a classification model represents the traditional procedure, with model training and testing performed on two random partitions of the same static dataset. Here, the resulting random train and test splits are therefore identically distributed. This first study evaluates the model’s ability to handle unknown samples from the same feature distribution (i.e., same turbine, same EOCs). For this, the custom CNN (see Section 2.3) was trained and tested separately on both turbine datasets, vb06 and vb07 (see the results in column 1 and column 2 in Table 1).
The performance of the model on turbine vb06 becomes close to a perfect fit, while the data for vb07 seem more difficult to learn. A closer look at the distributions of vb06 and vb07 reveals that rotation speeds below 5 rpm (rotations per minute) are below 40 samples per class (i.e., rare cases). Almost all rotation speeds between 5 and 14 rpm are below 50 samples per class for vb06. Only 6.2–6.4 rpm and 12.2 to 13.8 rpm are above 700 samples per class for vb06. In comparison, vb07 has more than 700 samples per class for 5.6–5.8 rpm and 11.3 to 12.2 rpm. Therefore, the peaks in rotation speed are slightly shifted between both turbines. The other main difference is that vb07 has three times more measurements for underrepresented rotation speeds between 5 and 14 rpm i.e., around 130 samples per class.
With each measurement being a combination of EOCs with many degrees of freedom, there is a large bandwidth of radargram variations for a single rotation speed value. The assumption is that the model fails to correctly learn the underrepresented radargrams for both datasets. Only in the case of vb07 the number of (still) underrepresented examples contributes to more misclassifications in total, which leads to a reduced F1 score.
The complete dataset for vb06 with over 276,000 samples, is much more evenly distributed and contains at least 700 samples per class for all rotation speeds between 5 and 14 rpm. The small custom CNN trained with this dataset reached class-based F1 scores of about 0.95 and proved a solid discriminative ability toward the classification of each rotor blade. Based on this initial study, machine learning experiments beyond Section 3.1 will solely use data from vb06.

3.1.2. Re-Using Trained Classifiers for Other Similar Turbine

Practical scenarios commonly call for reusing a trained classifier in further comparable power plants. Next, a simple transfer learning check was performed by evaluating each of the two models on the test data from the other turbine. Since the core assumption of TRM rotor classifications consists of the uniqueness of each blade (compare Section 2.2.3), low accuracy is foreseen for this ad hoc check. However, the learned representation might be successfully reused with some additional effort (i.e., fine-tuning). For the sake of consistency, the blades in vb06 and vb07 are both denoted as ROT1 to ROT3.
Consequently, column 3 and column 4 in Table 1 display the expected poor performance of that ad hoc check. For comparison, random guessing (or “pure chance”) for three classes in a balanced dataset would lead to a class-based F1 score of 0.333. However, with one class (ROT1) at 0.193 in column 3, and one class (ROT2) at 0.452, there is one result below random guessing and one above. A similar pattern is found for column 4. Such results below random chance can be described as “systematic mistaking” and imply that the model is not only guessing wrong but actively and consistently biasing its errors towards incorrect classes.
In general, this observation of poor transfer performance represents another indicator (in addition to Section 3.1.1) that each rotor blade appears to be individual (or even unique) at a level that is measurable by the radar system. Moreover, the confusion matrix given in Figure 7 provides a closer look at the results in column 3 of Table 1. It gives an overview of the predictions when the model trained with vb06 data is evaluated on the vb07 test dataset. A clear trend is visible when comparing the middle column with the other two. Most samples from vb07 are predicted as class ROT2 of vb06. This means that most samples from vb07 have a greater similarity to ROT2 from vb06 than to ROT1 and ROT3. Furthermore, the predictions of the middle column are almost evenly distributed between the three rotors of vb07. This indicates that all three rotor blades of turbine vb07 have a statistically equal similarity to the rotor blade ROT2 of vb06 (see Figure 8 for an illustrative depiction).
Thus, a key question arises: On which learned features does the classifier base its decisions? Hence, in the later parts of this paper, visual explanations will be investigated.
Apart from the aforementioned ad hoc check, a more realistic and widely used approach employs fine-tuning. This fine-tuning approach was performed in two variations. The first variant was extended into a two-stage joint training, where the model is optimized first with one turbine and then with the complete training partition of the second turbine i.e., 60% of the total dataset (Table 1, columns 5 & 6). The second run represents the conventional way, where a small fine-tuning dataset is used, which is 10% of the complete dataset of the second turbine (Table 1, columns 7 & 8).
Regarding the first fine-tuning scenarios, the model that was trained with both turbine training datasets achieves test performances close to the model without pretraining (compare column 1 with 6 or column 2 with 5 in Table 1). However, if the model is pretrained in this way on the training partition of one turbine and subsequently fine-tuned on the complete training partition of the other turbine until saturation, a residual bias toward features of the pretraining turbine remains. These features can have similarities to all classes of the second turbine, which explains the measurable reduction of the maximum F1 score for all three rotor blades. If the model otherwise managed to learn the three rotor blades of each turbine having a high but reduced baseline performance for both turbine datasets, this reduction is still an indicator of increased uncertainty between the three nominal classes, since pairs of different rotor blades are mapped into single classes.
To further analyze this behavior, a conventional fine-tuning run was conducted. When comparing column 7 with 5 or columns 8 with 6 in Table 1, the test performance was significantly lower if the model was fine-tuned with only 10% of the complete dataset instead of the previous 60%. These expected results can plausibly be explained by a larger number of underrepresented EOC configurations and smaller average counts of examples for those underrepresented cases resulting from the smaller fine-tuning dataset.
Since rotor blades can be very unique in the fine-granulated scale of the fiber mantle and the inner blade skeleton, and the radar is able to partially penetrate the material to also measure subjacent structural information [21,40], the radar profile of each rotor blade is inherently different from the others. The presence of structural differences between equally manufactured rotor blades is further supported by measurable mass imbalances within individual blades and mass differences between the rotor blades of the same turbine [54,55]. Therefore, each rotor blade must be treated as another object in the scope of this work (more in Section 3.5).
The observed “systematic mistaking” in the ad-hoc transfer learning study indicates that features learned from one turbine are only partially present in the data of another turbine (Table 1, columns 3 & 4). In particular, the data augmentation study (Section 3.3.4) shows that correct blade discrimination relies on small, persistent scatter patterns located within the crescent-shaped structure of the radargram. Since these low-level features differ between turbines (Section 3.5), the classifier tends to rely on larger and more generic structures. This leads to consistent confusion between rotor blades, as observed in the ad-hoc transfer case. The extreme case is observed when a model is pretrained on one turbine and subsequently fine-tuned with the complete training partition of another turbine. As recently mentioned, it further shows that residual features from the pretraining turbine can introduce bias and additional uncertainty, resulting in reduced maximum F1 scores (Table 1, columns 5 & 6). Therefore, the conducted studies do not indicate the presence of useful transferable representations across different turbines.

3.2. Classifier Robustness Across EOCs

In Section 3.1, it was already indicated that each rotor blade can cause individual reflection profiles for the same radar sensor, making each rotor blade distinguishable from other rotor blades of the same wind turbine. A follow-up question would be whether such characteristic features of a specific rotor blade are also invariant to changing EOCs, which is investigated in this subsection with respect to the measurements of one wind turbine, namely vb06.
The focus is on the three wind parameters: rotation speed, nacelle orientation, and pitch angle. Each wind parameter study was conducted in the same way. The complete dataset of vb06 was divided into three partitions with respect to the value of the wind parameter. One partition contains only samples with lower values, another partition has higher values, and the third partition contains measurements with intermediate values between the limits of the first and second partitions. The custom intervals of the different parameters are carefully selected by hand to minimize the impact of their correlation with the inspected parameter. Each partition initially stored the samples sorted by their timestamp. After a deterministic shuffle along the timestamps via numpy random seed, the partition was further split by a 60:20:20 ratio into training, validation, and testing data. Therefore, all three sub-partitions have matching EOCs and a similar feature distribution.
There are other validation methods that closer match real-world scenarios, such as rolling time windows or timely holdout. However, the variation in wind parameters between consecutive months, weeks, or even days can be quite strong, which makes it systematically difficult to determine the contribution of individual wind conditions to the change in classification performance. By previously inspecting the distribution of EOCs and intentionally selecting intervals to minimize correlation, it granted much more control over the studies conducted and facilitated interpretation.

3.2.1. Rotation Speed

The classification results for the rotation speed are given in Table 2. Column 1 shows the model trained and tested with data from an interval with slower rotation speeds (3500+ training samples per class). Column 2 is set up similarly, but for high rotation speeds (2500+ training samples per class). Both columns show high performance for new data from the same feature distribution, i.e., same EOCs. The fifth column shows the performance of the model when trained on both intervals and tested on unseen samples from both trained intervals. The sampling ratio here is balanced for both intervals. This scenario shows a solid fitting result on the data that yields a high F1 score.
When comparing columns 1 and 2 of Table 2, the performance for the “high” partition is the poorest. Here, the high rotation speed leads to a relatively small crescent in radargrams. The correct recognition of small and tiny objects remains a well-known challenge in computer vision and has become a dedicated research subfield [56,57]. Small objects (i) are more prone to perturbations through noise/clutter because of their small visible area, and (ii) make it harder to learn a rich set of meaningful features due to the lack of a nuanced shape or a missing fine texture. Furthermore, (iii) many deep learning models are sensitive to the scale of objects and therefore lack the ability to recognize in small objects their known larger counterparts.
Accordingly, column 3 of Table 2 lists the performance of the CNN that is trained with data from slower rotations and tested on data with higher rotation speed for the same wind turbine vb06. Column 4 is the opposite composition. Both cases of systematic mistaking show that extrapolating to radar measurements of the same wind turbine but for unseen EOCs is not attainable.
Possibly, the classifier can learn diverse features (as proven in previous studies here), but these features are not correlated in a meaningful sense with the appearance of a smaller crescent. Again, the demand arises to understand which learned features the CNN draws on for its predictions. Here, the visual explanations in the later part of the article can help.
Column 6 on the right side of Table 2 represents the interpolation test. Here, interpolation does not refer to the phenomena heavily discussed with regard to overfitting, e.g., in Montanari and Zhong, and Mallinar et al. [58,59]. In the current paper, interpolation means that both intervals, one with slow rotation speeds and one with high rotation speeds, are ingested for training by the same CNN but then the model was tested on samples with rotation speeds between both intervals (900+ testing samples per class). It also differs from mathematical interpolation, since the model is applied to existing radargrams of unknown EOCs and is not attempting to reconstruct missing EOCs or corresponding radargrams. The interpolation results naturally show significantly better performance compared to the extrapolation, but still lack the precision of the regular static setup shown in column 1 and column 2 of Table 2. Results regarding the ResNet-18 are addressed in Section 3.3.2.
To assess the robustness of the reported results, multiple reruns were performed for selected runs summarized in Table 2, including a high-performance case (column 5), an interpolation case (column 6), and a case of systematic misclassification (column 3). Each case was repeated three times using different random seeds for data partitioning and model initialization. The resulting standard deviations ranged from 0.2–0.6% for column 5, approximately 2% for column 6, and 0.4–4.8% for column 3. Similar deviations were found for equivalent runs using ResNet-18. As expected, increasing discrepancies between training and test distributions led to higher variability. However, the magnitude of the observed standard deviations is too small to affect the conclusions of this study. Consequently, averaged results are only reported when statistical exactness is required.
The inference time of the model on an Nvidia RTX 4090 (Nvidia Corporation, Santa Clara, CA, USA) is below 0.5 s per input, and a new radargram is measured and stored every 0.8 s. Sufficient inference time is attainable on edge devices. Thus, real-time processing is technically feasible but not recommended at the current stage, as this design is not yet robust to unknown EOCs. Overall, the above results show that radar measurements of different rotor blades hold some type of characteristic features that define each rotor blade from one another, but those features are partially mutable through changes in EOCs, limiting interpolation and extrapolation. According to these findings, a model for a wind turbine cannot be trained on a fixed dataset, which is collected in a short amount of time, to be robust for all upcoming events. The difficulty is within the high number of degrees of freedom that can alter the characteristic features of each rotor blade under different EOCs. Furthermore, since EOCs such as wind speed and wind direction cannot be controlled and such wind dynamics are not predictable, it makes it highly difficult to collect respective data that covers all potential EOCs of that wind turbine (see Table 3).
At least three possible solutions exist: they either consist of continual learning, where the classifier should be re-trained on recent data that evolves with the altering EOCs, in order to steadily minimize the interpolation or extrapolation gap. Secondly, dedicated augmentations or synthetic data could mimic potential different EOCs, even if they have not yet occurred naturally (see Section 3.3.3). Alternatively, larger and more complex networks [25] are known to generalize better (see Section 3.3.2), but cannot compensate for missing features from unknown EOCs. The impact of changing environmental factors or correlated variables was further discussed in a recent publication with respect to simulated wind turbine data [60].
The hypothesis in Section 3.5, which unifies this work’s results, indicates that latent structural features based on surface and subsurface reflections are characterizing individual rotor blades and enable a classification of those during operation. Aging effects and defects are another form of structural change on and within the rotor blade and are expected to be similarly revealed via radar signal reflections. Those structural alterations happen gradually and subtly change the characteristic features of a rotor blade over time. Such alterations will presumably be unknown to the model, but can be similarly addressed by the previously mentioned continual learning methodology.

3.2.2. Nacelle Orientation and Pitch Angle

The following Table 4 and Table 5 results are obtained by the same procedures previously explained and also lead to similar partition sizes and findings. To save space and avoid redundancy, the respective tables are provided with short remarks about anything different from the first example regarding the rotation speed. Details about the sampled EOCs are provided in Table 6 and Table 7, respectively.

3.3. Visual Explanations of Radargrams

The above results indicate the existence of characteristic features for each rotor blade, enabling discrimination between multiple rotor blades. To make more specific statements and build trust in the model, GuidedGradCAM will be used to visualize the dominant features that a CNN has learned to correctly classify three rotor blades of the same wind turbine. For the inference stage during GuidedGradCAM, the batch size was set to 1 in order to compute visual explanations for individual radargrams.

3.3.1. Custom CNN on Original Radargrams

First, GuidedGradCAM attribution evaluations were performed on the last convolution layer of the custom CNN. The attributions were upsampled to match the input size (795 × 251 pixels).
Figure 9 treats an exemplary radargram of ROT1, which is correctly predicted by the custom CNN: visual explanations for class 1 (correct), classes 2 and 3 (both wrong) are depicted. Since mainly the middle section of the crescent in all three visualizations is activated, the more dominant first reflections on the rotor blade surface seem to be of interest to the model for its decision. Here, positive and negative contributions (i.e., for and against a given class) are located at the core reflection site.
Figure 10 covers an illustrative radargram of ROT2, which is accurately classified by the custom CNN: visual explanations for classes 1 (wrong), 2 (correct), and 3 (wrong) are shown. Results still exhibit similar attributions for class 1 and class 2, but a different pattern for class 3: apart from positive (green) contributions at the central part, a significant number of peripheral negative (red) pixels along the crescent are visible.
All investigations had been carried out with three datasets holding measurements with low rotation speeds, high rotation speeds, or low and high rotation speeds, and led to very similar results.

3.3.2. ResNet-18 on Original Radargrams

The above ablation study with the custom CNN showed mainly small, spatially concentrated visual explanations. Keeping the shallowness and size of the custom model in mind, such behavior was likely caused by the small field of view of the final convolutional layers. Therefore, a second deeper model (ResNet-18) was selected to also test a widely known/investigated model architecture. Additionally, the observation that green attributions for each class are observed on the main reflection front, even though the model has correctly classified the radargram could hint towards a weakness of the custom CNN. If reinforcing patterns of all three classes are very closely located on the radargram, and the radargram happens to be slightly distorted (caused by changing EOCs), one may expect the discriminative ability to rapidly decline.
Thus, the last three columns in Table 2 show a robustness comparison regarding the interpolation capacity of the custom CNN and the ResNet-18. Here, the decline in classification performance in the interpolation task is more significant for the custom CNN (as expected). With the ResNet-18 considered the more capable architecture, an additional study using pretrained weights from the ImageNet-1k dataset was conducted to further assess the limitations of the dataset and task. The experiment trained on low and high rotation speeds was first repeated three times with different random seeds with randomly initialized weights. The same configuration and partitions were then used for the same ResNet-18 architecture initialized with pretrained ImageNet-1k weights. The pretrained variant reached training saturation approximately three times faster with a 1–2% higher baseline F1 score when evaluated on test data drawn from the same EOC distribution. This result is notable given that ImageNet-1k consists of natural images and does not contain sensor-specific representations such as radargrams. A common explanation in the transfer-learning literature is that early convolutional layers learn generic feature detectors (i.e., edges, simple geometric structures, and textures) [46] that are transferable across domains [61,62] and provide a more favorable initialization for training convergence [63].
Interpolation performance improved when using the pretrained ResNet-18, as shown in the added column 8 (compared to column 7) in Table 2, indicating increased robustness under previously unseen EOCs. This behavior is consistent with the comparison between the custom CNN and the ResNet-18 trained solely on radargram data, where the more expressive architecture showed better generalization. Overall, the results suggest that more effective feature representations improve robustness and generalization, even when the pretraining domain differs substantially from the target domain.
To provide a clearer picture of why ResNet-18 seems more robust, the model’s class attributions were also computed via GuidedGradCAM. Figure 11 generally exhibits activation patterns that are broader than in the case of the custom CNN. Similar observations were found in Selvaraju et al. [49] where a ResNet yields more extended visual explanations than the Visual Geometry Group (VGG) architecture [64]. In Figure 11 the wrong classes (i.e., 2 and 3) interestingly lead to explanations far off the crescent where only noise resides.
When a radargram contains a second, smaller sickle-shaped reflection on the right-hand side of the first reflection, as in Figure 12, it can influence the decision-making of the model (see wrong class 3). This smaller reflection represents the echo of the radar sensor. It occurs when the first signal sent from the radar is partly reflected by the sensor and reaches the reflecting object a second time. It therefore is of reduced intensity and has traveled twice as far, which explains its location at double the distance.
Analogous behavior can be observed in Figure 13, where the decision on the correct class is based on this second-order effect.

3.3.3. Custom CNN and ResNet-18 on Cropped Radargrams

In the next step, the impact of random cropping was investigated, since models were shown to focus on distant noise patterns (Section 3.3.2.) From the original radargrams (size 795 × 251 pixels), the center section (size 450 × 251 pixels) was first extracted to remove most parts without the actual event. Then, random crops (size 224 × 224) were extracted for model training. The model was trained from scratch twice with the exact same dataset, either by combining cropped or raw radargrams of low and high rotation speeds (Table 8). Random cropping seems to reduce the training and testing performance of the ResNet-18. This means that random cropping makes it more difficult for the model to classify the three rotor blades for data with the same feature distribution. Either previous performance was based on spurious features (i.e., shortcut learning), which cropping has taken away. Alternatively, cropping has been too aggressive and eliminated relevant parts of the crescent.
Interestingly, the interpolation test revealed additional helpful insights. As given in Table 8, random cropping improved the overall ability of the model to correctly classify the rotor blades when it was tested on radargrams with unknown rotation speeds between the intervals used for training. In particular, the identification of class 2 (ROT2) improved significantly using random cropping. This overall increase in the F1 scores may indicate that random cropping supports the model in having a better general understanding of the individual rotor blade reflection and in ignoring (a) local noise patterns around the main reflection front and (b) general background noise before and after the actual event during the complete measurement. This would also explain why the training time until the optimization was saturated was longer for random cropping. This was probably because each radargram was randomly cropped in a different location, making it harder for the model to learn the relevant patterns of each class.
To verify these assumptions, GuidedGradCAM was also applied to the ResNet-18 model trained with randomly cropped radargrams. Figure 14 demonstrates the usual attributions found for random cropping in this model. The training and testing dataset for this cropping example only used radargrams with low rotation speed. The plotted attributions are for class 3 (ROT3) in case of a correct classification of class 3. Compared to correct predictions with no random cropping, the attributions found have a very compact rectangular shape and capture a limited area centered around the main reflection front. This indicates that the model was better conditioned to comprehend the reflection information of the rotor blades themselves rather than the surrounding background noise.
A less expected example occurred for a ResNet-18 trained with radargrams of low and high rotation speed using the same cropping as in the previous test. In Figure 15, the attributions of class 2 for a correct prediction of the same class exhibit a pattern that contributes to the correct class decision, which is not part of the two implied reflection fronts. This means that the model has learned areas of unwanted background noise to correctly classify the rotor blades. This type of unsuitable learning behavior is most likely caused by too extreme random cropping. With the dimensions of the cropping kept unchanged and added radargrams of high rotation speed having shorter reflections in the vertical, since they move faster through the sensor field, the ResNet-18 was exposed to several examples where the crops do not contain sufficient parts of the actual rotor blade reflection. This must have forced the model to extract meaning from the remaining parts of the images, i.e., the background noise.
Considering the previous observations, random cropping shows potential, but only when applied correctly and fit for the data. This means that one needs to know the dimensions of the expected objects of interest within the images. Since an operational wind turbine has multiple ways to calibrate to wind speed and direction, by nacelle orientation and pitch angle, affecting rotation speed, angle, and distance to the sensor, there are many possible different radargrams. One must also consider that some architectures are not capable of handling multiple resolutions or performing on resolutions other than the training resolution. Models of the latter kind cannot predict new radargrams with the original shape when they were trained on cropped data. Although the approach enhanced the model’s focus on rotor blade reflections and improved interpolation performance, the maximum training and test accuracy slightly declined. This decrease is attributed to the combination of centered and random cropping, which prevented the model from memorizing systematic background noise patterns that otherwise facilitated radargram classification. Traditional random cropping can introduce noise that harms classifier performance as demonstrated in Figure 15, because misplaced crops misalign with regions of interest (ROI), i.e., crescents representing rotor blade reflections. Previous work showed that masking or guiding the cropping procedure can reduce such noise and improve the accuracy in CNN-based classifiers or object detectors [65]. An overview of popular adaptive ROI extraction methods, such as learned region proposals or instance segmentation models that perform ROI cropping and alignment, can be found in Liu et al. [66]. One article specifically focused on a novel ROI extraction layer in CNNs that explicitly learns where to crop based on the image content [67]. While these approaches appear as promising improvements over random cropping, an extensive and systematic study of such learning procedures to maximize model performance is beyond the scope of this work.

3.3.4. ResNet-18 on Augmented Radargrams

Finally, another way to assess model robustness is to evaluate it on perturbed test data (after training on the original radargrams). The evaluated methods are Gaussian blur, defocus, and downscale and were used from the albumentations library [68] (version 1.4.15) for PyTorch (version 2.1.0 with CUDA version 12.1). The degree of the applied augmentations is chosen to be rather severe to create a more noticeable change in attributions (Figure 16).
The first augmentation technique is Gaussian blur, which applies a Gaussian filter with a random kernel size and a sigma value as the standard deviation of the kernel. By reducing image noise and detail, it creates a smoothing effect. Subfigure (b) shows a change in attributes, with less attention to the middle section of the main reflection front and even less attention to the immediate surrounding area. The class-based test F1 scores on the augmented version of the test data samples are 0.174, 0.509, and 0.0, respectively. This indicates that with all low-level patterns smoothed out, and therefore removed, the model fails to correctly detect each rotor blade. Performance fell below random guessing, with class 2 systematically misclassified as either class 1 or 3.
The defocus method also uses a Gaussian blur technique, but with disc kernels, creating a camera-like smoothing effect. The blur effect in Subfigure (c) is even stronger than the one in Subfigure (b), creating different attributions for the same image and observed class. The computed F1 scores are 0.0, 0.497, and 0.0. displaying a systematic confusion of the model with a 100% bias towards class 2.
The third listed augmentation technique downscales an image to a lower resolution. The downscaled image is then upscaled again, but without recovering the lost information from the downscaling. The image then represents a more pixelated version. The downscaling and upscaling procedure reveals some pixels with failed measurements in Subfigure (d) that have a level of intensity below the background noise, given by the few blue pixels. Similarly to the blurring methods shown previously, the F1 score fell significantly to values of 0.057, 0.255, and 0.491. This method is another variant to simulate a loss of detail. It becomes a regular radargram with lower resolution imitating a radar system with different settings, i.e., wider range bins and lower frequency, which leads to fewer ramps for the same time window and fewer range bins for the same distance interval. Although the attributions are shaped and located similarly to the baseline version (Subfigure (a)), the reduced performance indicates that the model is primed to the radargram settings of the original training data.
The stronger the pixel-level augmentation parameters were set, the worse the overall classification performance of the model, indicating the main dependency of low-level features for a successful rotor blade discrimination. Here, high-level features refer to the shape or orientation of the crescent, and low-level features encompass texture. As the attributions for all tested data augmentation techniques have shown, the main reflection front is still partially detected by the model, but it fails to recover characteristic low-level patterns from the pixel-level manipulated images, which are necessary for the correct classification.
Apparently, some image augmentations, such as downscaling, have an intuitive correspondence to a change in radar configuration. However, a reliable correspondence between radar hardware degradation (e.g., antenna, transmitter/receiver chain, ADC, …) and pixel-level modification in the resulting radargrams is non-trivial. The coupled signal processing within modern radar sensors is too complex, and mapping their degradations to image augmentations and related intensities would be speculative. Studies conducted in this section were intended to perform a stress test on the model’s behavior and the impact on visual explanations, not to model specific physical failure modes.

3.4. Impact of Weather Conditions on Radargrams

While wind conditions have a direct impact on the problem geometry, i.e., the movement of the rotor blade relative to the sensor (Section 2.2.1), the weather can also have a systemic influence on the overall radargram. Increased system temperature may occur during hot summer months in Europe and potentially increase thermal noise in the sensor. Thermal noise in the radar receiver was shown to negatively affect the accuracy of range and velocity measurements [69]. Another relevant condition is the impact of rain. In laboratory and field measurements on rain detection [70], the atmospheric attenuation of mild rain was negligible, while heavy rain was expected to have a measurable impact on radar data. Water molecules in the form of raindrops have a higher absorption and scattering of electromagnetic waves at typical microwave and radar frequencies than dry air. However, both conditions are rather critical in applications with small objects tracked at great distances and less problematic in this experimental setup with a single large rotor blade at comparatively short distances (<15 m). Thermal noise increases the overall noise floor, and rain causing attenuation reduces signal strength. Thus, both effects degrade the signal-to-noise ratio (SNR), but they are systemic and statistical, i.e., they affect all measurements similarly. Additionally, CNNs trained on radargrams focus on spatial patterns and relative contrast between pixels, which implicitly compensates for moderate changes in the SNR.

3.5. “Structural Fingerprint” Hypothesis

The core hypothesis of this work is that each rotor blade of a wind turbine exhibits persistent characteristic structural differences arising from manufacturing uncertainties and aging effects. Manufacturing variability in particular, is well known for rotor blade production and is largely related to heavy, repetitive manual labor and limited process control [55]. Such variability manifests, among other aspects, in measurable mass imbalances both within individual rotor blades and between the blades of the same turbine [54]. Since radar waves partially penetrate glass fiber composites and are sensitive to their surface and subsurface structural properties, radar-based measurements provide a plausible physical basis for capturing these differences in wind turbine blades [21,40].
Based on the conducted studies, this hypothesis needs to be critically evaluated. A typical radargram, as shown in Figure 4, is a strong compression of a three-dimensional motion into a two-dimensional image. It encodes the reflected signal intensity as a function of slow time (ramp index, y-axis) and distance (range bins, x-axis). The dominant time-varying cause of reflections in this setup is a segment of one rotor blade entering and leaving the sensor field, which represents the pixels of high intensity in the radargram.
Since each pixel is an integrated intensity over the three-dimensional cross-section in that range bin, it effectively is the aggregated reflection of the respective rotor blade segment, i.e., spatial information is lost during that compression. Nevertheless, such a pixel computed by all reflections of that rotor blade segment is observed across many range bins and emission times during a rotor blade pass, which forms a crescent-shaped trajectory in the two-dimensional radargram. Positive attribution maps of a trained CNN show a focus on those crescents for the correct classifications (Figure 12 and Figure 14). Paired with a robust performance across known EOCs (Section 3.1.1), this indicates the existence of reliable blade-specific structural features, despite substantial compression. These features manifest as consistent pixel compositions for given EOCs, which enable the discrimination of individual rotor blades.
Furthermore, moderate variations in the installation of individual rotor blades on the same turbine are unlikely to impact classification performance. For example, if one rotor blade was reinstalled in a slightly stronger tilt than before, relative to the tower and therefore the radar sensor, this would lead to a slight shift in the radar crescent for a given EOC. It would not change the reflective properties and, therefore, it would not affect the model performance as long as comparable shifts were present in the training data across different EOCs. These observations suggest that the composition of reflections in the radargram not only resemble gross geometric motion, but also encodes surface-dependent scattering, characteristic of each individual rotor blade, influenced by material properties, structural layering, and local surface and subjacent imperfections. While such structural variations are assumed to represent the primary physical mechanism underlying the demonstrated discrimination of nominally identical rotor blades via radargrams, establishing a direct correspondence between specific material deviations and radar-derived representations remains challenging and is beyond the scope of this study.

4. Conclusions

In order to advance the emerging field of tower-radar monitoring, the robustness and explainability of convolutional classifiers were covered in this paper. It was found that fine-grained features are necessary for rotor-blade classification, which are sensitive to the possibly occurring perturbations. This task of identifying individual rotor blades is a fundamentally different challenge for a CNN than training it to recognize the overall structure or appearance of a turbine blade in contrast to unrelated objects. Using a larger model, i.e., ResNet-18 versus the compact custom CNN, yields some robust performance; however, latency limits and constrained resources could be hindering in real-time real-world systems.
The visual explanations revealed the fixation on second-order reflections of the rotor blade for some classes. These reflections only occur for some EOCs, which makes it less reliable for a general discrimination of the rotor blades. Introducing a subtle cropping based on these findings helped the model to use the main crescent of the radargram for all rotor blades. Deliberately induced data augmentations, such as random crops, helped to boost the interpolation ability. Accordingly, on the basis of what has been attained herewith, a multitude of compelling research avenues exist:
  • Thanks to the visual explanations, the high-intensity areas in the radargrams were identified, which are decisive for blade classification. In contrast, the remaining background of the crescent turned out to be obstructive in some instances. Thus, a foreground detection step could be introduced, leading to background removal or context-based adaptive cropping to aid blade recognition.
  • In the current work, perturbations were investigated, and the kinship among image transformations was pointed out. Here, augmentations help to enlarge or balance datasets; tower-radar monitoring, as a recent computer vision topic, however, will likely also benefit from the imagery of resembling domains as well as synthetic data [12].
  • The intricate but isolated task of recognizing individual rotor blades was tackled. Practical tower-radar monitoring systems nonetheless face multi-task challenges. The growth of wind energy often conflicts with the conservation of wildlife. In order to protect species, wind turbines will not even become approved in sensitive locations, or existing ones have to be temporarily switched off at designated times. In order to resolve this green-green dilemma (climate versus wildlife protection), novel detection capabilities are being developed with respect to birds and bats in the vicinity of wind parks [71,72].
  • By addressing the impact of changing EOCs, the fundamentally dynamic setting of tower-radar monitoring was acknowledged. In addition, the continual learning challenge should be treated in which the classifier adapts to new and changing data as it emerges during monitoring. Here, only weak or noisy labels are available [73] and, e.g., regularization [74] must strike a balance between preservation versus flexibility.

Author Contributions

Conceptualization, S.A. and J.M.; Methodology, S.A.; Software, S.A.; Validation, S.A.; Formal Analysis, S.A.; Investigation, S.A.; Resources, J.M.; Data Curation, S.A.; Writing—Original Draft Preparation, S.A. and C.K.; Writing—Review & Editing, S.A., C.K. and J.M.; Visualization, S.A.; Supervision, C.K. and J.M.; Project Administration, J.M.; Funding Acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the BMWK (German Federal Ministry for Economic Affairs and Climate Action, grant number 03EE2035A).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Due to confidentiality agreements with the wind power plant operator, the entire dataset supporting the findings of this study cannot be publicly shared. A small subset of the data has been deposited in Zenodo [75] and the source code for this work is publicly available on GitHub https://github.com/SaILaIDiN/TRM-of-Wind-Turbine-Blades-with-CNNs (accessed on 2 February 2026).

Acknowledgments

The authors thank Moritz Mälzer and Sebastian Beck from the department of mechanical engineering of the University of Siegen for their technical support in the data acquisition that led to this research.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
EOCEnvironmental and operational condition
FMCWFrequency-modulated continuous wave
TRMTower-radar monitoring
MLMachine learning
AIArtificial intelligence
LIMELocal interpretable model-agnostic explanations
UAVUnmanned aerial vehicle
CNNConvolutional neural network
SARSynthetic aperture radar
SCADASupervisory control and data acquisition
FFTFast Fourier transform
ReLURectified linear unit
MaxPoolMaximum pooling
ROIRegion of interest
SNRSignal-to-noise ratio

References

  1. Niemi, J.; Tanttu, J.T. Deep learning-based automatic bird identification system for offshore wind farms. Wind Energy 2020, 23, 1394–1407. [Google Scholar] [CrossRef]
  2. Rosa, I.M.D.; Marques, A.T.; Palminha, G.; Costa, H.; Mascarenhas, M.; Fonseca, C.; Bernardino, J. Classification success of six machine learning algorithms in radar ornithology. Ibis 2016, 158, 28–42. [Google Scholar] [CrossRef]
  3. Zhao, H.; Chen, G.; Hong, H.; Zhu, X. Remote structural health monitoring for industrial wind turbines using short-range Doppler radar. IEEE Trans. Instrum. Meas. 2021, 70, 8002609. [Google Scholar] [CrossRef]
  4. The European Parliament and the Council of the European Union. Regulation Laying Down Harmonised Rules on Artificial Intelligence. 2024. Available online: http://data.europa.eu/eli/reg/2024/1689/oj (accessed on 11 August 2025).
  5. Samek, W.; Müller, K.R. Towards explainable artificial intelligence. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning; Springer International Publishing: Cham, Switzerland, 2019; pp. 5–22. [Google Scholar] [CrossRef]
  6. Meyers, R.A. Mathematics of Complexity and Dynamical Systems; Springer Science & Business Media: New York, NY, USA, 2011. [Google Scholar] [CrossRef]
  7. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar] [CrossRef]
  8. Truong, J.; Rudolph, M.; Yokoyama, N.H.; Chernova, S.; Batra, D.; Rai, A. Rethinking sim2real: Lower fidelity simulation leads to higher sim2real transfer in navigation. In Proceedings of the 6th Conference on Robot Learning; PMLR: New York, NY, USA, 2023; Volume 205, pp. 859–870. [Google Scholar]
  9. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2414–2423. [Google Scholar] [CrossRef]
  10. Jing, Y.; Yang, Y.; Feng, Z.; Ye, J.; Yu, Y.; Song, M. Neural style transfer: A review. IEEE Trans. Vis. Comput. Graph. 2019, 26, 3365–3385. [Google Scholar] [CrossRef]
  11. Mälzer, M.; Beck, S.; Alipek, S.; Reichart, E.; Moll, J.; Krozer, V.; Oikonomopoulos, C.; Kassner, J.; Hägelen, M.; Heinecke, T.; et al. Radar-based structural monitoring of wind turbines blades: Field results from two operational wind turbines. In Proceedings of the 14th International Workshop on Structural Health Monitoring, Stanford, CA, USA, 12–14 September 2023. [Google Scholar] [CrossRef]
  12. Kexel, C. SiWiRoRa—Simulated Wind-Turbine Rotor-Blade Radargrams; Zenodo: Geneva, Switzerland, 2024. [Google Scholar] [CrossRef]
  13. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  14. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar] [CrossRef]
  15. Mumuni, A.; Mumuni, F. Data augmentation: A comprehensive survey of modern approaches. Array 2022, 16, 100258. [Google Scholar] [CrossRef]
  16. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast Image Augmentation Library. 2024. Available online: https://albumentations.ai (accessed on 11 August 2025).
  17. Gohar, I.; Yew, W.K.; Halimi, A.; See, J. Review of state-of-the-art surface defect detection on wind turbine blades through aerial imagery: Challenges and recommendations. Eng. Appl. Artif. Intell. 2025, 144, 109970. [Google Scholar] [CrossRef]
  18. Liu, H.; Liu, S.; Liu, Z.; Niu, B.; Xie, J.; Luo, C.; Shi, Z. Wind turbine blade surface defect detection model based on improved you only look once version 10 small and integrated compression. Eng. Appl. Artif. Intell. 2025, 159, 111645. [Google Scholar] [CrossRef]
  19. Jia, X.; Chen, X. AI-based optical-thermal video data fusion for near real-time blade segmentation in normal wind turbine operation. Eng. Appl. Artif. Intell. 2024, 127, 107325. [Google Scholar] [CrossRef]
  20. Streser, E.; Alipek, S.; Rao, M.; Simon, J.; Moll, J.; Kraemer, P.; Krozer, V. Radar-based damage detection in a wind turbine blade using convolutional neural networks: A proof-of-concept under fatigue loading. Sensors 2025, 25, 3337. [Google Scholar] [CrossRef] [PubMed]
  21. Moll, J.; Arnold, P.; Mälzer, M.; Krozer, V.; Pozdniakov, D.; Salman, R.; Rediske, S.; Scholz, M.; Friedmann, H.; Nuber, A. Radar-based structural health monitoring of wind turbine blades: The case of damage detection. Struct. Health Monit. 2018, 17, 815–822. [Google Scholar] [CrossRef]
  22. Tang, W.; Blanche, J.; Mitchell, D.; Harper, S.; Flynn, D. Characterisation of composite materials for wind turbines using frequency modulated continuous wave sensing. J. Compos. Sci. 2023, 7, 75. [Google Scholar] [CrossRef]
  23. Alipek, S.; Mälzer, M.; Beck, S.; Kexel, C.; Moll, J.; Krozer, V.; Kassner, J.; Heinecke, T.; Rose, J.; Berger, M. Potential and limitations of anomaly detection via tower-radar monitoring of wind turbine blades in regular operation with convolutional networks. e-J. Nondestruct. Test. 2024, 29, 1–11. [Google Scholar] [CrossRef]
  24. Tu, W.; Deng, W.; Gedeon, T. A closer look at the robustness of contrastive language-image pre-training (clip). Adv. Neural Inf. Process. Syst. 2023, 36, 13678–13691. [Google Scholar]
  25. Hendrycks, D.; Dietterich, T. Benchmarking neural network robustness to common corruptions and perturbations. arXiv 2019, arXiv:1903.12261. [Google Scholar] [CrossRef]
  26. Mintun, E.; Kirillov, A.; Xie, S. On interaction between augmentations and corruptions in natural corruption robustness. Adv. Neural Inf. Process. Syst. 2021, 34, 3571–3583. [Google Scholar]
  27. Oveis, A.H.; Giusti, E.; Ghio, S.; Meucci, G.; Martorella, M. LIME-Assisted automatic target recognition with SAR images: Toward incremental learning and explainability. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 9175–9192. [Google Scholar] [CrossRef]
  28. Qiu, Z.; Rivaz, H.; Xiao, Y. Is visual explanation with Grad-CAM more reliable for deeper neural networks? A case study with automatic pneumothorax diagnosis. In Machine Learning in Medical Imaging; Springer: Cham, Switzerland, 2023; pp. 224–233. [Google Scholar] [CrossRef]
  29. Galli, A.; Marrone, S.; Moscato, V.; Sansone, C. Reliability of explainable artificial intelligence in adversarial perturbation scenarios. In Pattern Recognition. ICPR International Workshops and Challenges; Springer: Cham, Switzerland, 2021; pp. 243–256. [Google Scholar] [CrossRef]
  30. Czerkawski, M.; Clemente, C.; Michie, C.; Andonovic, I.; Tachtatzis, C. Robustness of deep neural networks for micro-Doppler radar classification. In Proceedings of the 2022 23rd International Radar Symposium (IRS), Gdansk, Poland, 12–14 September 2022; pp. 480–485. [Google Scholar] [CrossRef]
  31. Chen, T.; Zhang, L.; Guo, W.; Zhang, Z.; Datcu, M. Analyzing the Adversarial Robustness and Interpretability of Deep SAR Classification Models: A Comprehensive Examination of Their Reliability. Remote Sens. 2025, 17, 1943. [Google Scholar] [CrossRef]
  32. Oveis, A.H.; Giusti, E.; Cantelli-Forti, A.; Martorella, M. XAI-Driven Resilient Image Classification in the Presence of Adversarial Perturbations. IET Radar Sonar Navig. 2025, 19, e70041. [Google Scholar] [CrossRef]
  33. Tien, T.B.; Quang, T.V.; Ngoc, L.N.; Ngoc, H.T. Enhancing time series data classification for structural damage detection through out-of-distribution representation learning. Structures 2024, 65, 106766. [Google Scholar] [CrossRef]
  34. Shu, J.; Ding, W.; Zhang, J.; Lin, F.; Duan, Y. Continual-learning-based framework for structural damage recognition. Struct. Control Health Monit. 2022, 29, e3093. [Google Scholar] [CrossRef]
  35. Golubović, D.; Erić, M.; Vukmirović, N. High-resolution Doppler and azimuth estimation and target detection in HFSWR: Experimental study. Sensors 2022, 22, 3558. [Google Scholar] [CrossRef]
  36. Walenczykowska, M.; Kawalec, A. Application of continuous wavelet transform and artificial naural network for automatic radar signal recognition. Sensors 2022, 22, 7434. [Google Scholar] [CrossRef]
  37. Dai, W.; Dai, C.; Qu, S.; Li, J.; Das, S. Very deep convolutional neural networks for raw waveforms. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 421–425. [Google Scholar] [CrossRef]
  38. Manwell, J.F.; McGowan, J.G.; Rogers, A.L. Wind Energy Explained: Theory, Design and Application; John Wiley & Sons: Hoboken, NJ, USA, 2010. [Google Scholar] [CrossRef]
  39. Kim, B.S.; Jin, Y.; Lee, J.; Kim, S. FMCW radar estimation algorithm with high resolution and low complexity based on reduced search area. Sensors 2022, 22, 1202. [Google Scholar] [CrossRef]
  40. Moll, J.; Maetz, T.; Maelzer, M.; Krozer, V.; Mischke, K.; Krause, S.; Bagemiel, O.; Nuber, A.; Kremling, S.; Kurin, T.; et al. Radar-based monitoring of glass fiber reinforced composites during fatigue testing. Struct. Control Health Monit. 2021, 28, e2812. [Google Scholar] [CrossRef]
  41. Cortés, E.; Sánchez, F.; O’Carroll, A.; Madramany, B.; Hardiman, M.; Young, T.M. On the material characterisation of wind turbine blade coatings: The effect of interphase coating–laminate adhesion on rain erosion performance. Materials 2017, 10, 1146. [Google Scholar] [CrossRef]
  42. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  43. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  44. Wightman, R. PyTorch Image Models. 2019. Available online: https://github.com/rwightman/pytorch-image-models (accessed on 11 August 2025).
  45. Masters, D.; Luschi, C. Revisiting small batch training for deep neural networks. arXiv 2018, arXiv:1804.07612. [Google Scholar] [CrossRef]
  46. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Computer Vision—ECCV 2014; Springer: Cham, Switzerland, 2014; pp. 818–833. [Google Scholar] [CrossRef]
  47. Springenberg, J.T.; Dosovitskiy, A.; Brox, T.; Riedmiller, M. Striving for simplicity: The all convolutional net. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar] [CrossRef]
  48. Sundararajan, M.; Taly, A.; Yan, Q. Axiomatic Attribution for Deep Networks. arXiv 2017. [Google Scholar] [CrossRef]
  49. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
  50. Meta. Captum—Model Interpretability for PyTorch. 2020. Available online: https://captum.ai/api/guided_grad_cam.html (accessed on 11 August 2025).
  51. Dubey, S.R.; Singh, S.K.; Chaudhuri, B.B. Activation functions in deep learning: A comprehensive survey and benchmark. Neurocomputing 2022, 503, 92–108. [Google Scholar] [CrossRef]
  52. Agarap, A.F. Deep Learning using Rectified Linear Units (ReLU). arXiv 2019, arXiv:1803.08375. [Google Scholar] [CrossRef]
  53. Farrar, C.; Dervilis, N.; Worden, K. The past, present and future of structural health monitoring: An overview of three ages. Strain 2025, 61, e12495. [Google Scholar] [CrossRef]
  54. Chen, Y.; Griffith, D.T. Blade mass imbalance identification and estimation for three-bladed wind turbine rotor based on modal analysis. Mech. Syst. Signal Process. 2023, 197, 110341. [Google Scholar] [CrossRef]
  55. Krogh, C.; Hermansen, S.M.; Lund, E.; Kepler, J.; Jakobsen, J. A matter of course: Generating optimal manufacturing instructions from a structural layup plan of a wind turbine blade. Compos. Part A Appl. Sci. Manuf. 2023, 172, 107599. [Google Scholar] [CrossRef]
  56. Tong, K.; Wu, Y. Deep learning-based detection from the perspective of small or tiny objects: A survey. Image Vis. Comput. 2022, 123, 104471. [Google Scholar] [CrossRef]
  57. Talebi, H.; Milanfar, P. Learning to resize images for computer vision tasks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 497–506. [Google Scholar] [CrossRef]
  58. Montanari, A.; Zhong, Y. The interpolation phase transition in neural networks: Memorization and generalization under lazy training. Ann. Stat. 2022, 50, 2816–2847. [Google Scholar] [CrossRef]
  59. Mallinar, N.R.; Simon, J.B.; Abedsoltan, A.; Pandit, P.; Belkin, M.; Nakkiran, P. Benign, Tempered, or Catastrophic: Toward a Refined Taxonomy of Overfitting. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
  60. Kexel, C.; Mueller, H.; Alipek, S.; Moll, J. Mast-bound and Too Curious: Overcoming Drift in Wind-Tower Radar for Blade Monitoring Using Pre-training, Augmentation and Weight Consolidation Due to Correlated Conditions. In Proceedings of the 2nd Latin-American Workshop on Structural Health Monitoring (LATAM-SHM), Santiago, Chile, 7–9 January 2026. [Google Scholar] [CrossRef]
  61. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? Adv. Neural Inf. Process. Syst. 2014, 27, 3320–3328. [Google Scholar] [CrossRef]
  62. Ardalan, Z.; Subbian, V. Transfer learning approaches for neuroimaging analysis: A scoping review. Front. Artif. Intell. 2022, 5, 780405. [Google Scholar] [CrossRef] [PubMed]
  63. Kornblith, S.; Shlens, J.; Le, Q.V. Do better imagenet models transfer better? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2661–2671. [Google Scholar] [CrossRef]
  64. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
  65. Yang, R.; Wang, R.; Deng, Y.; Jia, X.; Zhang, H. Rethinking the random cropping data augmentation method used in the training of CNN-based SAR image ship detector. Remote Sens. 2020, 13, 34. [Google Scholar] [CrossRef]
  66. Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X.; Pietikäinen, M. Deep learning for generic object detection: A survey. Int. J. Comput. Vis. 2020, 128, 261–318. [Google Scholar] [CrossRef]
  67. Rossi, L.; Karimi, A.; Prati, A. A novel region of interest extraction layer for instance segmentation. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 2203–2209. [Google Scholar] [CrossRef]
  68. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and flexible image augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef]
  69. Ali, Z.; Elsayed, M.; Tiwari, G.; Ahmad, M.; Le Kernec, J.; Heidari, H.; Gupta, S. Impact of receiver thermal noise and PLL RMS jitter in radar measurements. IEEE Trans. Instrum. Meas. 2024, 73, 2002710. [Google Scholar] [CrossRef]
  70. Zadeh, A.T.; Mälzer, M.; Simon, J.; Beck, S.; Moll, J.; Krozer, V. Range–Doppler analysis for rain detection at Ka-band: Numerical and experimental results from laboratory and field measurements. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1027–1033. [Google Scholar] [CrossRef]
  71. Mälzer, M.; Taremi Zadeh, A.; Beck, S.; Moll, J.; Krozer, V. Towards radar barriers for animal fatality detection at wind turbines: Numerical and preliminary experimental results. IET Radar Sonar Navig. 2020, 14, 1767–1772. [Google Scholar] [CrossRef]
  72. Zadeh, A.T.; Diyap, M.; Moll, J.; Krozer, V. Towards localization and classification of birds and bats in windparks using multiple FMCW-radars at Ka-band. Prog. Electromagn. Res. M 2022, 109, 1–12. [Google Scholar] [CrossRef]
  73. Verwimp, E.; Aljundi, R.; Ben-David, S.; Bethge, M.; Cossu, A.; Gepperth, A.; Hayes, T.L.; Hüllermeier, E.; Kanan, C.; Kudithipudi, D.; et al. Continual learning: Applications and the road forward. Trans. Mach. Learn. Res. 2024, 2835–8856. Available online: https://openreview.net/forum?id=axBIMcGZn9 (accessed on 2 February 2026).
  74. Mundt, M.; Hong, Y.; Pliushch, I.; Ramesh, V. A wholistic view of continual learning with deep neural networks: Forgotten lessons and the bridge to active and open world learning. Neural Netw. 2023, 160, 306–336. [Google Scholar] [CrossRef] [PubMed]
  75. Mälzer, M.; Beck, S.; Moll, J. Radar-based sensing of wind turbines blades based on 35 GHz FMCW sensors installed at operational wind turbine towers. In Proceedings of the 14th International Workshop on Structural Health Monitoring (IWSHM 2023); Zenodo: Geneva, Switzerland, 2024. [Google Scholar] [CrossRef]
Figure 1. Adjacent concepts from computer vision applied to radargrams: Adversarial example, common corruption (blurriness), data augmentation (contrast or brightness changes), sim2real gap, geometric image transformation (EOCs), style transfer (rotor blade is mistaken for moon crescent). Input image taken from Mälzer et al. [11] and simulation from Kexel [12].
Figure 1. Adjacent concepts from computer vision applied to radargrams: Adversarial example, common corruption (blurriness), data augmentation (contrast or brightness changes), sim2real gap, geometric image transformation (EOCs), style transfer (rotor blade is mistaken for moon crescent). Input image taken from Mälzer et al. [11] and simulation from Kexel [12].
Sensors 26 01083 g001
Figure 2. Frequency-modulated continuous wave (FMCW) radar sensor unit [23]. A high-speed camera (100 images/s) is installed next to the FMCW radar (1000 frequency ramps/s). The latter is visible by the green circuit board that is the front-end holding the Rx and Tx antennas.
Figure 2. Frequency-modulated continuous wave (FMCW) radar sensor unit [23]. A high-speed camera (100 images/s) is installed next to the FMCW radar (1000 frequency ramps/s). The latter is visible by the green circuit board that is the front-end holding the Rx and Tx antennas.
Sensors 26 01083 g002
Figure 3. View of the experimental setup [11]. The sensor box is mounted on the wind turbine tower in the main wind direction of 230°. A barcode foil is attached on the surface of the rotor blade facing the sensor box.
Figure 3. View of the experimental setup [11]. The sensor box is mounted on the wind turbine tower in the main wind direction of 230°. A barcode foil is attached on the surface of the rotor blade facing the sensor box.
Sensors 26 01083 g003
Figure 4. Comparison of radargrams of different EOCs [23]. (Left): Nacelle Orientation = 198°, Rotation Speed = 11 rpm. (Middle): Nacelle Orientation = 227°, Rotation Speed = 6 rpm. (Right): Nacelle Orientation = 256°, Rotation Speed = 10.5 rpm (rpm—rotations per minute).
Figure 4. Comparison of radargrams of different EOCs [23]. (Left): Nacelle Orientation = 198°, Rotation Speed = 11 rpm. (Middle): Nacelle Orientation = 227°, Rotation Speed = 6 rpm. (Right): Nacelle Orientation = 256°, Rotation Speed = 10.5 rpm (rpm—rotations per minute).
Sensors 26 01083 g004
Figure 5. Simplified logical structure of how environmental conditions correlate with a wind turbine’s functioning that shapes the final radargram. Wind conditions, defined by the weather, drive operational adjustments of the turbine which ultimately define how the rotor blade moves in relation to the sensor field. Coloring of control mechanisms (purple) is in contrast to measurements (black).
Figure 5. Simplified logical structure of how environmental conditions correlate with a wind turbine’s functioning that shapes the final radargram. Wind conditions, defined by the weather, drive operational adjustments of the turbine which ultimately define how the rotor blade moves in relation to the sensor field. Coloring of control mechanisms (purple) is in contrast to measurements (black).
Sensors 26 01083 g005
Figure 6. Illustrating the hypothesis behind blade classification. Top-view at (sub-)centimeter radar emissions interacting with the layered internal structure of the radar-facing portion of a wind turbine rotor blade (cross-section shown).
Figure 6. Illustrating the hypothesis behind blade classification. Top-view at (sub-)centimeter radar emissions interacting with the layered internal structure of the radar-facing portion of a wind turbine rotor blade (cross-section shown).
Sensors 26 01083 g006
Figure 7. Confusion matrix for a model trained on vb06 data and tested on vb07 data (see column 3 in Table 1).
Figure 7. Confusion matrix for a model trained on vb06 data and tested on vb07 data (see column 3 in Table 1).
Sensors 26 01083 g007
Figure 8. Each rotor blade of both wind turbines is represented by a geometric shape, which symbolizes the unique set of patterns each rotor blade can hold. This symbolic comparison specifically resembles the case of column 3 in Table 1. As retrieved from the confusion matrix in Figure 7 the rotor blades from vb07 have the highest similarity to ROT2 from vb06. This is symbolized by the trapezoid, which is more similar to the forms of vb06.
Figure 8. Each rotor blade of both wind turbines is represented by a geometric shape, which symbolizes the unique set of patterns each rotor blade can hold. This symbolic comparison specifically resembles the case of column 3 in Table 1. As retrieved from the confusion matrix in Figure 7 the rotor blades from vb07 have the highest similarity to ROT2 from vb06. This is symbolized by the trapezoid, which is more similar to the forms of vb06.
Sensors 26 01083 g008
Figure 9. Visualized decisions of the custom convolutional neural network (CNN) for the three classes in the case of the predicted class and given ground truth label. Regions of interest are additionally plotted with upscale factor 4 to highlight small attributions (Case 1). (a) Class 1, predicted class 1, ground truth class 1. (b) Class 2, predicted class 1, ground truth class 1. (c) Class 3, predicted class 1, ground truth class 1.
Figure 9. Visualized decisions of the custom convolutional neural network (CNN) for the three classes in the case of the predicted class and given ground truth label. Regions of interest are additionally plotted with upscale factor 4 to highlight small attributions (Case 1). (a) Class 1, predicted class 1, ground truth class 1. (b) Class 2, predicted class 1, ground truth class 1. (c) Class 3, predicted class 1, ground truth class 1.
Sensors 26 01083 g009
Figure 10. Visualized decisions of the custom CNN for the three classes in case of the predicted class and given ground truth label (Case 2). (a) Class 1, predicted class 2, ground truth class 2. (b) Class 2, predicted class 2, ground truth class 2. (c) Class 3, predicted class 2, ground truth class 2.
Figure 10. Visualized decisions of the custom CNN for the three classes in case of the predicted class and given ground truth label (Case 2). (a) Class 1, predicted class 2, ground truth class 2. (b) Class 2, predicted class 2, ground truth class 2. (c) Class 3, predicted class 2, ground truth class 2.
Sensors 26 01083 g010
Figure 11. Visualized decisions of the ResNet-18 for the three classes in the case of the predicted class and given ground truth label (Ground truth class 1). (a) Class 1, predicted class 1, ground truth class 1. (b) Class 2, predicted class 1, ground truth class 1. (c) Class 3, predicted class 1, ground truth class 1.
Figure 11. Visualized decisions of the ResNet-18 for the three classes in the case of the predicted class and given ground truth label (Ground truth class 1). (a) Class 1, predicted class 1, ground truth class 1. (b) Class 2, predicted class 1, ground truth class 1. (c) Class 3, predicted class 1, ground truth class 1.
Sensors 26 01083 g011
Figure 12. Visualized decisions of the ResNet-18 for the three classes in the case of the predicted class and given ground truth label (Ground truth class 2). (a) Class 1, predicted class 2, ground truth class 2. (b) Class 2, predicted class 2, ground truth class 2. (c) Class 3, predicted class 2, ground truth class 2.
Figure 12. Visualized decisions of the ResNet-18 for the three classes in the case of the predicted class and given ground truth label (Ground truth class 2). (a) Class 1, predicted class 2, ground truth class 2. (b) Class 2, predicted class 2, ground truth class 2. (c) Class 3, predicted class 2, ground truth class 2.
Sensors 26 01083 g012
Figure 13. Visualized decisions of the ResNet-18 for the three classes in the case of the predicted class and given ground truth label. (Ground truth class 3). (a) Class 1, predicted class 3, ground truth class 3. (b) Class 2, predicted class 3, ground truth class 3. (c) Class 3, predicted class 3, ground truth class 3.
Figure 13. Visualized decisions of the ResNet-18 for the three classes in the case of the predicted class and given ground truth label. (Ground truth class 3). (a) Class 1, predicted class 3, ground truth class 3. (b) Class 2, predicted class 3, ground truth class 3. (c) Class 3, predicted class 3, ground truth class 3.
Sensors 26 01083 g013
Figure 14. Typical attributions found for a centered cropping in case of a correct classification.
Figure 14. Typical attributions found for a centered cropping in case of a correct classification.
Sensors 26 01083 g014
Figure 15. Example of unsuitable attributions from too aggressive cropping.
Figure 15. Example of unsuitable attributions from too aggressive cropping.
Sensors 26 01083 g015
Figure 16. Comparison of pixel-level data augmentation techniques on the same measurement example.
Figure 16. Comparison of pixel-level data augmentation techniques on the same measurement example.
Sensors 26 01083 g016
Table 1. F1 scores on various dataset compositions. (Column 1): the model was trained on the training partition for turbine vb06 and evaluated on another partition of the same turbine. (Column 2): equivalent setup with vb07. (Column 3 & 4): models were each trained on one turbine and evaluated on the test partition of the other. (Column 5): the model was first trained on vb06, then it was fine-tuned on 60% of the complete dataset of vb07 and then evaluated on test data from vb07. (Column 6): represents the vice versa of column 5. (Column 7): same configuration as column 5 but fine-tuned with only 10% of the complete dataset. (Column 8): represents the vice versa of column 7.
Table 1. F1 scores on various dataset compositions. (Column 1): the model was trained on the training partition for turbine vb06 and evaluated on another partition of the same turbine. (Column 2): equivalent setup with vb07. (Column 3 & 4): models were each trained on one turbine and evaluated on the test partition of the other. (Column 5): the model was first trained on vb06, then it was fine-tuned on 60% of the complete dataset of vb07 and then evaluated on test data from vb07. (Column 6): represents the vice versa of column 5. (Column 7): same configuration as column 5 but fine-tuned with only 10% of the complete dataset. (Column 8): represents the vice versa of column 7.
Train →
Finetune →
Test →
vb06
-
vb06
vb07
-
vb07
vb06
-
vb07
vb07
-
vb06
vb06
vb07
vb07
vb07
vb06
vb06
vb06
vb07
vb07
vb07
vb06
vb06
ROT10.9810.9100.1930.4780.8810.9680.7490.917
ROT20.9790.8980.4520.2080.8630.9710.7270.896
ROT30.9800.8630.3360.2170.8130.9710.7160.910
Table 2. F1 scores for systematic speed variation. (Columns 1 & 2): test case when EOCs match the training data. (Columns 3 & 4): lack of extrapolation for rotation speeds unseen at training. (Column 5): model trained with high and low rotation speed measurements. (Column 6): interpolation performance by excluding intermediate rotation, and therefore, wind speeds. All results are for the custom CNN except the last. (Column 7): ResNet-18 performance from Section 3.3.2 is indicated by an asterisk (*) and provided here for convenient comparison. (Column 8): Same as column 7 but initialized with pretrained weights from ImageNet-1k, which is indicated by a second asterisk (**).
Table 2. F1 scores for systematic speed variation. (Columns 1 & 2): test case when EOCs match the training data. (Columns 3 & 4): lack of extrapolation for rotation speeds unseen at training. (Column 5): model trained with high and low rotation speed measurements. (Column 6): interpolation performance by excluding intermediate rotation, and therefore, wind speeds. All results are for the custom CNN except the last. (Column 7): ResNet-18 performance from Section 3.3.2 is indicated by an asterisk (*) and provided here for convenient comparison. (Column 8): Same as column 7 but initialized with pretrained weights from ImageNet-1k, which is indicated by a second asterisk (**).
Train →
Test →
Low
Low
High
High
Low
High
High
Low
Both
Both
Both
Mid
Both *
Mid *
Both **
Mid **
ROT10.9620.9290.4780.3440.9700.5430.6790.868
ROT20.9650.9130.3240.3980.9650.5450.6570.866
ROT30.9790.9440.0070.1730.9750.6350.7690.921
Table 3. EOCs for systematic speed variation. For fair comparison all other relevant meta-parameters are reasonably set: e.g., the pitch angle is allowed to vary within a very small interval at same wind speeds and direction, in order to provide a sufficient number of train and test samples.
Table 3. EOCs for systematic speed variation. For fair comparison all other relevant meta-parameters are reasonably set: e.g., the pitch angle is allowed to vary within a very small interval at same wind speeds and direction, in order to provide a sufficient number of train and test samples.
Meta-ParameterLowMidHigh
rotation speed [rpm]{7.0, 9.0}{9.5, 11.5}{12.5, 13.5}
nacelle orientation [°]{200, 270} *
pitch angle [°]{−2.5, −1.5} *
wind direction [°]{150, 300} *
wind speed [m/s]{4, 7}{7, 8}{8, 11.5}
* Same parameter interval for all three partitions.
Table 4. F1 scores for systematic nacelle variation with similar findings as in Table 2.
Table 4. F1 scores for systematic nacelle variation with similar findings as in Table 2.
Train →
Test →
Low
Low
High
High
Low
High
High
Low
Both
Both
Both
Mid
ROT10.9260.9990.2350.4300.9710.458
ROT20.9350.9910.4000.2640.9710.344
ROT30.9500.9920.2910.1010.9790.538
Table 5. F1 scores for systematic pitch angle variation with similar findings as in Table 2.
Table 5. F1 scores for systematic pitch angle variation with similar findings as in Table 2.
Train →
Test →
Low
Low
High
High
Low
High
High
Low
Both
Both
Both
Mid
ROT10.9850.9720.5030.4590.9860.889
ROT20.9850.9740.4700.0820.9850.696
ROT30.9900.9730.0650.0130.9860.821
Table 6. EOCs for systematic nacelle variation. The difference between nacelle orientation and wind direction were matched for each sample to be below ten degrees.
Table 6. EOCs for systematic nacelle variation. The difference between nacelle orientation and wind direction were matched for each sample to be below ten degrees.
Meta-ParameterLowMidHigh
rotation speed [rpm]{0, 13.5} *
nacelle orientation [°]{200, 220}{225, 245}{250, 270}
pitch angle [°]{−5, 2} *
wind direction [°]{190, 230}{225, 245}{240, 280}
wind speed [m/s]{0, 10.5} *
* Same parameter interval for all three partitions.
Table 7. EOCs for systematic pitch variation. To reveal the potential impact of the pitch angle, the rotation speed must be kept in a very small interval since it is highly changing the radar profile, potentially inducing a bias. For a fixed wind direction, nacelle orientation, and rotation speed, the wind speed correlates with the pitch angle, because the pitch angle defines how fast the rotor blade would move under the contemporary wind speed. Therefore, if the rotation speed is held fixed, the pitch angle and wind speed cannot be decoupled, but it appears to be the only way to investigate the impact of the pitch angle.
Table 7. EOCs for systematic pitch variation. To reveal the potential impact of the pitch angle, the rotation speed must be kept in a very small interval since it is highly changing the radar profile, potentially inducing a bias. For a fixed wind direction, nacelle orientation, and rotation speed, the wind speed correlates with the pitch angle, because the pitch angle defines how fast the rotor blade would move under the contemporary wind speed. Therefore, if the rotation speed is held fixed, the pitch angle and wind speed cannot be decoupled, but it appears to be the only way to investigate the impact of the pitch angle.
Meta-ParameterLowMidHigh
rotation speed [rpm]{12.25, 13.25} *
nacelle orientation [°]{237.5, 262.5} *
pitch angle [°]{−3, 3}{5, 10}{12, 18}
wind direction [°]{237.5, 262.5} *
wind speed [m/s]{8.25, 11.5}{10, 16}{14.5, 18.5}
* Same parameter interval for all three partitions.
Table 8. The first six columns show the F1 scores for a ResNet-18 trained and tested on a joint dataset of radargrams with low and high rotation speed. The last three columns list the F1 scores for a ResNet-18 trained with data containing low and high rotation speeds and tested on radargrams with unknown rotation speeds between training data intervals. The first row shows results without data augmentation, and the second row with random cropping.
Table 8. The first six columns show the F1 scores for a ResNet-18 trained and tested on a joint dataset of radargrams with low and high rotation speed. The last three columns list the F1 scores for a ResNet-18 trained with data containing low and high rotation speeds and tested on radargrams with unknown rotation speeds between training data intervals. The first row shows results without data augmentation, and the second row with random cropping.
Train
ROT1
Train
ROT2
Train
ROT3
Test
ROT1
Test
ROT2
Test
ROT3
Test *
ROT1
Test *
ROT2
Test *
ROT3
raw0.9960.9960.9990.9800.9760.9880.7440.6770.820
crop0.9530.9530.9710.9460.9460.9640.7570.8140.823
* Additional interpolation test.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alipek, S.; Kexel, C.; Moll, J. Explainable Machine Learning for Tower-Radar Monitoring of Wind Turbine Blades: Fine-Grained Blade Recognition Under Changing Operational Conditions. Sensors 2026, 26, 1083. https://doi.org/10.3390/s26041083

AMA Style

Alipek S, Kexel C, Moll J. Explainable Machine Learning for Tower-Radar Monitoring of Wind Turbine Blades: Fine-Grained Blade Recognition Under Changing Operational Conditions. Sensors. 2026; 26(4):1083. https://doi.org/10.3390/s26041083

Chicago/Turabian Style

Alipek, Sercan, Christian Kexel, and Jochen Moll. 2026. "Explainable Machine Learning for Tower-Radar Monitoring of Wind Turbine Blades: Fine-Grained Blade Recognition Under Changing Operational Conditions" Sensors 26, no. 4: 1083. https://doi.org/10.3390/s26041083

APA Style

Alipek, S., Kexel, C., & Moll, J. (2026). Explainable Machine Learning for Tower-Radar Monitoring of Wind Turbine Blades: Fine-Grained Blade Recognition Under Changing Operational Conditions. Sensors, 26(4), 1083. https://doi.org/10.3390/s26041083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop