Next Article in Journal
Tribological Performance of an Automatic Transmission Fluid Additized with a Phosphonium-Based Ionic Liquid Under Electrified Conditions
Previous Article in Journal
On the Effectiveness of Optimisation Algorithms for Hydrodynamic Lubrication Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Recognition Method for Ferrography Wear Debris Images Using Improved Mask R-CNN Methods

by
Xiangwen Xiao
1,
Weixuan Zhang
1,
Qing Wang
2,
Yuan Liu
3 and
Yishou Wang
1,*
1
School of Aerospace Engineering, Xiamen University, Xiamen 361005, China
2
Xiamen Airlines, Xiamen 361003, China
3
AECC Hunan Aviation Powerplant Research Institute, Zhuzhou 412000, China
*
Author to whom correspondence should be addressed.
Lubricants 2025, 13(5), 208; https://doi.org/10.3390/lubricants13050208
Submission received: 2 April 2025 / Revised: 26 April 2025 / Accepted: 8 May 2025 / Published: 9 May 2025

Abstract

:
The accurate characterization of wear debris is crucial for assessing the health of rotating engine components and for conducting simulation experiments in debris detection. This study proposed an intelligent recognition method for ferrography wear debris images, leveraging several improved Mask Region-based Convolutional Neural Network (Mask R-CNN) algorithms to quantitatively calculate both the number of debris particles and their coverage areas. The improvement on the Mask R-CNN focuses on two key aspects: enhancing feature extraction through the feature pyramid network structure and integrating attention mechanisms. The most suitable attention mechanism for wear debris detection was determined through ablation experiments. The improved Mask R-CNN combined with the Convolutional Block Attention Module achieves the best Mean Pixel Accuracy of 87.63% at a processing speed of 7.6 frames per second, demonstrating its high accuracy and efficiency in wear particle segmentation. Furthermore, the quantitative and qualitative analysis of wear debris, including the number and area of debris particles and their classification, provides valuable insights into the severity of wear. These insights are essential for understanding the extent of wear damage and guiding maintenance decisions.

1. Introduction

In current simulations and experiments related to oil wear debris monitoring, spheres with varying diameters are commonly used to simulate and characterize wear debris. Wang et al. proposed a coaxial capacitive sensor network to monitor wear debris and simulated the effects of different spherical debris sizes on capacitance signals [1,2]. However, actual wear debris predominantly exhibits irregular shapes [3,4,5]. The wear debris detection methods that are verified by idealized spherical particles presents several challenges in engineering, including low recognition rates, high false alarm rates, and insufficient reliability. The size and concentration of wear debris are directly related to the wear condition of rotating components, as shown in Figure 1 [6]. Improving the reliability of oil wear debris recognition methods requires accurate information about the debris, including its quantity, size, and shape.
The standardized procedure for aviation oil debris analysis comprises four operational phases, as schematically depicted in Figure 2. Upon alert activation by the Oil Debris Monitoring System (ODMS) or during scheduled maintenance intervals, maintenance crews extract oil samples from strategic collection points—specifically the low-pressure circulation lines and reservoir sumps of turbine lubrication systems—to capture gravimetrically settled particulates. Following vacuum degassing to eliminate entrapped air bubbles, the protocol implements dual diagnostic modalities: (1) ferrographic microscopy for morphological characterization of particulate matter, and (2) spectrometric analysis for elemental composition determination to trace debris origination. This process culminates in the generation of standardized Magnetic Chip Detector (MCD) reports, as shown in Figure 2.
Magnetic Chip Detector (MCD) reports contain ferrography wear debris images, offering valuable information about wear debris, including its magnetism, size, and appearance. The extraction and identification of the number, size, and shape of wear debris in ferrography images have become a focal point of current research.
The application of digital image processing technology provides an effective approach for analyzing wear debris images. Image detection techniques in digital image processing can be categorized into two main types: traditional methods and neural network-based techniques [7].
Among traditional digital image processing methods, techniques such as graying, binarization, filtering, and morphological operations are commonly employed to detect the edges of ferrography wear debris images. Subsequently, morphological features including aspect ratio, roundness, debris size, crimp, and other morphological factors are extracted. Concurrently, principal component analysis (PCA) is used to classify the wear debris. Wu et al. applied morphological operations to segment wear debris images and found that the size of the wear debris after segmentation follows a Weibull distribution [8]. However, the large number of morphological feature parameters poses challenges in selecting the appropriate parameters and developing a high-accuracy detection model. These challenges limit the practical application of traditional digital image processing methods in engineering. In contrast, neural network-based image processing technology offers greater robustness, as it eliminates the need for manual parameter selection. A high-accuracy detection model can be developed by combining neural networks with labeled image datasets. In recent years, significant progress has been made in the recognition of ferrography wear debris images using neural networks, both domestically and internationally.
Cao and Xie applied wavelet transform to wear particle images, filtered noise and background, extracted wavelet-based feature vectors, and developed a BP neural network classifier for wear particle classification [9]. Shi et al. utilized the YOLO V5S model to detect wear debris images, enhancing its ability to detect wear particles of various sizes [10]. Jia et al. incorporated a self-attention residual module into YOLOv3 to improve detection of small- and multi-scale wear debris in ferrography images [11]. Additionally, Jia et al. employed DenseNet (Densely Connected Convolutional Networks) as the base network and applied transfer learning to develop the FWDNet (Ferrography Wear Debris Networks) model, achieving 90% accuracy [12]. The WPC-SS model, proposed by Fan et al., uses semantic segmentation and attention mechanisms to improve the multi-label classification of online wear particle images, surpassing standard classification methods in terms of accuracy [13]. He et al. proposed CBAM-YOLOv5, an intelligent recognition algorithm for ferrography wear particles that addresses challenges such as image blurring, background complexity, overlapping particles, and low-light conditions, significantly improving detection accuracy and speed [14]. Peng et al. introduced WP-DRnet for automatic ferrography image analysis [15]. Liu et al. developed a DCNN model for semantic segmentation of multiple types of wear particles in ferrography images [16]. Fan et al. proposed an improved YOLOv8-CGA-ASF-DBB method for multi-class wear debris recognition in online ferrography images, achieving a recognition accuracy of 94.53% with a processing time of 0.69 ms per image, thanks to enhanced feature extraction and recognition capabilities [17].
The aforementioned studies primarily utilized object detection or semantic segmentation techniques. While the object detection method localizes objects within a bounding box, the precise contours of the objects cannot be captured. Semantic segmentation method delineates the contours but struggles to distinguish adjacent objects. In the ferrography wear debris images, the adjacent wear debris particles are common, as shown in Figure 3. Consequently, how to extract comprehensive information from these images remains a challenge.
He et al. proposed an instance segmentation method, Mask Region-based Convolutional Neural Networks (Mask R-CNN) [18], which simultaneously performs both object detection and semantic segmentation. This method not only classifies objects at the pixel level but also distinguishes between different instances of similar objects. In the context of ferrography wear debris image recognition, the instance segmentation can be used to separate adjacent wear debris particles, enabling the calculation of size and shape characteristics for each individual debris particle. Examples of object detection, semantic segmentation, and instance segmentation are shown in Figure 4.
To achieve more accurate segmentation of wear debris images, this paper proposed several enhanced Mask R-CNN methods with multi-feature fusion. The proposed methods incorporate two main improvements. The first involves enhancing the Feature Pyramid Network (FPN) by adding additional layers to the feature extraction network. The second strategy integrates some attention mechanisms including the Convolutional Block Attention Module (CBAM) [19] to further improve feature extraction capabilities. The effectiveness of the enhanced networks are validated using a wear debris image dataset constructed specifically for this study.
The rest of this paper is organized as follows: Section 2 introduces the research framework, including the construction of the debris dataset, segmentation model training, and debris quantitative analysis. Section 3 provides a detailed description of the improvements made to the feature extraction network of Mask R-CNN. Section 4 outlines the model training process, presents the experimental results, and discusses the findings. Finally, Section 5 summarizes the conclusions.

2. Intelligent Recognition Method of Wear Debris Image Based on Improved Mask R-CNN and Wear Debris Area Calculation

The intelligent recognition method for wear debris images proposed in this paper consists of three key steps: constructing the debris dataset, training the segmentation model, and quantitatively analyzing the wear debris images. The recognition process is illustrated in Figure 5:
The detailed steps are listed as follows: Step 1 involves labeling wear debris images from Magnetic Chip Detector (MCD) reports to construct the debris image dataset. Step 2 entails using this dataset to train the neural network based on the improved Mask R-CNN. Step 3 focuses on recognizing wear debris images with the trained network and performing quantitative analysis by calculating the number and area of the debris.
For clarity, the structure of the improved Mask R-CNN and the process of quantitative debris analysis are described in the following subsections, respectively.

2.1. Improved Mask R-CNN Structure

The network structure of Mask R-CNN, as shown in Figure 6, primarily consists of three components: the backbone network, the Region Proposal Network (RPN), and the RoIAlign layer. The backbone network employs a Feature Pyramid Network (FPN) to extract feature information from the target image and generate multi-scale feature maps. The RPN generates candidate regions from the target image, which are then mapped onto the feature maps and passed into the Region of Interest Align (RoIAlign) layer. In RoIAlign, bilinear interpolation is used to obtain feature vectors with fixed dimensions. These feature vectors are then fed into the fully connected layer (FC Layer) for predicting the target’s bounding box offset (Box Regression) and classification. Simultaneously, the mask branch performs instance segmentation on the Region of Interest (ROI), determining whether each pixel in the region belongs to the target. This approach significantly improves the accuracy of pixel-level recognition and classification.
The Mask R-CNN model has powerful pixel-level classification capabilities, but it is typically trained on standard public datasets, which provide a sufficient amount of data for training and help avoid issues such as misdiagnosis or overfitting caused by insufficient data. However, in practice, acquiring ferrography wear debris images is challenging, making it difficult to obtain a large volume of labeled data. As a result, wear debris image segmentation is often treated as a small-sample learning task. To address the challenges of small-sample wear debris image segmentation, two main approaches are commonly employed: data augmentation and specialized algorithms.
In terms of data augmentation, this paper applied the Sobel operator to process wear debris images, enhancing the clarity of the debris edges (Section 4.1). On the algorithm side, the backbone of the Mask R-CNN model has been improved to capture more detailed wear debris features, thereby enhancing detection accuracy (Section 3). The structure of the improved network is shown in Figure 7. The specific improvements are as follows:
(1)
The CBAM attention mechanisms are added after the Resnet network, allowing the model to better focus on the location and channel features of the debris region. This approach effectively reduces the required number of training epochs and enhances detection accuracy.
(2)
Based on the concept of multi-layer feature fusion, an additional set of feature maps is stacked and fused on top of the original FPN network. The new FPN network further integrates the wear debris features extracted from multiple layers of convolution (Resnet) into multiple feature maps, enabling more accurate differentiation of debris edge contours.

2.2. Qualitative and Quantitative Analysis

The size, quantity, and concentration of debris are directly linked to the engine’s wear condition. To assess the engine’s wear status, a quantitative analysis of the debris is essential. The most intuitive quantitative parameter that can be derived from ferrography images is the debris coverage area [20]. Therefore, quantifying wear severity can be achieved by detecting the coverage area of the debris.
The traditional method for calculating the wear debris coverage area involves applying filtering and binarization techniques to ferrography wear debris images. This approach requires manual selection of an appropriate threshold, complicating automatic analysis and leading to reduced accuracy. Additionally, the scale of wear debris images is often non-uniform (see Figure 8), making it difficult to automatically determine the actual coverage area of the wear debris.
When analyzing images with a computer, the images are composed of pixels, and image operations are therefore based on pixel manipulation. To automatically calculate the actual wear debris area, it is necessary to convert the pixel area of the debris into real-world dimensions. The process involves the following steps:
(1)
The Improved Mask R-CNN is applied to segment the wear debris contours in the ferrography images.
(2)
The pixel area of the wear debris is calculated, and a conversion relationship between pixel values and physical dimensions is established.
(3)
The dimensions of the wear debris particles are then converted from pixel units to actual sizes.
(4)
The external ellipses of the wear debris contours are computed, and the physical characteristics of the debris are described using the major and minor axes of the ellipses.
(5)
The original image is annotated with the number of wear debris particles, their physical characteristics (Major Axis, Minor Axis, Aspect Ratio, and Area), and their types.
The conversion from pixel values to physical dimensions is illustrated in Figure 9a. The OCR-recognized dimensions are standardized to millimeter (mm) units and then divided by the length of the scale bar. Additionally, since most wear debris contours are irregular, they are fitted with ellipses for easier qualitative analysis (Figure 9b). The physical characteristics of the debris are described using three parameters: Major Axis, Minor Axis, and Aspect Ratio. Here, the major and minor axes correspond to the major and minor axes of the ellipse, and the Aspect Ratio is the ratio of the Major Axis to the Minor Axis. Based on these characteristics, it is possible to analyze the type of wear debris particles and perform qualitative analysis.
The actual coverage area of wear debris is calculated using the information obtained from the segmentation performed by the Improved Mask R-CNN model and the image calibration. This approach enables the automatic analysis of lubricating oil wear debris in ferrography images by automatically determining the coverage area of each wear debris particle. The test results are shown in Figure 10.

3. Optimization of Image Segmentation for Lubricating Oil Wear Debris

3.1. FPN Structure Improvement

The backbone network of Mask R-CNN consists of ResNet101 and FPN. ResNet101, a widely used residual neural network, was developed to address the vanishing and exploding gradient issues that can arise with increasing model depth. The multi-scale feature maps from images are extracted through a series of convolutional and pooling layers. The features extracted from layers with fewer convolutions are referred to as shallow features, primarily representing the color characteristics of the image. In contrast, the features from deeper layers are known as deep features, which capture more abstract semantic information. The FPN processes feature maps with different sizes by normalizing them to a uniform size and fusing them to generate multi-scale features.
In the task of detecting wear debris in lubricating oil, the wear debris regions occupy a significant portion of the image, and the edges are often indistinct. As shown in Figure 11, the shallow and deep features extracted from the processing of ferrography wear debris images by ResNet101 exhibit distinct characteristics. The shallow features primarily represent the color of the wear debris, while the deep features mainly capture the edges of the debris. However, in the shallow feature maps, the boundaries between adjacent debris particles are unclear, leading to a higher likelihood of misclassification. In contrast, the deep feature maps show incomplete debris edges. Therefore, to accurately identify the wear debris in the image, it is essential to deeply fuse both shallow and deep features.
The structure of the FPN is illustrated in Figure 12a. In the original FPN architecture, the deep feature layer P5 is derived solely from multiple convolutional layers of ResNet101, without integration with other feature layers. To address this, Liu et al. proposed adding a new FPN layer to deeply fuse high- and low-level feature layers of the image [21]. In response, this study further integrates the four feature maps produced by the FPN. The shallow features undergo an additional 2D convolution and are then fused with adjacent feature maps. The newly formed feature maps, M2, M3, M4, and M5, offer a more comprehensive fusion of multi-scale features in the debris images. The revised FPN network structure is shown in Figure 12b.

3.2. Attention Mechanism

Ferrography wear debris images are grayscale, with shallow features providing the primary information. To enhance the network’s feature extraction capability, an attention mechanism is integrated into the network. Prominent attention mechanisms in computer image processing include Squeeze-and-Excitation Networks (SENet) [22], Efficient Channel Attention (ECA) [23], CBAM [19], and others.
SENet is a channel attention mechanism that performs global pooling on images, utilizes two fully connected layers to derive channel weights, and then multiplies these weights with the original image channels. ECANet improves upon SENet by replacing the two fully connected layers with 1D convolutional layers, adjusting the channel consideration for each weight based on the convolution kernel size. Despite these advancements, both SENet and ECANet primarily focus on channel attention mechanisms, which may not fully capture spatial information in images.
CBAM addresses this limitation by integrating both channel and spatial attention mechanisms. A study demonstrates that CBAM outperforms both SENet and ECANet [19]. CBAM operates as a sequential combination of a channel attention module and a spatial attention module. The structure of CBAM is shown in Figure 13.
The Convolutional Block Attention Module (CBAM) systematically enhances feature discriminability through a dual-attention mechanism, formally expressed as
F = M S ( M C ( F ) F ) M C ( F ) F
where
F R H × W × C : Input feature map (Height × Width × Channels);
M C R 1 × 1 × C : Channel attention map;
M S R H × W × 1 : Spatial attention map;
: Hadamard product (element-wise multiplication);
H , W , C : Spatial height, width, and channel dimensions.
The specific formula for the attention mechanism is as follows:
(a)
Channel Attention Submodule:
M C ( F ) = σ ( W 1 ( W 0 ( F avg c ) ) + MLP ( W 1 ( W 0 ( F max c ) )
where
F avg c = 1 H × W i = 1 H j = 1 W F ( i , j , : ) : (Channel-wise average pooling);
F max c = max i , j F ( i , j , : ) (Channel-wise max pooling);
W 0 R C × C / r , W 1 R C / r × C : MLP weights with reduction ratio r = 16 ;
σ : Sigmoid activation.
(b)
Spatial Attention Submodule:
M S ( F ) = σ ( f 7 × 7 ( [ F avg S ; F max S ] ) )
where
F avg S = 1 C k = 1 C F ( : , : , k ) : (Spatial average pooling);
F max S = max k F ( : , : , k ) : (Spatial max pooling);
f 7 × 7 : 7 × 7 convolution with kernel weights W S ;
[ , ] : Channel concatenation.
The attention mechanism is strategically embedded into Mask R-CNN’s feature hierarchy through these implementation phases, as shown in Table 1:

4. Experimental Results

The program for recognizing ferrography wear debris images is implemented in Python (3.7) using TensorFlow-gpu (2.2.0) as the deep learning framework. The computer configuration used for this analysis includes an Intel Core i7-12700F CPU and an NVIDIA GeForce RTX 3060 12GB GPU.
To conserve computational resources and reduce training time, pre-trained weights and frozen training are utilized based on transfer learning principles. Improper configuration of the initial model parameters can significantly impact the model’s performance, leading to challenges in effectively extracting image features, prolonged training times, or even failure to converge. Research has shown that pre-trained models are generally effective across various datasets [24]. The primary advantage of pre-trained weights is that they provide an efficient feature extraction network, which transforms images into computer-recognizable features. To preserve the integrity of the initial feature extraction network, frozen training is employed to lock the weights of this network. This approach not only reduces video memory usage but also enhances training efficiency. Once frozen training is completed, defrost training is performed to adjust the weights of the entire network.

4.1. Annotation and Construction of Lubricating Oil Wear Debris Dataset

A notable validity discrepancy exists in current tribological research data sources: While most laboratory studies employ electrically driven controlled wear simulators to generate particulate samples-standardized conditions, their energy conversion paradigm (electrical→mechanical) fundamentally differs from the thermodynamic cycle (Brayton cycle) governing aero-turbofan engines. In contrast, airline MCD reports capture native wear particles directly generated by critical tribopairs like high-pressure compressor bearings and turbine shaft seals through in-service oil monitoring. These particles form through authentic mechanisms (fretting/adhesive wear under TLA-dynamic conditions with combustion-driven thermal stresses), accurately reflecting material degradation in operational environments. Consequently, our research prioritizes extracting ferrography images from airline operational MCD reports to construct an aerodynamically relevant wear signature dataset with high operational fidelity.
In this paper, a total of 147 wear debris images were extracted from more than 50 MCD reports, with each image having a resolution of 332 × 258 pixels. The original images contain metadata such as image names, scales, and annotation information (e.g., the white annotations shown in the left figure of Figure 11), which can negatively impact neural network training and detection performance. To address this, the original images were cropped to 213 × 213 pixels, and the ‘Inpaint’ function in OpenCV was used to remove redundant information and repair the images. The results of the cropping and repair process are shown in Figure 14.
Due to insufficient pixel density and unclear details in the original images, image enhancement is necessary. To improve edge features and facilitate annotation, the Laplacian operator and Sobel operator are employed to sharpen the images. The Laplacian operator works based on the following principle: First, the image is converted to grayscale, resulting in a matrix F with dimensions (Width: w; Height: h), where each element represents the pixel value at that point. The formula for Laplacian sharpening is given by (4):
G ( x , y ) = F ( x + 1 , y ) + F ( x 1 , y ) + F ( x , y + 1 ) + F ( x , y 1 ) 4 F ( x , y )
In this context, F x , y denotes the value of the element located at row y and column x in matrix F, while G x , y represents the value of the element at row y and column x in matrix G. The sharpened image is obtained by converting matrix G into a grayscale image.
The Sobel operator is a template matching technique that uses a 3 × 3 kernel. This kernel is applied to a 3 × 3 window within the image, and the results are computed. The maximum value from these results is then output as the sharpening result for the central element of the window. Unlike the gradient-based Laplacian operator, the Sobel operator not only sharpens the image but also smooths out noise, offering improved performance. The Sobel operation is described by the Formulas (5) and (6):
G x = 1 0 1 2 0 2 1 0 1 × A , G y = 1 2 1 0 0 0 1 2 1 × A
G ( x , y ) = G x 2 + G y 2
Here, G x , y denotes the value of the element located at row y and column x in matrix G. Matrix A represents the grayscale version of the original image. The sharpening result is obtained by converting matrix G into a grayscale image.
To facilitate programming and simplify the calculation process, the square root is used to approximate the sum of absolute values, as illustrated in Equation (7):
G ( x + y ) = G x + G y
The sharpening results of the images using the Laplacian and Sobel operators are presented in Figure 15. It is evident that the Sobel operator produces superior sharpening results compared to the Laplacian operator. Therefore, the Sobel operator is used for image sharpening in this study.
After image processing, the wear debris contours in the ferrography images were annotated under expert guidance to construct the dataset, as shown in Figure 16.
The marked results of the annotated dataset reveal that 147 wear debris images collectively contain over 2500 individual wear particles, a quantity sufficient for preliminary neural network training. To enhance model robustness and mitigate overfitting risks, systematic data augmentation was implemented on the wear debris image dataset. The augmentation protocol included geometric transformations (random cropping and rotation) combined with noise injection operations, as schematically illustrated in Figure 17. For each original image, one stochastic transformation was algorithmically selected to generate novel synthetic samples. This approach effectively expanded the training set by a factor of five, resulting in a final augmented dataset comprising 735 distinct wear particle images.

4.2. Evaluation Indicator

The performance evaluation of instance segmentation models requires a multidimensional metric system, given their dual functionality in object detection and semantic segmentation. Our proposed evaluation framework comprises the following core components:
1. Object Detection Metrics
Mean Intersection over Union (MIoU): Quantifies spatial alignment between predicted and ground truth bounding boxes. The schematic diagrams of P∩G and P∪G are shown in Figure 18,
MIoU = 1 N n = 1 N P n G n P n G n
where P n denotes the i-th predicted box, G n denotes the corresponding ground genuine box, and N denotes the total instances.
Average Precision (AP) and Average Recall (AR): AP represents the proportion of true positive samples among the predicted positive samples; AR represents the proportion of correctly detected samples among all actual positive samples.
AP = T P T P + F P
AR = T P T P + F N
where true positives (TPs) mean positive instances correctly classified as positive. False positives (FPs) mean negative instances misclassified as positive. False negatives (FNs) mean positive instances misclassified as negative.
2. Semantic Segmentation Metric
Mean Pixel Accuracy (MPA): MPA is the average of per-class pixel accuracy.
MPA = 1 k i = 1 k p i i j = 1 k p i j
In the above formulas, the parameter k represents the number of target types to be detected. The true value is denoted as I, and the predicted value is denoted as j. Here, p i j signifies that the i-th category is predicted as the j-th category.

4.3. Comparison of Experimental Results

To validate the effectiveness of the Improved Mask R-CNN models proposed in this paper, experiments were conducted on the ferrography wear debris image dataset. The specific training parameters are shown in Table 2:
The detection performance of the improved network was compared with that of the original Mask R-CNN network and other instance segmentation methods, such as Yolact [25], YOLOv8 [26], and YOLOv11 [27]. The model’s accuracy was assessed using MIoU, PA, PR, and MPA metrics, while the detection speed was evaluated by the common index FPS (Frames Per Second). Additionally, the ablation experiments were conducted to determine the most suitable attention mechanism for wear debris image segmentation. The image segmentation results are shown in Figure 19, and the performance evaluation results are summarized in Table 3.
As shown in Table 3, the improved neural network demonstrates higher detection accuracy compared to the original model, albeit with a slightly reduced detection speed. This decrease in speed can be attributed to the introduction of a bottom-up path augmentation in the FPN structure, which enhances the entire feature pyramid by incorporating precise localization signals from lower layers, thereby increasing the network’s depth.
Additionally, the Yolact network, a representative single-stage instance segmentation model, was tested for comparison. YOLO11 exhibited a significantly faster detection speed, reaching 88.3 FPS, which makes it suitable for real-time detection. However, the detection accuracy is noticeably lower than that of the two-stage Mask R-CNN network.
Wear debris image detection primarily serves as a technical foundation for ferrography analysis and is typically conducted offline. In this context, the segmentation accuracy is of paramount importance to accurately identify the parameters such as debris size, shape, and type, while the requirement for detection speed is comparatively lower. Thus, the Mask R-CNN model as the foundational network for wear debris detection provides notable advantages in terms of accuracy and reliability.
Additionally, the combination of the backbone network and the attention mechanism was optimized, with the Improved Mask R-CNN model integrated with CBAM achieving the best performance. The proposed configuration achieved an MIoU of 84.89% with a MAP of 87.63%. Figure 19 demonstrates that the Improved Mask R-CNN model exhibits enhanced capability in distinguishing adjacent wear debris.
To assess the effectiveness of Sobel sharpening in wear debris detection, we first constructed two datasets containing pre-sharpened and post-sharpened debris images with identical mark data. Using the training parameters specified in Table 2, we trained detection models on both datasets. The evaluation metrics employed in this study include MIoU, PA, PR, and MPA. The comparative results are presented in Table 4.
These experimental results demonstrate that applying the Sobel operator to wear debris images significantly enhances the detection accuracy of Improved Mask R-CNN model with CBAM.
The evaluation metrics discussed above primarily assess the classification performance of neural networks. However, identifying the quantity of wear debris is also crucial for quantitative analysis. Therefore, additional evaluation metrics are introduced to measure the neural network’s ability to distinguish the number of wear debris particles. These metrics include precision rate (PR), miss rate (MR), and false rate (FR). The calculation formulas for PR, MR, and FR are provided in Equations (12), (13), and (14), respectively.
P R = p 2 p 1
M R = p 3 p 1
F R = p 4 p 1
In the above equations, p 1 represents the total number of wear debris across all images; p 2 denotes the number of wear debris correctly detected; p 3 indicates the number of wear debris that were missed; p 4 refers to the number of wear debris that were incorrectly identified.
A total of 80 ferrography wear debris images were selected for detection, containing a total of 2398 debris particles. The detection results are presented in Table 5.
Table 5 and Figure 19 show that the original Mask R-CNN model exhibits a high miss rate and lacks sufficient capability to segment and identify adjacent wear debris. In contrast, the Improved Mask R-CNN model demonstrates a significant reduction in miss rate, with a precision rate reaching 97.70%, which is approximately 15% higher than that of the original model. Additionally, the improved model shows enhanced detection capability for adjacent wear debris.

4.4. Qualitative and Quantitative Analysis of Wear Debris

The relationship between wear debris types and their physical characteristics is shown in Table 6 [7]. It is well known that different types of wear debris exhibit distinct sizes and features. Moreover, due to the varying scales of ferrography images, debris with the same pixel size may differ in actual physical dimensions. While neural network methods can effectively segment wear debris contours, they are insufficient for accurate classification. Therefore, this study first employs neural network methods to segment the wear debris contours and then classifies them in terms of the characteristics listed in Table 3. The classification results are shown in Figure 20.
The classification accuracy achieved by the proposed wear debris detection method is 97.65%, which is close to the detection accuracy of debris quantity (Table 5), indicating the reliability of this classification method. Figure 21 shows the proportion of various types of wear debris. The occurrence of different types of wear debris is closely related to specific damage mechanisms in the engine.
Rolling wear debris accounts for the largest proportion at 61.6%, possibly indicating high-load operating conditions of the engine bearings, which could result from prolonged operation or inadequate maintenance. Sliding wear debris represents 25.6%, a relatively high percentage, likely related to sliding friction under specific engine operating conditions. Cutting wear debris constitutes the smallest proportion at only 2.0%. The low presence of cutting wear debris suggests good manufacturing quality and operational status of the engine, though further analysis is needed to rule out potential material or design flaws.
Interestingly, under normal conditions, wear debris caused by normal wear should be the most abundant. However, in this study, they account for only 10.8%. This discrepancy is likely because the dataset originates from MCD reports, where the engines had already experienced significant wear, leading to a substantial increase in abnormal debris.
In conclusion, the proposed wear debris image detection method combined with neural network-based segmentation and threshold-based classification demonstrates excellent performances in accurately identifying the types and sizes of wear debris. Additionally, the proportions and size ranges of different wear debris provide valuable support for future experiments.

5. Conclusions and Outlook

5.1. Conclusions

This paper presented a wear debris image recognition method based on the Improved Mask R-CNN. First, ferrography wear debris images from airline MCD reports are processed and annotated to create a lubricating oil wear debris dataset. Second, the Mask R-CNN feature extraction network, based on the FPN structure, is enhanced and an attention mechanism is introduced to develop a more reliable wear debris detection method. This improvement increases detection accuracy and reduces the miss rate. Finally, the method performs quantitative analysis by calculating the wear debris coverage area and classifies the debris based on their features, providing technical support for the automatic recognition of ferrography wear debris images.

5.2. Outlook

While the improved Mask R-CNN method presented in this paper has demonstrated significant advancements in wear debris image recognition, several promising avenues for future research and development remain. One potential area of exploration is the integration of more advanced attention mechanisms, such as multi-modal attention or transformer-based attention, which could further enhance the model’s ability to focus on critical features and improve detection accuracy. Additionally, expanding the dataset to include a wider variety of wear debris types and conditions—such as those from different lubrication systems or under various operating conditions—would strengthen the model’s generalization capabilities and robustness.
Another important direction is the optimization of the model’s computational efficiency. As the complexity of deep learning models increases, reducing inference time and memory usage becomes crucial for real-time applications, especially in resource-constrained environments such as mobile devices or embedded systems. Techniques such as model compression, pruning, and quantization could be employed to achieve this goal.
Lastly, incorporating additional contextual information, such as temporal data or metadata from the lubrication system, could enhance the model’s interpretability and decision-making capabilities. This would enable a more comprehensive understanding of wear debris patterns and their implications, ultimately contributing to more effective maintenance and reliability strategies.

Author Contributions

Conceptualization, Y.W.; Resources, Q.W.; Data curation, Y.L.; Writing—original draft, X.X.; Writing—review & editing, Y.W.; Project administration, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

The research was funded by the Xiamen Major Science and Technology Program Project (Future Industry Field) (No. 3502Z20241035), the National Natural Science Foundation of China (Grant No. 51975494) and the Fundamental Research Funds for the Central Universities (No. 20720180120).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

Author Qing Wang was employed by the company Xiamen Airlines. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Wang, Y.; Lin, T.; Wu, D.; Zhu, L.; Qing, X.; Xue, W. A New In Situ Coaxial Capacitive Sensor Network for Debris Monitoring of Lubricating Oil. Sensors 2022, 22, 1777. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, Y.; Wu, D.; Zhu, L. Progress on on-line sensing technology for wear debris in lubricant. J. Electron. Meas. Instrum. 2021, 35, 73–83. [Google Scholar]
  3. Ma, W.; Sun, Y.; Zeng, Z. Analysis of different size wear debris in oil by electrical capacitance tomography. J. Electron. Meas. Instrum. 2017, 31, 1984–1990. [Google Scholar]
  4. Chen, Z.; Zuo, H.; Zhan, Z.; Zhang, Y.; Cai, J. Study of oil system oil-line debris electrostatic monitoring technology. Acta Aeronaut. Et Astronaut. Sinica 2012, 33, 446–452. [Google Scholar]
  5. Zhu, L.; Xiao, X.; Wu, D.; Wang, Y.; Qing, X.; Xue, W. Qualitative Classification of Lubricating Oil Wear Particle Morphology Based on Coaxial Capacitive Sensing Network and SVM. Sensors 2022, 22, 6653. [Google Scholar] [CrossRef] [PubMed]
  6. Zhu, X.; Zhong, C.; Zhe, J. Lubricating oil conditioning sensors for online machine health monitoring—A review. Tribol. Int. 2017, 109, 473–484. [Google Scholar] [CrossRef]
  7. Anderson, D.P. Wear Particle Atlas: Revised; Telus: Singapore, 1982. [Google Scholar]
  8. Wu, T.; Wu, H.; Du, Y.; Kwok, N.; Peng, Z. Imaged wear debris separation for on-line monitoring using gray level and integrated morphological features. Wear 2014, 316, 19–29. [Google Scholar] [CrossRef]
  9. Cao, Y.; Xie, X. Wear Particles Classification Based On Wavelet Transform and Back-Propagation Neural Network. In Proceedings of the Multiconference on Computational Engineering in Systems Applications, Beijing, China, 4–6 October 2006; IEEE: New York, NY, USA, 2006; Volume 2. [Google Scholar]
  10. Shi, X.; Cui, C.; He, S.; Xie, X.; Sun, Y.; Qin, C. Research on recognition method of wear debris based on YOLO V5S network. Ind. Lubr. Tribol. 2022, 74, 488–497. [Google Scholar] [CrossRef]
  11. Jia, F.; Wei, H.; Sun, H.; Song, L.; Yu, F. An object detection network for wear debris recognition in ferrography images. J. Braz. Soc. Mech. Sci. Eng. 2022, 44, 67. [Google Scholar] [CrossRef]
  12. Jia, F.; Wei, H. FWDNet: A Novel Recognition Network for Ferrography Wear Debris Image Analysis. Wirel. Commun. Mob. Comput. 2022, 2022, 6511235. [Google Scholar] [CrossRef]
  13. Fan, S.; Zhang, T.; Guo, X.; Zhang, Y.; Wulamu, A. Wpc-ss: Multi-label Wear Particle Classification Based on Semantic Segmentation. Mach. Vis. Appl. 2022, 33, 43. [Google Scholar] [CrossRef]
  14. He, L.; Wei, H. CBAM-YOLOv5: A Promising Network Model for Wear Particle Recognition. Wirel. Commun. Mob. Comput. 2023, 2023, 2520933. [Google Scholar] [CrossRef]
  15. Peng, Y.; Cai, J.; Wu, T.; Cao, G.; Kwok, N.; Peng, Z. WP-DR-net: A novel wear particle detection and recognition network for automatic ferrograph image analysis. Tribol. Int. 2020, 151, 106370. [Google Scholar] [CrossRef]
  16. Liu, X.; Wang, J.; Sun, K.; Cheng, L.; Wu, M.; Wang, X. Semantic segmentation of ferrography images for automatic wear particle analysis. Eng. Fail. Anal. 2021, 122, 105268. [Google Scholar] [CrossRef]
  17. Fan, B.; Wang, Z.; Feng, S.; Wang, J.; Peng, W. An improved YOLOv8-CGA-ASF-DBB method for multi-class wear debris recognition of online visual ferrograph image. Meas. Sci. Technol. 2024, 35, 126123. [Google Scholar] [CrossRef]
  18. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 386–397. [Google Scholar] [CrossRef] [PubMed]
  19. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  20. Kirk, T.B.; Panzera, D.; Anamalay, R.V.; Xu, Z.L. Computer image analysis of wear debris for machine condition monitoring and fault diagnosis. Wear 1995, 181, 717–722. [Google Scholar] [CrossRef]
  21. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  22. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  23. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; 11534–11542. [Google Scholar]
  24. Zhu, Z.; Lin, K.; Jain, A.K.; Zhou, J. Transfer Learning in Deep Reinforcement Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 13344–13362. [Google Scholar] [CrossRef] [PubMed]
  25. Bolya, D.; Zhou, C.; Xiao, F.; Li, Z. YOLACT: Real-time instance segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9157–9166. [Google Scholar]
  26. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8. (AGPL-3.0). 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 1 March 2025).
  27. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLO11. (AGPL-3.0). 2024. Available online: https://github.com/ultralytics/ultralytics (accessed on 3 March 2025).
Figure 1. Relation diagram of abrasive particle size, concentration, and wear status of rotating parts [6].
Figure 1. Relation diagram of abrasive particle size, concentration, and wear status of rotating parts [6].
Lubricants 13 00208 g001
Figure 2. MCD report (left: ferrographic microscopy; right: spectrometric analysis).
Figure 2. MCD report (left: ferrographic microscopy; right: spectrometric analysis).
Lubricants 13 00208 g002
Figure 3. Ferrography image with adjacent wear debris.
Figure 3. Ferrography image with adjacent wear debris.
Lubricants 13 00208 g003
Figure 4. (a): Original; (b): object detection; (c): semantic segmentation; (d): instance segmentation (Red color: wear debris).
Figure 4. (a): Original; (b): object detection; (c): semantic segmentation; (d): instance segmentation (Red color: wear debris).
Lubricants 13 00208 g004
Figure 5. Intelligent recognition process of wear debris image.
Figure 5. Intelligent recognition process of wear debris image.
Lubricants 13 00208 g005
Figure 6. Mask R-CNN network structure.
Figure 6. Mask R-CNN network structure.
Lubricants 13 00208 g006
Figure 7. Improved Mask R-CNN network structure.
Figure 7. Improved Mask R-CNN network structure.
Lubricants 13 00208 g007
Figure 8. Wear debris images with different scales. (The black text “电子图像” at the top of the image denotes the electronic image, while the white text “谱图” in the middle of the image indicates the spectrogram. Both are annotation information from the original ferrograph image).
Figure 8. Wear debris images with different scales. (The black text “电子图像” at the top of the image denotes the electronic image, while the white text “谱图” in the middle of the image indicates the spectrogram. Both are annotation information from the original ferrograph image).
Lubricants 13 00208 g008
Figure 9. Image calibration results. (a): Image scale conversion; (b): Wear Debris ellipse fitting. (The black text “电子图像” at the top of the image (a) denotes the electronic image, while the white text “谱图” in the middle of the image (a) indicates the spectrogram. Both are annotation information from the original ferrograph image).
Figure 9. Image calibration results. (a): Image scale conversion; (b): Wear Debris ellipse fitting. (The black text “电子图像” at the top of the image (a) denotes the electronic image, while the white text “谱图” in the middle of the image (a) indicates the spectrogram. Both are annotation information from the original ferrograph image).
Lubricants 13 00208 g009
Figure 10. Results of intelligent detection and quantitative analysis of wear debris. (The black text “电子图像” at the top of the left image denotes the electronic image, while the white text “谱图” in the middle of the left image indicates the spectrogram. Both are annotation information from the original ferrograph image).
Figure 10. Results of intelligent detection and quantitative analysis of wear debris. (The black text “电子图像” at the top of the left image denotes the electronic image, while the white text “谱图” in the middle of the left image indicates the spectrogram. Both are annotation information from the original ferrograph image).
Lubricants 13 00208 g010
Figure 11. Wear debris feature map extracted based on Resnet101. (Left: shallow feature; Right: deep feature).
Figure 11. Wear debris feature map extracted based on Resnet101. (Left: shallow feature; Right: deep feature).
Lubricants 13 00208 g011
Figure 12. FPN network structure. (Red dashed lines: original feature propagation paths; blue dashed lines: new feature propagation paths).
Figure 12. FPN network structure. (Red dashed lines: original feature propagation paths; blue dashed lines: new feature propagation paths).
Lubricants 13 00208 g012
Figure 13. CBAM structure [19].
Figure 13. CBAM structure [19].
Lubricants 13 00208 g013
Figure 14. Image of lubricating oil wear debris. (Left: original image; Right: image after cutting and repair). (The black text “电子图像” at the top of the left image denotes the electronic image, while the white text “谱图” in the middle of the left image indicates the spectrogram. Both are annotation information from the original ferrograph image).
Figure 14. Image of lubricating oil wear debris. (Left: original image; Right: image after cutting and repair). (The black text “电子图像” at the top of the left image denotes the electronic image, while the white text “谱图” in the middle of the left image indicates the spectrogram. Both are annotation information from the original ferrograph image).
Lubricants 13 00208 g014
Figure 15. Image sharpening results ((a): Original, (b): Laplacian operator, (c): Sobel operator).
Figure 15. Image sharpening results ((a): Original, (b): Laplacian operator, (c): Sobel operator).
Lubricants 13 00208 g015
Figure 16. Marked results (Up: original images; Down: marked images). (Label the contours of abrasive particles with distinct colors to construct an instance segmentation dataset capable of differentiating objects of the same class).
Figure 16. Marked results (Up: original images; Down: marked images). (Label the contours of abrasive particles with distinct colors to construct an instance segmentation dataset capable of differentiating objects of the same class).
Lubricants 13 00208 g016
Figure 17. (a): Original image; (b): cropping; (c): rotation; (d): Gaussian noise.
Figure 17. (a): Original image; (b): cropping; (c): rotation; (d): Gaussian noise.
Lubricants 13 00208 g017
Figure 18. MIoU schematic diagram.
Figure 18. MIoU schematic diagram.
Lubricants 13 00208 g018
Figure 19. Instance segmentation results: (a) Original, (b) Mask RCNN, (c) Improved Mask R-CNN. (The red circles in the figure indicate overlapping wear debris. A comparative demonstration highlights the advantage of the proposed method in segmenting overlapping wear debris).
Figure 19. Instance segmentation results: (a) Original, (b) Mask RCNN, (c) Improved Mask R-CNN. (The red circles in the figure indicate overlapping wear debris. A comparative demonstration highlights the advantage of the proposed method in segmenting overlapping wear debris).
Lubricants 13 00208 g019
Figure 20. Classification results of wear particles.
Figure 20. Classification results of wear particles.
Lubricants 13 00208 g020
Figure 21. Proportion of different types of wear debris.
Figure 21. Proportion of different types of wear debris.
Lubricants 13 00208 g021
Table 1. CBAM integration in Mask R-CNN architecture.
Table 1. CBAM integration in Mask R-CNN architecture.
Integration StagePositionImplementation DetailsOutput Dimension
Backbone (ResNet)After each residual blockInsert CBAM after final convolution:
X o u t = CBAM ( F ( X i n ) ) + X i n
256 × 256 × 256
Feature Pyramid NetworkLateral connectionsApply CBAM before feature fusion:
P i = CBAM ( Conv ( C i ) ) + C i
[M2-M5] scales
Region Proposal NetworkBefore classification layerSpatial attention for proposal scoring:
S r p n = M S Conv ( F )
14 × 14 × 512
RoI HeadsPost RoIAlign poolingChannel attention for mask prediction:
F r o i = M C FC ( F p o o l e d )
7 × 7 × 2048
Table 2. Training parameters.
Table 2. Training parameters.
ParametersValueParametersValue
Epoch200OptimizerStochastic Gradient Descent (SGD)
Batch size8Momentum0.9
Initial learning rate0.04Weight decay1 × 10−4
Table 3. Comparison of experimental results.
Table 3. Comparison of experimental results.
NetworkMIoU/%PA/%PR/%MAP/%FPS
Mask R-CNN79.3477.6278.3982.598.8
Mask R-CNN + SENet80.8778.1778.9983.218.6
Mask R-CNN + CBAM82.6480.5979.0486.178.3
Mask R-CNN + ECA81.8381.2679.6285.368.5
Improved Mask R-CNN82.3581.7280.6883.747.9
Improved Mask R-CNN + SENet83.2682.7780.7185.377.8
Improved Mask R-CNN + CBAM84.8983.7381.1887.637.6
Improved Mask R-CNN + ECA83.7181.2381.4586.187.7
Yolact70.3569.5868.7471.6738.7
YOLOv870.4969.8269.7671.2877.51
YOLO1172.4370.7571.5173.4088.3
Table 4. Comparative analysis of detection models with versus without Sobel operator.
Table 4. Comparative analysis of detection models with versus without Sobel operator.
NetworkMioU/%AP/%AR/%MPA/%
Improved Mask R-CNN + CBAM without Sobel81.7482.7580.5884.19
Improved Mask R-CNN + CBAM with Sobel84.8983.7381.1887.63
Table 5. Abrasion debris test results.
Table 5. Abrasion debris test results.
NetworkMask R-CNNImproved Mask R-CNN + CBAM
p 1 23982398
p 2 19872343
p 3 38551
p 4 365
P/%82.8697.70
M/%16.052.12
F/%1.50.21
Table 6. Characteristics of different types of wear debris.
Table 6. Characteristics of different types of wear debris.
Wear TypeWear Debris Characteristics
Major AxisAspect RatioMorphology
Normal Wear<15 μm<2Flake-like
Cutting Wear>25 μm>5Elongated
Rolling Wear>15 μm<2Irregular Contour
Severe Sliding Wear>15 μm>2Striated
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiao, X.; Zhang, W.; Wang, Q.; Liu, Y.; Wang, Y. Intelligent Recognition Method for Ferrography Wear Debris Images Using Improved Mask R-CNN Methods. Lubricants 2025, 13, 208. https://doi.org/10.3390/lubricants13050208

AMA Style

Xiao X, Zhang W, Wang Q, Liu Y, Wang Y. Intelligent Recognition Method for Ferrography Wear Debris Images Using Improved Mask R-CNN Methods. Lubricants. 2025; 13(5):208. https://doi.org/10.3390/lubricants13050208

Chicago/Turabian Style

Xiao, Xiangwen, Weixuan Zhang, Qing Wang, Yuan Liu, and Yishou Wang. 2025. "Intelligent Recognition Method for Ferrography Wear Debris Images Using Improved Mask R-CNN Methods" Lubricants 13, no. 5: 208. https://doi.org/10.3390/lubricants13050208

APA Style

Xiao, X., Zhang, W., Wang, Q., Liu, Y., & Wang, Y. (2025). Intelligent Recognition Method for Ferrography Wear Debris Images Using Improved Mask R-CNN Methods. Lubricants, 13(5), 208. https://doi.org/10.3390/lubricants13050208

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop