Next Article in Journal
PLA/MWNTs Conductive Polymer Composites as Stress Sensors—The Role of Supramolecular Ordering
Previous Article in Journal
SDN-Oriented 6G Industrial IoT Architecture Design and Application to Optimal RIS Placement and Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DCRDF-Net: A Dual-Channel Reverse-Distillation Fusion Network for 3D Industrial Anomaly Detection

1
School of Computer and Information Science, College of Software, Southwest University, Chongqing 400715, China
2
Kawasaki (Chongqing) Robotics Engineering Co., Ltd., Chongqing 400700, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(2), 412; https://doi.org/10.3390/s26020412
Submission received: 7 November 2025 / Revised: 22 December 2025 / Accepted: 5 January 2026 / Published: 8 January 2026
(This article belongs to the Section Sensor Networks)

Abstract

Industrial surface defect detection is essential for ensuring product quality, but real-world production lines often provide only a limited number of defective samples, making supervised training difficult. Multimodal anomaly detection with aligned RGB and depth data is a promising solution, yet existing fusion schemes tend to overlook modality-specific characteristics and cross-modal inconsistencies, so that defects visible in only one modality may be suppressed or diluted. In this work, we propose DCRDF-Net, a dual-channel reverse-distillation fusion network for unsupervised RGB–depth industrial anomaly detection. The framework learns modality-specific normal manifolds from nominal RGB and depth data and detects defects as deviations from these learned manifolds. It consists of three collaborative components: a Perlin-guided pseudo-anomaly generator that injects appearance–geometry-consistent perturbations into both modalities to enrich training signals; a dual-channel reverse-distillation architecture with guided feature refinement that denoises teacher features and constrains RGB and depth students towards clean, defect-free representations; and a cross-modal squeeze–excitation gated fusion module that adaptively combines RGB and depth anomaly evidence based on their reliability and agreement.Extensive experiments on the MVTec 3D-AD dataset show that DCRDF-Net achieves 97.1% image-level I-AUROC and 98.8% pixel-level PRO, surpassing current state-of-the-art multimodal methods on this benchmark.

1. Introduction

With the rapid advancement of industrial automation and intelligent manufacturing, precise detection of product surface defects has become a critical technology for ensuring production quality [1]. In real industrial environments, the vast majority of samples are normal, while anomalous ones are scarce, which makes Unsupervised Anomaly Detection (UAD) trained solely on normal samples a core direction in applied research [2]. RGB-based surface anomaly detection has seen significant progress, with the performance of many methods approaching its practical limit on the widely adopted MVTec AD benchmark [2]. However, RGB is sensitive to illumination variations and lacks three-dimensional geometric information, which limits its performance in defect scenarios dominated by 3D deformations such as undulations and material loss.
With the increasing prevalence of 3D sensors in industrial settings [3] and the release of the MVTec 3D-AD dataset [4], multimodal defect detection has emerged as an important trend that leverages complementary texture and geometric information to compensate for the limitations of unimodal methods. Early multimodal approaches largely followed a memory-bank paradigm (e.g., BTF, M3DM, CPMF),constructing a repository of normal features during training and performing anomaly discrimination through feature distance retrieval during inference [5,6]. Although these methods are robust, their reliance on substantial storage and computationally intensive distance matching hinders their deployment in resource-constrained industrial environments.
Current representative technical routes for multimodal industrial anomaly detection primarily encompass reconstruction-based methods and teacher–student distillation. Reconstruction-based methods typically perform self-supervised learning with synthetic anomalies and label anomalies via reconstruction residuals; however, they often overlook cross-modal joint denoising and consistency constraints, leading to anomaly restoration and background sensitivity, which, in complex scenes, cause false positives and false negatives. Teacher–student distillation adopts a frozen teacher to extract normal representations and trains a student to regress the teacher’s outputs; yet prior work has often employed isomorphic single-branch designs for the teacher and student, wherein fused representations tend to wash out modality-specific anomaly cues, causing cross-modal interference, insufficient anomaly contrast, and reduced separability for fine-grained defects. To address these issues, we propose DCRDF-Net (Dual-Channel Reverse-Distillation Fusion Network), which learns modality-specific normality manifolds on aligned RGB and 3D data and performs detection and localization through a unified deviation measure. Our main contributions can be summarized as follows:
  • Perlin-guided perturbation and texture matching: In the foreground region, multi-frequency Perlin multiplicative perturbations driven by a shared random seed are synchronously applied to both color and depth data at corresponding spatial locations. For the RGB branch, textures are randomly sampled from the Describable Textures Dataset (DTD) for texture matching and statistically consistent fusion; for the depth branch, geometrically consistent guided smoothing is performed.
  • Dual-channel reverse distillation: Reverse distillation is employed for both the RGB and depth branches, with a Guided Feature Denoising Network inserted between the teacher and student networks. This network suppresses defect-corrupted responses in the teacher’s outputs, ensuring the student network learns a pristine representation of normal samples.
  • Cross-modal squeeze–excitation gated fusion: Channel recalibration leveraging cross-modal global statistics is integrated with depthwise-separable spatial gating. This framework learns pixel-wise weight maps based on the evidential strengths of the two student networks and their inter-modal discrepancies, thus achieving adaptive fusion of the two modalities.

2. Related Work

2.1. Two-Dimensional Industrial Anomaly Detection

Image anomaly detection aims to localize defects from a single RGB image of the target. The release of public datasets such as MVTec AD [2] and VisA [7] has substantially accelerated the development of related methods. Current mainstream approaches fall into three paradigms: reconstruction-based, normalizing-flow-based, and teacher–student distillation.
Reconstruction-based methods typically employ encoder-decoder architectures to reconstruct images or their intermediate features, using reconstruction error as the anomaly criterion. DRAEM [8] constructs a parallel architecture of reconstruction and discriminative subnetworks, simultaneously learning joint representations of anomalous images and their reconstruction results, as well as the classification boundary between them. To alleviate the coarse artifacts produced by autoencoder reconstructions, Liu et al. [9] proposed a two-stage pipeline that coordinates an Impression Extractor Network (IE-Net) with an expert network, which significantly improves the reconstruction fidelity of normal sample. Wolleb et al. [10] proposed a strategy within the diffusion model framework that combines iterative noise injection with classifier-guided image-to-image translation, enabling the generation of high-resolution anomaly heatmaps without requiring complex training. For low-light and complex illumination scenarios, some works have leveraged paired well-lit/low-light images for unsupervised learning, reconstructing well-lit features from low-light features and using the resulting inconsistency to identify anomalies, achieving superior detection and localization performance on datasets like LL-IAD [11].
Normalizing flow-based methods project the features of normal samples onto a tractable latent distribution through reversible affine transformations, and then determine the anomaly status of test samples based on their likelihood under the learned distribution. CFLOW-AD [12] integrated a pre-trained convolutional encoder with a multi-scale generative decoder under a conditional normalizing flow framework. It directly estimated the likelihood of encoded features, enabling real-time unsupervised anomaly detection and localization while maintaining memory and computational efficiency. CS-Flow [13] proposed a fully convolutional cross-scale normalizing flow architecture. Multi-scale feature maps were processed in parallel to capture both local and global context, and a spatial likelihood map was generated while preserving image structure, achieving interpretable and precise pixel-level defect localization. DifferNet [14] combined convolutional features with normalizing flows by first modeling the distribution of normal samples in a multi-scale feature space. Anomaly scores were then derived from pixel-wise likelihood differences, and robustness was improved through multiple image transformations.
Uninformed Students [15] introduced an end-to-end student–teacher distillation framework, in which a teacher network, pre-trained on large-scale natural images, extracted multi-scale local descriptors. The student network was trained to regress the teacher’s outputs, and a dense anomaly map was generated using prediction variance and regression error. Anomalies were detected when the student failed to accurately reproduce the teacher’s features on anomalous samples. IKD [16] tackled overfitting in distillation by introducing a Context Similarity Loss (CSL) to strengthen the perception of the normal data manifold structure, and employed Adaptive Hard Sample Mining (AHSM) to dynamically focus on informative challenging samples, thereby distilling more discriminative knowledge. RD++ [17] incorporated self-supervised optimal transport into a reverse distillation framework to improve the compactness of normal features. It also used simplex noise to generate pseudo-anomalous samples, reducing the student network’s sensitivity to anomalous signals and enhancing discrimination between normal and anomalous samples. In addition, for industrial surface segmentation and specific component inspection, NC-Net [18] improved pixel-level segmentation robustness via nested U-blocks and attention-based edge/mask guidance. In chip inspection scenarios, an efficient backbone combined with reverse feature matching and masking was used to achieve interpretable anomaly localization, supported by a real industrial dataset and ablation studies [19].

2.2. Three-Dimensional Industrial Anomaly Detection

2.2.1. Memory Bank Methods

In 3D anomaly detection, memory-bank-based methods have become an established technique, which identifies anomalies by comparing input data against a stored repository of features from normal samples. M3DM [5] designed a multi-3D memory model that leveraged multiple memory banks together with a hybrid feature-fusion strategy to enhance cross-modal interaction and reduce information loss. CPMF [6] aggregated multimodal features and stored the fused representation in a single memory bank, reducing the memory footprint and improving inference speed. Shape-Guided [20] proposed a shape-guided expert-learning framework that constructed bimodal memory banks for expert training and linked them via correspondences, thereby fully exploiting geometric and appearance cues. M3DM-NR [21], building on M3DM, added a two-stage denoising network to process the inputs and replaced farthest-point sampling with an aligned multi-scale feature-extraction module, improving local feature extraction for point clouds. These methods require substantial memory to construct the banks, and the resulting memory pressure slows inference, making them ill-suited to real-time industrial defect-detection scenarios.

2.2.2. Reconstruction-Based Methods

Reconstruction-based methods typically learn, on normal data, to self-supervise the reconstruction of the input modalities (RGB images and depth maps), and use the reconstruction error as the anomaly criterion. EasyNet [22] designed a dual-branch multimodal reconstruction network that independently restored RGB and depth inputs and used an information-entropy comparison strategy to decide whether to fuse depth features, achieving fast inference and deployment practicality. DBRN [23] likewise adopted a dual-branch reconstruction architecture, introduced Perlin noise on depth maps to generate anomaly masks, used grayscale images from natural-texture datasets to simulate realistic defect textures, and designed an importance-scoring module to assess and fuse the two modalities. 3DSR [24] proposed a Depth-Aware Discrete Autoencoder (DADA) that jointly modeled RGB and 3D data in a discrete latent space, improving the representation and detection of 3D surface anomalies. Uni-3DAD [25] proposed an architecture with a feature-based branch and a reconstruction-based branch to detect surface defects and missing regions, and finally fused the two outputs using an OCSVM. These approaches often prioritized self-supervised reconstruction while neglecting cross-modal joint denoising and fusion strategies, which led to false positives and false negatives in complex scenes.

2.2.3. Teacher–Student Methods

In multimodal anomaly detection, the teacher–student paradigm leverages the teacher’s strong representation of normal samples to guide the student; when a test sample deviates from normal patterns, the discrepancy between student and teacher outputs serves as an anomaly cue. AST [26] adopted an asymmetric teacher–student design, using a conditional normalizing flow as the teacher and a convolutional network as the student, and fed concatenated RGB and depth features into the model; by exploiting the bijective property of flows, it further amplified the teacher’s output disparity between normal and abnormal samples, thereby strengthening the student’s anomaly response. MMRD [27] proposed a multimodal reverse-distillation approach: on the teacher side, twin encoders extracted RGB and depth features in parallel to enable parameter-free fusion, whereas on the learnable student decoder side, cross-modal interaction and reconstruction-error computation drove the student to more accurately recover the distribution of normal samples. LPFSTNet [28] designed, for 3D industrial scenarios, a lightweight, parameter-free “head-attention” teacher–student network: the student comprised only simplified convolutions, reducing parameters and accentuating the architectural gap from the teacher to improve generalization. Existing methods commonly suffer from representation homogenization due to isomorphic teacher–student pairs and from anomaly-signal smoothing caused by single-branch fusion.
Across these three lines of work, a common limitation is that RGB and depth are either coupled too early or fused in a heuristic manner. Memory-bank methods incur substantial storage and retrieval overhead and typically treat fused features as static templates. Reconstruction-based models emphasize self-supervised restoration but rarely perform joint cross-modal denoising, which makes them prone to restoring anomalies and amplifying background clutter. Teacher–student approaches often rely on isomorphic, single-branch designs, so that fusion tends to homogenize modality-specific cues and smooth out fine-grained anomaly signals. In contrast, DCRDF-Net is explicitly designed to (i) avoid external memory banks and learn modality-specific normality manifolds under a shared frozen teacher via dual-channel reverse distillation, (ii) enrich supervision with a Perlin-guided pseudo-anomaly generator that injects appearance–geometry-consistent perturbations at aligned RGB–depth locations, and (iii) perform adaptive, pixel-wise fusion through a cross-modal squeeze–excitation gated module that exploits both unimodal evidence and cross-modal discrepancy.

3. Method

3.1. Overview of the Methodology

The overall pipeline and key module designs of the proposed DCRDF-Net are shown in Figure 1. The method begins by synchronously injecting appearance-geometry-consistent pseudo-anomalies into registered RGB-Depth pairs. Specifically, for the RGB modality, we employ multi-frequency Perlin noise—statistically aligned with external textures—to induce foreground-aware and boundary-preserving perturbations. Concurrently, structurally consistent geometric perturbations are applied to the depth channel via proportional scaling. In the feature-learning stage, a shared teacher encoder, coupled with a Guided Feature Refinement Network (GFRN), performs projection-based compression and linear-attention refinement on the perturbed multi-scale features, thereby recovering clean representations that closely approximate the defect-free prior. Subsequently, two independent student branches conduct reverse reconstruction for the RGB and depth modalities, respectively, to obtain within-modality anomaly response maps. Finally, a cross-modal squeeze–excitation gated fusion module is constructed, which leverages global channel priors and depthwise-separable spatial gating to generate pixel-wise adaptive weights.
The construction strategy for the consistency-preserving pseudo-anomalies is detailed in Section 3.2. Section 3.3 then elaborates on the teacher–student reverse-distillation network and the Guided Feature Refinement Network (GFRN). Finally, the design of the cross-modal gated fusion network is described in Section 3.4.

3.2. Perlin-Guided Percentage-Based Perturbation and Statistical Texture Alignment

To synchronously inject appearance–geometry–consistent pseudo anomalies at identical pixel coordinates, we denote the aligned RGB image and depth map by R and D, respectively. We first construct a soft-foreground mask M s from a binary foreground M f with boundary attenuation, where the edge-decay weight w edge depends exponentially on the distance to the foreground boundary dist ( M f ) , τ is the decay scale, and α is the softening coefficient:
w edge ( x , y ) = exp dist ( M f ) ( x , y ) τ .
M s ( x , y ) = min 1 , M f ( x , y ) + α w edge ( x , y ) .
Concretely, dist ( M f ) is computed using the standard Euclidean distance transform on the binary foreground mask M f , yielding for each pixel its Euclidean distance to the nearest foreground boundary. This distance is then normalized by the maximum distance inside the foreground region so that it lies in [ 0 ,   1 ] , and subsequently used in Equation (1) to generate the edge-decay weight w edge .
The texture intensity T is obtained from luminance gradients and local variance and is mapped to a normalized weight T ^ via a sigmoid. Using a single random seed s, we generate three-frequency multi-scale Perlin noise with amplitude compression, where P ( f i ; s ) is the 2D Perlin field at frequency f i , η i are mixture weights, and κ is the compression coefficient:
P ˜ = tanh κ i = 1 3 η i P ( f i ; s ) .
In practice, for 256   ×   256 inputs we generate Perlin fields on 16   ×   16 , 32   ×   32 , and 64   ×   64 grids, corresponding to three frequencies f 1 , f 2 , f 3 that cover coarse-, mid-, and fine-scale structures in the foreground. The mixture weights are fixed to ( η 1 , η 2 , η 3 ) = ( 0.5 ,   0.3 ,   0.2 ) to emphasize mid-scale components while suppressing excessive high-frequency noise, and the compression coefficient is set to κ = 1.0 . After the tanh ( · ) nonlinearity, P ˜ remains in a moderate range and provides sufficient contrast for the percentage-based modulation in Equations (6) and (7). These hyperparameters are selected once on a small held-out subset and kept fixed for all categories and experiments on the MVTec 3D-AD dataset.
Within the foreground, we construct a per-pixel probability map U from the structure field, texture weight, and boundary weight; here σ ( · ) is the sigmoid, θ is a quantile threshold, [ · ] is the indicator function, and Close 3 × 3 ( · ) denotes morphological closing. The final mask M is given by
U = σ P ˜ · T ^ · w edge ,
M = Close 3 × 3 M s · [ U > θ ] .
For the appearance branch, we select a texture image N from the DTD library and align it via affine transformation and channel-statistics matching to obtain N ^ , which matches the foreground statistics of R. Concretely, we compute the tight bounding box of M s and resize/crop the sampled DTD texture to this box with a small random scale and rotation, yielding an affine mapping from the DTD coordinates to the image plane. We then match the per-channel mean and standard deviation of the transformed texture to those of R within the foreground region so that N ^ is visually consistent with the object appearance before blending. Let the pixel transparency ρ ( x , y ) be determined by texture and gradient cues, and define the foreground modulation field S f = M · w edge · T ^ . With percentage amplitude β p for color channels, element-wise/channel-wise product ⊙, and intensity clipping clip ( · ) , we compute
R a = ( 1 M ) R + M ( 1 ρ ) clip R ( 1 + β p S f P ˜ ) + ρ N ^ ,
where α in Equation (2) is the softening coefficient for M s , while ρ ( x , y ) here denotes the per-pixel transparency. This procedure ensures that the injected texture is spatially aligned with the object region and smoothly blended with the original appearance according to the mask M and transparency map ρ ( x , y ) .
For the geometry branch, we maintain consistency with the surface structure. We inpaint to obtain D fill and guided-smooth P ˜ with D fill to produce P ˜ d . With depth-channel percentage amplitude β d , the perturbed depth is
D a = ( 1 M ) · D fill + M · D fill · 1 + β d S f P ˜ d ,
where R a and D a denote the RGB image and depth map after pseudo-anomaly injection, and M is the final mask produced by the generation process.

3.3. Dual-Channel Reverse-Distillation Network

Given the heterogeneity of defect representations across modalities, RGB images are more sensitive to color and luminance variations and can precisely capture subtle changes of this kind, whereas depth maps possess stronger discrimination for geometric relief and surface micro-topography, effectively characterizing related structural attributes. Early fusion at shallow network layers tends to cause fusion instability, whereby one modality dominates and suppresses the other. Accordingly, we adopt branch-wise independent modeling with decision-level fusion: first, we construct a dual-channel reverse-distillation network, enabling the RGB and depth branches to perform feature denoising, modality alignment, and defect detection within their respective channels; then, at the decision stage, we introduce a cross-modal squeeze–excitation gated fusion module that achieves adaptive weighted fusion via global channel statistics and pixel-wise evidence analysis. In contrast to classic teacher–student pipelines that jointly adapt the teacher on the target domain, we keep the teacher frozen and only update the denoising and student branches so that the teacher acts as a stable cross-modal semantic prior and the teacher–student discrepancy remains sensitive to local anomalies.

3.3.1. Teacher Encoder

We employ a ResNet-50 backbone, pre-trained on ImageNet and kept frozen throughout training, as the shared teacher encoder Φ for both modalities, serving as a modality-agnostic, structure-sensitive prior rather than being fine-tuned on the target dataset. The original RGB image R and depth map D (where D is first normalized and replicated along the channel dimension to form a three-channel “pseudo-RGB” input that matches the ResNet-50 input format), together with their Perlin-perturbed counterparts R a and D a , are processed by the teacher encoder to extract multi-scale features { F i 0 , D i 0 , F i a , D i a } i = 1 3 ; the clean features { F i 0 , D i 0 } provide direct supervision targets for both the guided feature denoising network and the student network while also defining a common semantic coordinate system shared by the RGB and depth branches.

3.3.2. Guided Feature Refinement Network

The guided feature refinement network (GFRN) aims to recover the teacher’s defect-free multi-scale representations, aligning the student-side pyramid with { F i 0 , D i 0 } in distribution while preserving sharp boundaries and fine details. At each scale, a stack of projection layers performs constrained denoising and distribution contraction (channel recalibration and instance normalization), pushing anomaly-induced first-order responses back onto the teacher’s low-variance submanifold and yielding clean, prior-centered features. Concretely, each projection stack is implemented as two consecutive 3   ×   3 convolutional blocks, where each block consists of a 3   ×   3 convolution followed by InstanceNorm and a LeakyReLU activation; this design focuses on locally suppressing anomaly-induced noise and shrinking the feature distribution toward the normal teacher manifold. We then use linear attentionas a non-local refiner to aggregate long-range self-similarity at the same spatial resolution; a position-aware bias and residual write-back suppress cross-boundary leakage and over-smoothing, preserving steep boundary transitions. In practice, we adopt a lightweight linear-attention block that directly operates on the feature map and is wrapped in a residual connection, so that global consistency is enhanced without introducing additional downsampling or heavy transformer layers. The purified features are denoted by F ˜ i and D ˜ i . Projection layers handle local denoising and variance reduction, whereas linear attention enforces global consistency and boundary fidelity. We instantiate two GFRNs with identical architecture but independent parameters for the RGB and depth branches, allowing each modality to adapt to its own statistics; for the depth branch, the GFRN operates on features extracted from the replicated and normalized three-channel depth input, so that stable modality-specific biases in the frozen teacher responses are implicitly absorbed by the refinement network and the depth-side student, while the remaining teacher–student discrepancy primarily reflects anomaly-induced deviations rather than deterministic cross-modal differences.
Training minimizes a scale-wise denoising-distillation loss that directly matches the denoised outputs to the teacher’s clean features:
L denoise = i = 1 3 F ˜ i F i 0 2 2 + i = 1 3 D ˜ i D i 0 2 2 .

3.3.3. Student Decoder

A ResNet-50 decoder reconstructs each modality branch. The denoised features { F ˜ i , D ˜ i } i = 1 3 first pass through a lightweight OCBE (Output Calibration and Boundary Enhancement) module for distribution regularization and bandwidth control, then are decoded to produce student reconstructions { F i S , D i S } i = 1 3 .
We compute multi-scale distances between the student’s reconstructed features and the teacher’s pristine features { F i 0 , D i 0 } , followed by upsampling and aggregation to obtain per-modality anomaly prediction maps p r and p d . To inherit the teacher’s soft distribution while achieving pixel-level alignment with anomaly masks, we employ two complementary losses: feature-level knowledge distillation L KD for channel-distribution and confidence-shape alignment, and pixel-level focal loss L focal to handle class imbalance and improve small-defect recall.
The overall student objective is:
L stu = λ KD L KD + λ seg L focal ,
where λ KD , λ seg 0 (both set to 1 during training).

3.4. Cross-Modal Squeeze–Excitation Gated Fusion

We construct a cross-modal squeeze–excitation gated fusion (CM-SEGF) module that generates pixel-wise adaptive weights via global channel recalibration and depthwise-separable spatial gating, enabling calibrated fusion of the two modality predictions. First, multi-scale features from the two student branches are upsampled and concatenated to obtain F r , F d R B × C × H × W . For each modality, global average/max pooling yields a r , m r , a d , m d R B × C . We form cross-modal descriptors z r = [ a r , m r , a d , m d ] and z d = [ a d , m d , a r , m r ] and map them through an MLP to channel gates g r , g d [ 0 ,   1 ] C , then recalibrate features via
F ˜ m = g m F m , m { r , d } .
From the recalibrated features we extract three single-channel cues—RGB evidence E r = AvgCh ( | F ˜ r | ) , depth evidence E d = AvgCh ( | F ˜ d | ) , and cross-modal discrepancy Δ H = AvgCh ( | F ˜ r F ˜ d | ) —and stack them as X = [ E r , E d , Δ H ] R B × 3 × H × W . A depthwise-separable block models spatial context and mixes channels to produce a pre-activation, and the pixel-wise weight field is
s = σ DSConv [ AvgCh ( | F ˜ r | ) , AvgCh ( | F ˜ d | ) , AvgCh ( | F ˜ r F ˜ d | ) ] [ 0 ,   1 ] B × 1 × H × W .
The fused prediction explicitly balances the two modalities:
p fused = s p r + ( 1 s ) p d ,
where p r , p d [ 0 ,   1 ] are pixel-level predictions from the RGB and depth branches, respectively. The fusion objective uses Focal Loss as the sole supervision:
L fused = λ fused L focal p fused , M , λ fused = 1 .

4. Experimental

4.1. Experimental Details

Dataset: We conducted experiments on the MVTec 3D-AD dataset [4], a 3D industrial anomaly-detection benchmark comprising 10 object categories. The training set contains 2656 normal samples, each consisting of an RGB image and the corresponding 3D scan information. The test set includes 1197 samples—948 anomalous and 249 normal—and provides pixel-level annotations for evaluating both detection and localization. In addition, MVTec 3D-AD provides a validation set with 294 normal samples.
Implementation Details: We used co-registered 256 × 256 RGB–depth pairs as inputs for both training and inference. During training, we synthesized geometrically conformal pseudo anomalies within foreground regions using a Perlin-based, texture-aligned strategy to mimic real defect distributions; at inference, raw image pairs were fed directly. The backbone was ResNet-50, and models were trained and evaluated independently per category. We optimized with Adam (learning rate 0.005, batch size 16) for a total of 150 epochs. All experiments were conducted on a single NVIDIA RTX 2080. (NVIDIA, Santa Clara, CA, USA) At inference, the network outputs anomaly heatmaps for the RGB branch, the depth branch, and their fused result.
Evaluation Metrics: To comprehensively assess the proposed method’s detection and localization performance, we adhere to the evaluation protocol established in recent studies on MVTec 3D-AD and multimodal industrial anomaly detection and use two standard metrics at the image and pixel levels. At the image level, we compute the image-level area under the ROC curve (I-AUROC) over all test samples to evaluate the model’s ability to recognize anomalous objects. I-AUROC is a threshold-independent metric that is relatively insensitive to class imbalance and has been widely adopted as the standard indicator for reporting image-level detection performance in industrial anomaly detection benchmarks. At the pixel level, we report the Per-Region Overlap (PRO) to quantify the model’s spatial accuracy in localizing anomalous regions. In contrast to purely pixel-counting measures, PRO aggregates the overlap over connected components of the ground-truth mask and is therefore more sensitive to the spatial continuity and completeness of detected defect regions.

4.2. Experimental Results and Analysis

Table 1 and Table 2 report comparisons of image-level detection (I-AUROC) and pixel-level localization (PRO) on the MVTec 3D-AD dataset.Qualitative visualizations are shown in Figure 2. Overall, the proposed method leads on both metrics: the class-wise mean I-AUROC reaches 97.1% and the class-wise mean PRO reaches 98.8%, attaining best or near-best scores on most categories and exhibiting stable, well-balanced detection–localization performance.
Image-level detection (I-AUROC): As shown in Table 1, our method achieves the highest I-AUROC on Carrot, Cookie, Foam, and Tire (99.5%, 99.3%, 98.4%, and 92.8%, respectively), while other categories remain in a high range (e.g., Bagel 99.1%, Rope 98.1%), yielding an overall mean of 97.1%. Compared with 3D methods using only point clouds or depth (e.g., DepthGAN/AE/VM), our I-AUROC improves markedly across categories; relative to representative multimodal baselines (BTF, AST, M3DM, MMRD), our approach delivers strong results on both texture-dominated (Carrot, Cookie) and geometry-dominated (Foam, Tire) targets, indicating superior adaptability to heterogeneous defect patterns. A closer comparison shows that, for Foam/Tire, gains over single-modality or statically fused models stem from a more sensitive response to geometric undulations and local concavities; for Carrot/Cookie, compared with models that reconstruct only in the pixel or depth domain, our method achieves higher separability for fine-grained texture anomalies.
Pixel-level localization (PRO): As shown in Table 2, our method attains the highest PRO on Bagel, Carrot, Cookie, and Rope (99.1%, 99.5%, 98.5%,98.8%, respectively) and ranks among the top for most other categories, resulting in a mean of 98.8%, surpassing representative methods (e.g., 97.6% for MMRD, 97.6% for Shape-Guided, and 96.4% for M3DM). Notably, in scenes with complex textures (Foam, Peach) and slender structures (Rope, Cable Gland), our approach maintains high recall while effectively suppressing false positives, producing smoother anomaly heatmaps with sharper boundaries. These results indicate that the proposed denoising distillation and adaptive fusion preserve structural consistency while emphasizing anomalous details, thereby stabilizing localization metrics across diverse object morphologies.
Runtime efficiency: We further evaluated the runtime characteristics of DCRDF-Net on an experimental platform equipped with a single NVIDIA RTX 2080 GPU (12 GB) using a PyTorch (version 2.4.1) implementation. For aligned RGB–depth pairs of size 256   ×   256 with a batch size of 1, the model attains an inference throughput of about 3 FPS, with a peak GPU memory footprint of around 3.2 GB during testing. Owing to the frozen teacher backbone, lightweight dual student branches, and the absence of external memory banks or heavy generative modules, the overall computational cost remains compatible with typical cycle-time requirements in in-line industrial inspection.

4.3. Ablation Study

To quantify the contribution of individual modalities and validate bimodal synergy, we evaluated three configurations under identical experimental conditions: using only depth (3D-only), only RGB, and both modalities (RGB+3D). As shown in Table 1 and Table 3, the class-wise mean I-AUROC for the three settings was 89.1%, 90.3%, and 97.1%, respectively; the bimodal setting achieved a clear and consistent improvement over either unimodal counterpart and exhibited stable advantages across categories. Per-category analysis indicated that texture-dominated categories (e.g., Carrot, Cookie) benefited markedly from RGB data, whereas geometry-dominated ones (e.g., Foam, Rope, Tire) showed substantially improved anomaly separability when depth information was incorporated. When the two modalities were combined, their advantages compounded—for example, Foam increased from 88.9% (3D-only) and 91.3% (RGB-only) to 98.4%; Tire rose from 83.3%/87.0% to 92.8%; and Carrot improved from 92.0%/92.4% to 99.5%.These results stem from cross-modal complementarity and consistency constraints: RGB is highly sensitive to fine-grained textures and color changes, whereas depth offers stronger structural discrimination for deformations, surface undulations, and missing regions. Joint evaluation suppresses unimodal false alarms in texture-complex or cluttered scenes and compensates for RGB omissions in geometry-dominated cases, thereby systematically widening the distribution gap between normal and anomalous samples.
As summarized in Table 4 and visualized in Figure 3, the baseline model, based on a ResNet50 teacher–student architecture, demonstrated limited discriminative capability for subtle defects in foreground regions and produced relatively diffuse, noisy anomaly responses in the background. With Perlin-guided percentage perturbations and texture/geometry alignment, the training signal became more challenging; normal–abnormal separability improved and the model’s reliance on background cues decreased markedly, which is reflected in Figure 3 by more concentrated responses around true defect regions. Incorporating the Guided Feature Refinement Network (GFRN) alone purified multi-scale features, enhanced long-range consistency and boundary fidelity, and widened the normal–abnormal margin, leading to cleaner and sharper anomaly maps. When both components were enabled, their synergy—stronger training supervision coupled with cleaner representations—yielded steady gains across metrics and visibly less cluttered heatmaps in Figure 3.
Fusion strategies. As also illustrated in Figure 3, max fusion generally delivered higher sensitivity than equal-weight averaging but produced isolated high-response regions that impaired spatial continuity. By contrast, the proposed Cross-Modal Squeeze–Excitation Gated Fusion (CM-SEGF) leveraged student semantic features to recalibrate channels and perform pixel-wise adaptive weighting, so that modality contributions are increased near true defects and suppressed in background regions, resulting in more contiguous and accurate anomaly maps in Figure 3 and thereby achieving more robust performance at both the image and pixel levels.

5. Conclusions

We have presented DCRDF-Net, which achieves robust multimodal unsupervised industrial anomaly detection through three synergistic designs: (i) a consistency-preserving pseudo-anomaly generator that synchronously injects conformal perturbations into appearance and depth within the foreground via multi-frequency Perlin guidance and texture/geometry alignment, yielding position-controllable and difficulty-adjustable training signals; (ii) a Guided Feature Refinement Network (GFRN) that aligns the frozen teacher’s defect-free prior on the student side via multi-layer projection compression and linear attention, purifying multi-scale representations while preserving boundary fidelity; and (iii) Cross-Modal Squeeze–Excitation Gated Fusion (CM-SEGF), which generates pixel-wise weights from both channel and spatial perspectives based on student semantics, explicitly calibrating the RGB/depth contributions and improving regional consistency and robustness. Experiments demonstrated the proposed model’s state-of-the-art performance on MVTec 3D-AD.
Although DCRDF-Net achieves strong detection and localization performance on MVTec 3D-AD, it still has several limitations. First, the Perlin-guided pseudo anomalies cover only a restricted family of defect morphologies and scales and are not intended to visually mimic the full diversity of real industrial defects. In particular, we observe that for a small subset of samples with large-scale structural damage or extremely specular surfaces the synthetic perturbations provide weaker guidance, and the performance margin over strong baselines becomes smaller. Second, the current study assumes that RGB and depth inputs are reasonably well-formed and does not yet systematically evaluate robustness under severely corrupted or missing single-modality conditions (e.g., heavily saturated or highly noisy depth channels). In addition, we reused an ImageNet-pretrained RGB encoder as a frozen teacher for the depth modality via channel replication. While this design works well under our experimental setup, it may still leave room for improved domain adaptation in more complex or strongly geometry-dominated depth scenarios. Future work will therefore focus on enriching the pseudo-anomaly distribution with more physically grounded manufacturing priors, as well as on exploring depth-aware or multimodal pre-training and more robust designs for handling extreme single-modality degradation.

Author Contributions

Conceptualization, C.W. and H.Z.; methodology, C.W.; software, C.W.; validation, C.W.; formal analysis, C.W.; investigation, C.W.; resources, J.C.; data curation, C.W.; writing—original draft preparation, C.W.; writing—review and editing, H.Z.; visualization, C.W.; supervision, J.C. and H Z.; project administration, H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data supporting the findings of this study are available as follows: the MVTec 3D-AD benchmark dataset is publicly available at https://www.mvtec.com/company/research/datasets/mvtec-3d-ad (accessed on 4 January 2026) (subject to the dataset’s terms of use).

Acknowledgments

We gratefully acknowledge Kawasaki (Chongqing) Robotics Engineering Co., Ltd., for providing the experimental environment and computational resources.

Conflicts of Interest

Author Jianbo Chen was employed by the company Kawasaki (Chongqing) Robotics Engineering Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Lin, Y.; Chang, Y.; Tong, X.; Yu, J.; Liotta, A.; Huang, G.; Song, W.; Zeng, D.; Wu, Z.; Wang, Y.; et al. A survey on RGB, 3D, and multimodal approaches for unsupervised industrial image anomaly detection. Inf. Fusion 2025, 116, 103139. [Google Scholar] [CrossRef]
  2. Bergmann, P.; Fauser, M.; Sattlegger, D.; Steger, C. MVTec AD—A comprehensive real-world dataset for unsupervised anomaly detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 9592–9600. [Google Scholar]
  3. Wang, J.; Niu, Y.; Huang, B. Fusion-restoration model for industrial multimodal anomaly detection. Neurocomputing 2025, 637, 130073. [Google Scholar] [CrossRef]
  4. Bergmann, P.; Jin, X.; Sattlegger, D.; Steger, C. The MVTec 3D-AD dataset for unsupervised 3D anomaly detection and localization. arXiv 2021, arXiv:2112.09045. [Google Scholar] [CrossRef]
  5. Wang, Y.; Peng, J.; Zhang, J.; Yi, R.; Wang, Y.; Wang, C. Multimodal industrial anomaly detection via hybrid fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 8032–8041. [Google Scholar]
  6. Cao, Y.; Xu, X.; Shen, W. Complementary pseudo multimodal feature for point cloud anomaly detection. Pattern Recognit. 2024, 156, 110761. [Google Scholar] [CrossRef]
  7. Zou, Y.; Jeong, J.; Pemula, L.; Zhang, D.; Dabeer, O. Spot-the-difference self-supervised pre-training for anomaly detection and segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23 October 2022; pp. 392–408. [Google Scholar]
  8. Zavrtanik, V.; Kristan, M.; Skočaj, D. DRAEM: A discriminatively trained reconstruction embedding for surface anomaly detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 8330–8339. [Google Scholar]
  9. Liu, Y.; Zhuang, C.; Lu, F. Unsupervised two-stage anomaly detection. arXiv 2021, arXiv:2103.11671. [Google Scholar] [CrossRef]
  10. Wolleb, J.; Bieder, F.; Sandkühler, R.; Cattin, P.C. Diffusion models for medical anomaly detection. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Singapore, 18 September 2022; pp. 35–45. [Google Scholar]
  11. Hoang, D.-C.; Tan, P.X.; Nguyen, A.-N.; Pham, M.-K.; Duong, T.H.A.; Huynh, T.-M.; Bui, S.-A.; Nguyen, D.-M.; Ha, Q.-H.; Trinh, V.-A.; et al. Unsupervised industrial anomaly detection using paired well-lit and low-light images. J. Comput. Des. Eng. 2025, 12, 41–61. [Google Scholar] [CrossRef]
  12. Gudovskiy, D.; Ishizaka, S.; Kozuka, K. CFlow-AD: Real-time unsupervised anomaly detection with localization via conditional normalizing flows. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 98–107. [Google Scholar]
  13. Rudolph, M.; Wehrbein, T.; Rosenhahn, B.; Wandt, B. Fully convolutional cross-scale flows for image-based defect detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 1088–1097. [Google Scholar]
  14. Rudolph, M.; Wandt, B.; Rosenhahn, B. Same same but differnet: Semi-supervised defect detection with normalizing flows. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 1907–1916. [Google Scholar]
  15. Bergmann, P.; Fauser, M.; Sattlegger, D.; Steger, C. Uninformed students: Student-teacher anomaly detection with discriminative latent embeddings. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4183–4192. [Google Scholar]
  16. Cao, Y.; Wan, Q.; Shen, W.; Gao, L. Informative knowledge distillation for image anomaly segmentation. Knowl.-Based Syst. 2022, 248, 108846. [Google Scholar] [CrossRef]
  17. Tien, T.D.; Nguyen, A.T.; Tran, N.H.; Huy, T.D.; Duong, S.; Nguyen, C.D.T.; Truong, S.Q.H. Revisiting reverse distillation for anomaly detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 24511–24520. [Google Scholar]
  18. Park, K.-B.; Lee, J.Y. Novel industrial surface-defect detection using deep nested convolutional network with attention and guidance modules. J. Comput. Des. Eng. 2022, 9, 2466–2482. [Google Scholar] [CrossRef]
  19. Ullah, W.; Khan, S.U.; Kim, M.J.; Hussain, A.; Munsif, M.; Lee, M.Y.; Seo, D.; Baik, S.W. Industrial defective chips detection using deep convolutional neural network with inverse feature matching mechanism. J. Comput. Des. Eng. 2024, 11, 326–336. [Google Scholar] [CrossRef]
  20. Chu, Y.-M.; Chieh, L.; Hsieh, T.-I.; Chen, H.-T.; Liu, T.-L. Shape-guided dual-memory learning for 3D anomaly detection. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023. [Google Scholar]
  21. Wang, C.; Zhu, H.; Peng, J.; Wang, Y.; Yi, R.; Wu, Y.; Ma, L.; Zhang, J. M3DM-NR: RGB-3D noisy-resistant industrial anomaly detection via multimodal denoising. IEEE Trans. Pattern Anal. Mach. Intell. 2025, in press. [Google Scholar] [CrossRef] [PubMed]
  22. Chen, R.; Xie, G.; Liu, J.; Wang, J.; Luo, Z.; Wang, J.; Zheng, F. EasyNet: An easy network for 3D industrial anomaly detection. In Proceedings of the 31st ACM International Conference on Multimedia (ACM MM), Ottawa, ON, Canada, 29 October–3 November 2023; pp. 7038–7046. [Google Scholar]
  23. Bi, C.; Li, Y.; Luo, H. Dual-branch reconstruction network for industrial anomaly detection with RGB-D data. In Proceedings of the International Conference on Image, Signal Processing, and Pattern Recognition (ISPP 2024), Guangzhou, China, 23–25 February 2024; Volume 13180, pp. 767–774. [Google Scholar]
  24. Zavrtanik, V.; Kristan, M.; Skočaj, D. Cheating depth: Enhancing 3D surface anomaly detection via depth simulation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 2164–2172. [Google Scholar]
  25. Liu, J.; Mou, S.; Gaw, N.; Wang, Y. Uni-3DAD: GAN-inversion aided universal 3D anomaly detection on model-free products. Expert Syst. Appl. 2025, 272, 126665. [Google Scholar] [CrossRef]
  26. Rudolph, M.; Wehrbein, T.; Rosenhahn, B.; Wandt, B. Asymmetric student-teacher networks for industrial anomaly detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023; pp. 2592–2602. [Google Scholar]
  27. Gu, Z.; Zhang, J.; Liu, L.; Chen, X.; Peng, J.; Gan, Z.; Jiang, G.; Shu, A.; Wang, Y.; Ma, L. Rethinking reverse distillation for multi-modal anomaly detection. AAAI Conf. Artif. Intell. (AAAI) 2024, 38, 8445–8453. [Google Scholar] [CrossRef]
  28. Cheng, Y.; Chen, J.; Wen, G.; Tan, X.; Liu, X. LPFSTNet: A lightweight and parameter-free head attention-based student–teacher network for fast 3D industrial anomaly detection. Neurocomputing 2025, 623, 129408. [Google Scholar] [CrossRef]
Figure 1. Overall architecture of the proposed DCRDF-Net.
Figure 1. Overall architecture of the proposed DCRDF-Net.
Sensors 26 00412 g001
Figure 2. Qualitative comparison of anomaly localization on the MVTec 3D-AD dataset. The compared methods are shown using our implementations, and the last column presents the predictions of DCRDF-Net.
Figure 2. Qualitative comparison of anomaly localization on the MVTec 3D-AD dataset. The compared methods are shown using our implementations, and the last column presents the predictions of DCRDF-Net.
Sensors 26 00412 g002
Figure 3. Qualitative comparison of anomaly heatmaps for different ablation variants of DCRDF-Net on the MVTec 3D-AD dataset.
Figure 3. Qualitative comparison of anomaly heatmaps for different ablation variants of DCRDF-Net on the MVTec 3D-AD dataset.
Sensors 26 00412 g003
Table 1. Image-level anomaly detection results (I-AUROC%) on MVTec 3D-AD. (bold values denote the maximum values of each column).
Table 1. Image-level anomaly detection results (I-AUROC%) on MVTec 3D-AD. (bold values denote the maximum values of each column).
MethodBagelCable
Gland
CarrotCookieDowelFoamPeachPotatoRopeTireMean
DepthGAN53.837.258.060.343.053.464.260.144.357.753.2
DepthAE64.850.265.048.880.552.271.252.954.055.259.5
DepthVM51.355.147.758.161.771.645.042.159.862.355.5
BTF91.874.896.788.393.258.289.691.292.188.686.5
Shape-guided98.689.498.399.197.685.799.096.596.086.994.7
AST98.387.397.697.193.288.597.498.1100.079.793.7
EasyNet99.199.891.896.894.594.590.580.799.479.392.6
M3DM99.490.997.297.696.094.297.388.997.285.094.5
MMRD99.994.396.494.399.291.294.990.199.490.195.0
Ours99.198.999.599.396.298.495.193.798.192.897.1
Table 2. Pixel-level anomaly localization results (PRO%) on MVTec 3D-AD. (bold values denote the maximum values of each column).
Table 2. Pixel-level anomaly localization results (PRO%) on MVTec 3D-AD. (bold values denote the maximum values of each column).
MethodBagelCable
Gland
CarrotCookieDowelFoamPeachPotatoRopeTireMean
DepthGAN42.142.277.869.649.425.228.536.240.263.147.4
DepthAE43.215.880.849.184.140.626.221.671.647.848.1
DepthVM38.832.119.457.040.828.224.434.926.833.133.5
BTF97.696.997.997.393.388.897.598.195.097.195.9
Shape-guided98.197.398.297.196.297.898.198.397.497.597.6
AST97.094.798.193.991.390.697.998.288.994.094.4
EasyNet93.594.197.189.788.599.799.288.895.572.891.9
M3DM97.097.197.995.094.193.297.797.197.197.596.4
MMRD98.699.099.195.199.090.194.999.098.798.297.6
Ours99.198.999.598.596.298.496.795.798.895.898.8
Table 3. I-AUROC(%) score for anomaly detection of MVTec 3D-AD. (bold values denote the maximum values of each column).
Table 3. I-AUROC(%) score for anomaly detection of MVTec 3D-AD. (bold values denote the maximum values of each column).
GroupMethodBagelCable
Gland
CarrotCookieDowelFoamPeachPotatoRopeTireMean
3DDepthGAN53.037.660.760.349.748.459.548.953.652.152.3
DepthAE46.873.149.767.353.441.748.554.956.454.654.6
FPFH82.555.195.279.788.358.275.888.992.965.378.2
3D-ST86.248.483.289.484.866.376.368.795.848.674.8
Shape-guided98.368.297.899.896.073.799.397.996.687.191.6
AST88.157.696.595.767.979.799.091.595.661.183.3
M3DM94.165.196.596.990.576.088.097.492.676.587.4
EasyNet73.567.874.786.471.971.671.372.588.568.774.7
MMRD82.966.693.780.497.286.594.780.696.784.986.6
Ours92.689.492.091.887.788.987.687.290.683.389.1
RGBPatchCore87.688.079.168.291.270.169.561.884.170.277.0
Shape-guided91.193.688.366.297.477.278.564.188.470.681.5
AST94.792.885.182.598.195.189.561.399.282.188.0
EasyNet98.299.291.795.391.992.384.078.598.674.290.4
M3DM94.491.889.674.995.976.791.964.893.876.785.0
MMRD98.793.794.377.098.184.791.375.399.385.389.8
Ours92.091.892.492.289.191.389.086.692.087.090.3
Table 4. Architecture ablation on MVTec 3D-AD (I-AUROC and PRO, %). (bold values denote the maximum values of each column).
Table 4. Architecture ablation on MVTec 3D-AD (I-AUROC and PRO, %). (bold values denote the maximum values of each column).
VariantPerlin NoiseGFRNFusionI-AUROCPRO
BaselineNoNoAvg83.776.4
+ Perlin NoiseYesNoAvg86.188.2
+ GFRNNoYesAvg86.575.7
+ Perlin Noise + GFRN (Avg)YesYesAvg88.480.4
+ Perlin Noise + GFRN (Max)YesYesMax89.190.6
FullYesYesCM-SEGF97.198.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, C.; Chen, J.; Zhang, H. DCRDF-Net: A Dual-Channel Reverse-Distillation Fusion Network for 3D Industrial Anomaly Detection. Sensors 2026, 26, 412. https://doi.org/10.3390/s26020412

AMA Style

Wang C, Chen J, Zhang H. DCRDF-Net: A Dual-Channel Reverse-Distillation Fusion Network for 3D Industrial Anomaly Detection. Sensors. 2026; 26(2):412. https://doi.org/10.3390/s26020412

Chicago/Turabian Style

Wang, Chunshui, Jianbo Chen, and Heng Zhang. 2026. "DCRDF-Net: A Dual-Channel Reverse-Distillation Fusion Network for 3D Industrial Anomaly Detection" Sensors 26, no. 2: 412. https://doi.org/10.3390/s26020412

APA Style

Wang, C., Chen, J., & Zhang, H. (2026). DCRDF-Net: A Dual-Channel Reverse-Distillation Fusion Network for 3D Industrial Anomaly Detection. Sensors, 26(2), 412. https://doi.org/10.3390/s26020412

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop