Skip Content
You are currently on the new version of our website. Access the old version .
MathematicsMathematics
  • Article
  • Open Access

30 January 2026

Hybrid-Frequency-Aware Mixture-of-Experts Method for CT Metal Artifact Reduction

,
,
and
1
School of Medicine and Health, Harbin Institute of Technology, Harbin 150001, China
2
Institute of Medical and Health Sciences, Harbin Institute of Technology Zhengzhou Research Institute, Zhengzhou 450000, China
3
Faculty of Computing, Harbin Institute of Technology, Harbin 150001, China
4
School of Cyber Security and Smart Policing, Zhengzhou Police University, Zhengzhou 450000, China
This article belongs to the Special Issue Structural Networks for Image Application

Abstract

In clinical CT imaging, high-density metallic implants often induce severe metal artifacts that obscure critical anatomical structures and degrade image quality, thereby hindering accurate diagnosis. Although deep learning has advanced CT metal artifact reduction (CT-MAR), many methods do not effectively use frequency information, which can limit the recovery of both fine details and overall image structure. To address this limitation, we propose a Hybrid-Frequency-Aware Mixture-of-Experts (HFMoE) network for CT-MAR. The proposed method synergizes the spatial-frequency localization of the wavelet transform with the global spectral representation of the Fourier transform to achieve precise multi-scale modeling of artifact characteristics. Specifically, we design a hybrid-frequency interaction encoder with three specialized branches, incorporating wavelet-domain, Fourier-domain, and cascaded wavelet–Fourier modulation, to distinctively refine local details, global structures, and complex cross-domain features. Then, they are fused via channel attention to yield a comprehensive representation. Furthermore, a Frequency-Aware Mixture-of-Experts (MoE) mechanism is introduced to dynamically route features to specific frequency experts based on the degradation severity, thereby adaptively assigning appropriate receptive fields to handle varying metal artifacts. Evaluations on synthetic (DeepLesion) and clinical (SpineWeb, CLINIC-metal) datasets show that HFMoE outperforms existing methods in both quantitative metrics and visual quality. Our method demonstrates the value of explicit frequency-domain adaptation for CT-MAR and could inform the design of other image restoration tasks.

1. Introduction

Computed tomography (CT) is a fundamental tool in modern diagnostic radiology. However, high-density metallic implants, e.g., dental fillings, hip prostheses, and surgical clips, can corrupt projection data, leading to severe metal artifacts in reconstructed images [1]. These artifacts manifest as dark bands and bright streaks, which not only degrade perceived image quality but also obscure critical anatomical details of surrounding tissues. Moreover, metal artifacts can compromise downstream clinical tasks, such as target delineation in radiotherapy and computer-aided diagnosis. Therefore, developing effective CT metal artifact reduction (MAR) techniques is crucial for restoring diagnostic information.
Early solutions primarily relied on projection inpainting methods, such as linear interpolation (LI) [2] and normalized MAR (NMAR) [3]. Although these methods can reduce strong streaks, they often introduce secondary artifacts and blur fine details because they do not account for the nonlinear effects of beam hardening and scattering. In recent years, deep learning (DL) [4] has significantly advanced this field owing to its powerful nonlinear representation capacity and hierarchical architecture. Existing DL-based MAR methods are generally categorized into image domain [5,6], sinogram domain [7,8], and dual-domain collaboration approaches [9,10,11,12,13]. Among these, dual-domain methods have shown strong performance by combining information from both projection and image domains.
Despite progress, current deep learning methods remain limited in adapting to the varying frequency characteristics of artifacts. From the perspective of imaging physics, metal artifacts exhibit distinct spectral characteristics. Beam hardening typically induces low-frequency global shading (cupping effects), whereas photon starvation and scattering manifest as high-frequency radiating streaks and edge discontinuities. The distribution of these frequency components varies with the implant’s geometry, material, and location. However, conventional spatial-domain CNNs encounter a key limitation in balancing global context modeling and local detail preservation. Although increasing network depth can enlarge the theoretical receptive field, the effective receptive field (ERF) remains substantially constrained in practice [14], limiting the ability of the network to model long-range artifact streaks. As a result, many approaches adopt coarse-scale representations through spatial downsampling to aggregate global context, which unavoidably compromises high-frequency anatomical details. This design choice leads to a notable trade-off between artifact suppression and fine-structure preservation. Transformer-based architectures offer an alternative for modeling long-range dependencies [15]; however, their quadratic computational complexity with respect to image resolution hinders practical deployment in high-resolution medical imaging. Such limitations are clinically nontrivial, as the loss or blurring of subtle textures may introduce diagnostic ambiguity by obscuring or mimicking pathological structures. Consequently, it is necessary to propose a reconstruction framework that explicitly decouples artifact elimination from structural fidelity restoration.
To resolve this diagnostic trade-off and address these issues, we propose that an effective MAR model should adapt to frequency content and employ dynamic computation. We propose a Hybrid-Frequency-Aware Mixture-of-Experts (HFMoE) framework, which separates artifact processing into frequency bands and dynamically routes features to specialized modules. The core strategy is two-fold: First, to overcome the spectral bias, we introduce a hybrid frequency analysis strategy. We combine the wavelet transform, which provides joint spatial-frequency localization to preserve local textures, and the Fourier transform, which offers a global receptive field to effectively model and suppress long-range artifact signatures without destructive downsampling. Furthermore, a cascaded wavelet-Fourier branch is introduced to capture non-stationary spectral dependencies across different resolution scales. Second, to address the limitation of static networks, we integrate the mixture-of-experts (MoE) philosophy into the decoder. Unlike static layers, our frequency-aware experts specialize in different artifact types. A dynamic gating mechanism learns to assign higher weights to the most relevant experts based on the local frequency content of the input, realizing a “divide-and-conquer” strategy for artifact removal. The encoder includes a hybrid-frequency module with parallel wavelet, Fourier, and cascaded branches, fused via channel attention. In the decoder, the MoE architecture takes both spatial and wavelet high-frequency features as input, dynamically allocating computational resources to suppress artifacts while preserving tissue details. Experiments on Synthesized DeepLesion, Clinical SpineWeb, and CLINIC-metal datasets show that HFMoE improves upon current methods in artifact reduction and detail preservation.
In summary, our main contributions are as follows:
  • We propose HFMoE, a unified network that integrates wavelet and Fourier transforms to effectively exploit the complementary benefits of spatial-frequency and global spectral information for robust CT-MAR.
  • A hybrid-frequency interaction encoder is developed to capture multi-scale features, which incorporates concurrent wavelet, Fourier, and cascaded wavelet-Fourier modulation branches to facilitate multi-resolution spectral refinement, ensuring the restoration of local textures, global structures, and the complex cross-domain dependencies between them.
  • A Frequency-Aware MoE strategy is implemented in the decoder to dynamically allocate frequency-specific experts according to the degradation severity of the input, enabling adaptive artifact correction.
  • Extensive evaluations on multiple benchmarks, including synthesized and clinical datasets, validate that our method outperforms existing state-of-the-art methods in terms of both quantitative metrics and visual quality.

3. Materials and Methods

3.1. Frequency-Domain Foundations: DWT and FFT

Our method is based on the complementary analysis provided by the discrete wavelet transform (DWT) and the fast Fourier transform (FFT). Metal artifacts in CT scans exhibit a complex, hybrid-frequency nature, comprising both localized high-frequency streaking and globally distributed low-frequency shading. Notably, these components are often spatially entangled and non-stationary rather than cleanly separable in clinical CT images, due to scattering effects that depend on implant geometry and material properties. Individually, DWT and FFT offer distinct advantages for modeling different aspects of this problem, yet neither alone is sufficient to fully characterize such entangled artifacts. Their combination, therefore, forms the basis of our hybrid-frequency approach.

3.1.1. Discrete Wavelet Transform for Localized Analysis

The 2D DWT performs a multi-resolution decomposition of an image, enabling joint analysis in space and frequency. For an input image or feature map X R H × W , a single decomposition level applies quadrature mirror filter banks, producing four subbands:
{ X L L , X L H , X H L , X H H } = DWT ( X ) ,
where X L L represents the low-frequency approximation component, capturing coarse-grained structural information. The subbands X L H , X H L , and X H H correspond to high-frequency details along the horizontal, vertical, and diagonal directions, respectively. The original signal can be exactly reconstructed via the Inverse DWT (IWT): X = IWT ( X L L , X L H , X H L , X H H ) .
Specifically, DWT provides fine-grained space-frequency localization, which enables effective modeling of localized discontinuities and edge distortions induced by metal artifacts, i.e., sharp intensity transitions near metal–tissue interfaces and fine structural perturbations in surrounding anatomy. However, due to its limited receptive field within each subband, DWT alone is insufficient for capturing long-range, globally coherent streak artifacts that span large spatial extents. Thus, DWT is suitable for preserving local structural details but cannot fully represent the global organization of metal-induced corruption.

3.1.2. Fast Fourier Transform for Global Context

In contrast, the FFT provides a global representation of an image’s frequency content. For the 2D feature map X , its 2D FFT is defined as follows:
F ( X ) [ u , v ] = h = 0 H 1 w = 0 W 1 X [ h , w ] · e j 2 π ( u h / H + v w / W ) ,
with the inverse transform (IFFT) denoted by F 1 ( · ) . The magnitude spectrum | F ( X ) | reveals the global energy distribution across all frequencies, while the phase spectrum encodes essential structural information. The key attribute of the FFT is its inherent global receptive field; each point in the frequency domain depends on every pixel in the spatial domain.
This property makes the FFT effective for modeling globally correlated and directionally coherent components of metal artifacts, such as long-range streak backbones and low-frequency shading caused by beam hardening. The FFT and IFFT operations are efficient, with a computational complexity of O ( H W log ( H W ) ) . However, due to the lack of spatial localization in the FFT basis functions, purely Fourier-domain processing risks suppressing anatomical structures whose frequency orientations resemble those of metal streaks.

3.1.3. Motivation for a Hybrid Frequency-Aware Design

DWT and FFT have complementary characteristics [42]. While DWT offers precise spatial localization that is well-suited for high-frequency and localized artifact features, its ability to capture long-range dependencies is inherently limited. Conversely, the FFT provides a global spectral view that captures global artifact structure but sacrifices spatial selectivity.
Importantly, CT metal artifacts often depart from the implicit assumption that local and global components can be cleanly separated. They frequently exhibit globally coherent spectral patterns whose local appearance varies across spatial locations and resolution scales. Therefore, we propose to construct a unified wavelet–Fourier hybrid frequency-aware representation that is designed to model this entanglement rather than assume a simple decomposition. By jointly leveraging wavelet-based spatial-scale representations and Fourier-based global spectral analysis, the network can capture scale-dependent spectral characteristics of metal artifacts. Integrating both domains thus provides a richer and more robust feature space for artifact reduction compared to single-domain methods.

3.2. Overview

Our proposed HFMoE follows an encoder–decoder architecture with skip connections. Given a metal-corrupted CT image Y R H × W , the network first extracts shallow features via a convolutional stem: F 0 = Conv 3 × 3 ( Y ) .
The encoder comprises four stages, each performing feature downsampling via DWT. Each stage contains a Hybrid-Frequency Interaction Module (HFIM) (Section 3.3) that processes features in the spatial and frequency domains. The bottleneck layer employs multiple HFIMs for deep feature refinement.
The decoder symmetrically upsamples features via IWT. Crucially, each decoder stage incorporates the proposed Frequency-Aware Mixture-of-Experts Module (FAME) (Section 3.4), which dynamically routes features to experts specialized for different frequency patterns and receptive fields. Skip connections aggregate multi-scale features from the encoder to the decoder.
The final restored image X ^ is obtained through a reconstruction layer and a residual connection:
X ^ = Conv 3 × 3 ( F d ) + Y ,
where F d denotes the output of the decoder. This residual design enables the network learn the artifact component.

3.3. Hybrid-Frequency Interaction Module (HFIM)

The proposed HFIM is designed to jointly leverage the complementary strengths of wavelet and Fourier transforms for local detail preservation and global artifact suppression. As shown in Figure 1, given an input feature F i n H F I M R H × W × C , it is first projected to F = Conv 1 × 1 ( F i n ) . F is then split into three parallel branches.
Figure 1. The architecture of the Hybrid-Frequency Interaction Module (HFIM).
  • Branch 1: Wavelet-Domain Modulation. This branch focuses on localized details. F 1 is decomposed via DWT into four subbands: { F 1 L L , F 1 L H , F 1 H L , F 1 H H } . The high-frequency (HF) subbands are concatenated with the low-frequency (LF) component and processed by depthwise-separable convolution blocks f D ( · ) :
F 1 , H = Concat f D L H ( [ F 1 L L , F 1 L H ] ) , f D H L ( [ F 1 L L , F 1 H L ] ) , f D H H ( [ F 1 L L , F 1 H H ] ) .
The processed features are then reconstructed via IWT: F 1 = IWT ( F 1 L L , F 1 , H ) .
  • Branch 2: Fourier-Domain Modulation. This branch targets globally distributed artifacts. The feature is transformed via 2D FFT: F ( F 2 ) C H × W × C / 3 . A learnable frequency filter W F C H × W × C / 3 performs element-wise modulation, followed by the inverse FFT (iFFT):
    F 2 = F 1 W F F ( F 2 ) .
  • Branch 3: Cascaded Wavelet–Fourier Modulation. This branch models interactions across scales and frequencies. The HF subbands from an initial DWT of F 3 are first transformed to the Fourier domain, filtered, and then transformed back:
    F 3 , i = F 1 W F , i F f F i ( [ F 3 L L , F 3 i ] ) , i { L H , H L , H H } .
  • The final output is obtained via IWT, i.e., F 3 = IWT ( F 3 L L , F 3 , L H , F 3 , H L , F 3 , H H ) .
The outputs of the three branches are aggregated via channel attention and passed through a feedforward network (FFN), resulting in the final output F o u t H F I M of the HFIM.

3.4. Frequency-Aware Mixture-of-Experts Module (FAME)

Building upon the MoE paradigm, the FAME module dynamically selects specialized experts based on the frequency characteristics of input features. This architecture enables the network to adapt its ERF to varying artifact scales, a critical requirement for handling non-stationary metal degradations. As illustrated in Figure 2, the input feature F i n F A M E R H × W × C is first enriched with high-frequency cues extracted via DWT. This hybrid feature serves as input to the gating network and the expert pool.
Figure 2. The architecture of Frequency-Aware Mixture-of-Experts (FAME).
  • Expert Design. Our expert pool { E i } i = 1 N consists of three types of experts with heterogeneous receptive fields designed to handle different degradation patterns via multi-scale feature arbitration:
    Z ^ E i = Conv 1 × 1 Z E 1 ,   R F = 1 IWT Conv k × k DWT Z E i , 2 R F < 16 Z E i F 1 F w Q Z E i F w K Z E i H ,   R F 16
  • Local Expert ( R F = 1 ): Employs pointwise convolution ( 1 × 1 Conv) for pixel-wise adjustment, minimizing the risk of blurring fine textures during artifact-free region restoration.
  • Wavelet Expert ( 2 R F < 16 ): Applies a DWT-IWT sandwich with a k × k depthwise convolution in the wavelet domain, providing mid-range context to suppress localized streaks while preserving sub-band edges.
  • Fourier Expert ( R F 16 ): Approximates global self-attention via Fourier autocorrelation, offering a holistic RF to neutralize large-scale, cross-image streaks.
  • The i-th expert processes a routed feature Z i and produces an output Z ^ i .
  • Frequency-Aware Routing. The gating network G facilitates input-conditional execution by mapping frequency signatures to the most appropriate expert. A lightweight router, implemented as a linear layer followed by softmax, generates a routing weight vector g = G ( F ) R N . For each spatial position (or token), only the expert with the highest weight is activated (Top-1 routing). By dynamically selecting the optimal RF for each token, the model circumvents the limitations of static architectures, thereby ensuring robust MAR performance across varying levels of artifact severity. To encourage balanced utilization and align expert choice with frequency content, we employ an auxiliary load-balancing loss L a u x [41].
The final output of the FAME module is the aggregated expert outputs, modulated by a shared feature projection S ( F ) , and processed by an FFN:
F o u t F A M E = FFN i = 1 N g i · ( Z ^ i S ( F ) ) .

3.5. Loss Functions

Our training objective combines three complementary losses, as follows:
  • 1. Reconstruction Loss ( L r e c ): An l 1 loss [43] ensures pixel-level fidelity, weighted inside ( M ) and outside the metal trace region:
    L r e c = ( X ^ X g t ) M 1 + λ b g ( X ^ X g t ) ( 1 M ) 1 .
  • 2. Hybrid Frequency Loss ( L f r e q ): This loss enforces consistency in both wavelet and Fourier domains. Let D W ( · ) and D F ( · ) denote operations extracting high-frequency wavelet subbands and Fourier magnitude spectra, respectively.
    L f r e q = s { L H , H L , H H } α s D W s ( X ^ ) D W s ( X g t ) 1 + β D F ( X ^ ) D F ( X g t ) 1 .
  • 3. Routing Auxiliary Loss ( L a u x ): Adopted from [41], this loss balances expert utilization and encourages alignment between routing decisions and task complexity.
The total loss is a weighted sum: L t o t a l = L r e c + λ f L f r e q + λ a L a u x . The hyperparameters λ b g , λ f , λ a , α s , β are selected based on empirical observations.

4. Experiments

4.1. Physically-Informed Artifact Simulation and Clinical Dataset

To comprehensively evaluate the proposed method, we utilize one synthesized dataset for training and quantitative benchmarking, and two clinical datasets for assessing cross-domain generalizability.
  • Synthesized DeepLesion Dataset. Following the protocol established in [6,44], we constructed a synthesized dataset based on the DeepLesion collection. The training set consisted of 1000 paired clean and metal-corrupted volumes, generated by simulating metal artifacts on 1200 randomly sampled metal-free abdominal/thoracic CT images using 90 distinct implant types [7]. To faithfully replicate complex clinical scenarios, the pipeline integrated a polychromatic X-ray model (with 2 × 10 7 incident photons) to couple beam-hardening and partial-volume effects with realistic Poisson noise, while extensively randomizing metal mask size, orientation, and position within diverse anatomical regions to reflect clinical variations in patient habitus and implant trajectories. For evaluation, 200 pairs featuring 10 unseen implant types were reserved, focusing on physical attenuation characteristics and geometric morphology rather than specific chemical compositions. These test implants covered a wide range of spatial scales, from 35 pixels (tiny fragments) to 2061 pixels (large prostheses). Consistent with prior works [6,44], all volumes were resampled into 640 projection views under a standard fan-beam geometry and reconstructed at a resolution of 416 × 416 .
  • Clinical SpineWeb Dataset (available at https://csi-workshop.weebly.com/ or https://zenodo.org/records/7049844, accessed on 1 June 2014). To validate clinical generalization ability, we employed the SpineWeb dataset as an external testbed. This dataset consists of post-operative spinal CT scans with thoracolumbar instrumentation. We selected a set of spine volumes with implants that were not used in training. Metal implants were segmented using a standard threshold of 2500 HU. This dataset served to evaluate the model’s performance on real-world metallic hardware distinct from the training distribution.
  • Clinical CLINIC-metal Dataset. We also utilized the CLINIC-metal pelvic CT dataset [45] to test generalization to different anatomies. This multi-center dataset contained pelvic CT scans with severe artifacts, which challenged the model across different anatomies and acquisition protocols.

4.2. Implementation Details

Table 1 outlines the architectural specifications of HFMoE. We employed progressive multi-resolution training, increasing input size from 256 × 256 to 384 × 384 , then to 416 × 416 . The model was optimized using the AdamW optimizer with a cosine annealing schedule, where the learning rate decayed from an initial 1 × 10 3 to 1 × 10 6 following [46]. Data augmentation techniques, including random flipping, rotation, cropping, and MixUp, were applied to prevent overfitting. The framework was implemented in PyTorch 2.9.1. All experiments were conducted on a single NVIDIA RTX 4080S GPU. The model was trained for 2000 epochs, requiring approximately 72 h.
Table 1. HFMoE network configurations.

4.3. Experimental Results and Discussion

4.3.1. Evaluation Metrics

In this paper, we employ Peak Signal-to-Noise Ratio (PSNR) and the Structural Similarity Index (SSIM) as quantitative evaluation metrics to comprehensively and objectively assess the performance of the proposed model on the synthetic dataset. PSNR quantifies pixel-level fidelity based on the Mean Squared Error (MSE), while SSIM assesses structural preservation, which correlates more closely with diagnostic perception.
For two images, I and K (from [ 1000 , 3071 ] HU normalized to range [ 0 , 255 ] ), PSNR is defined as follows:
P S N R = 10 · log 10 M A X I 2 M S E , where M S E = 1 H W i , j [ I ( i , j ) K ( i , j ) ] 2
The SSIM index is calculated as follows:
S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
where μ and σ 2 represent the local mean and variance, σ x y is the covariance, and C 1 , C 2 are stability constants. These definitions ensure the evaluation criteria are transparent.

4.3.2. Quantitative Evaluation

Table 2 compares our proposed HFMoE with 16 existing MAR methods. For fair comparison, models with open-source code (e.g., OSCNet+, InDuDoNet) were evaluated using official pre-trained weights, while others were retrained following their original protocols. HFMoE achieves superior performance across all metrics. As detailed in Table 2, our method obtains the highest PSNR across all implant sizes. Specifically, HFMoE surpasses the best-performing MoE-based competitor, MoCE-IR, by 0.57 dB in PSNR, and outperforms the leading dual-domain method, InDuDoNet+, by a considerable margin of ∼3 dB. The improved SSIM scores also indicate enhanced structural preservation.
Table 2. Performance evaluation in terms of PSNR/SSIM on the Synthesized DeepLesion benchmark. Five distinct implant pairs with varying pixel counts ( [ 2061 , 890 ] , [ 81 , 451 ] , [ 254 , 124 ] , [ 118 , 112 ] , [ 53 , 35 ] , from large to small) were evaluated to validate the robustness of our HFMoE against different implant sizes. Bold numbers denote the highest scores.
To validate these improvements, a paired-sample t-test was conducted between HFMoE and the second-best method, i.e., MoCE-IR, yielding p < 0.001 ( p = 8.75 × 10 15 ). This confirms that the performance gains are statistically significant. Moreover, HFMoE achieved a lower standard deviation in PSNR ( 1.74 vs. 1.98 in Table 3), demonstrating superior stability across diverse anatomical structures and implant geometries. These quantitative advantages align with our radiologist study, where HFMoE was preferred in 81 % of clinical cases for its higher diagnostic fidelity.
Table 3. Standard deviation analysis of PSNR performance. Comparison of stability across varying metal artifact severities on the Synthesized DeepLesion benchmark. Lower standard deviation indicates higher consistency. Bold indicates the best stability.

4.3.3. Qualitative Evaluation

  • Synthesized DeepLesion. Figure 3 presents the reconstruction results for cases with varying implant sizes. HFMoE demonstrates robust artifact suppression, particularly in regions surrounding large implants. While competing methods often leave noticeable residual streaks (indicated by circles), our approach effectively mitigates these artifacts and better preserves anatomical boundaries. For small implants, HFMoE also achieves higher texture fidelity compared to existing methods.
    Figure 3. Visual comparisons of the MAR results on four representative test cases with varying implant sizes and geometries. (Row 1) Case 1175: 112 px staple-like implant; (Row 2) Case 100: 124 px columnar implant; (Row 3) Case 1362: 892 px parallel implants; (Row 4) Case 1618: 2061 px large pelvic scaffold. Red and green boxes highlight critical regions where our method effectively suppresses streaking and shading artifacts, successfully restoring fine anatomical textures and maintaining morphological fidelity (e.g., the circular structure in Case 1175 and sharp boundaries in Case 1618) that are lost or distorted in other methods. Display: [−175, 275] HU; metal (red). ROIs are enlarged. Zoom in for a better view.
  • Clinical Datasets. Qualitative results on the clinical SpineWeb and CLINIC-metal datasets are presented in Figure 4 and Figure 5, respectively. Unlike synthetic benchmarks, these clinical datasets involve real-world metal implants and noise patterns that are complex and difficult to model, presenting substantial challenges for robust artifact reduction. On SpineWeb, HFMoE effectively reduces beam-hardening streaks induced by spinal fixation hardware. It suppresses long-range artifacts while maintaining the continuity and visibility of vertebral structures, in contrast to baseline methods that either leave residual streaks or over-smooth anatomical details. On the CLINIC-metal dataset, which contains complex pelvic anatomy and diverse implant configurations, our method demonstrates robustness to various degradation patterns and generalizes well across different tissue textures, indicating its capacity to handle heterogeneous clinical scenarios. Moreover, HFMoE consistently achieves an effective trade-off between noise suppression and detail preservation. It preserves tissue boundaries and structural sharpness more effectively than other methods, which often introduce blurring, texture loss, or local distortions when confronted with severe noise and unknown implants.
    Figure 4. Qualitative comparison of clinical spine CT scans. (Row 1) An elliptical implant: the red box showcases our method’s superior noise reduction performance, providing a cleaner restoration of the underlying tissues, while the green box highlights its ability to preserve fine-grained anatomical textures during the denoising process. (Row 2) A small implant with prominent streak artifacts: the red box demonstrates effective artifact suppression, and the green box underscores our method’s ability to not only remove interfering streaks but also restore sharper anatomical boundaries and improve overall structural fidelity.Display window: [50, 500] HU. ROIs are enlarged. Zoom in for a better view.
    Figure 5. Qualitative comparison on the CLINIC-Metal dataset. (Top row): For the near-circular implant, the zoomed-in ROIs highlight that our method is the only one to achieve complete artifact elimination while maintaining sharp and clear structural boundaries. (Bottom row): In the presence of multiple metal implants, our approach exhibits the highest efficacy in suppressing streak artifacts and restoring latent tissue information, outperforming existing state-of-the-art techniques. Display window: [50, 500] HU. ROIs are enlarged. Zoom in for a better view.

4.4. Computational Efficiency

Figure 6 contrasts the inference efficiency of leading MAR methods. HFMoE has an inference time of 38 ms per slice ( 416 × 416 ), tested on a standard NVIDIA RTX 4080S GPU. This represents a 40% speed improvement over ACDNet (65 ms), while simultaneously achieving a ∼0.8 dB higher PSNR than MARFormer and MoCE-IR. The model has a small memory footprint: training uses ∼32 GB VRAM (batch size 4), and inference requires ∼180 MB per slice. Our algorithm relies on FFT and DWT operations that are natively supported by the vast majority of modern hardware, including high-end NVIDIA RTX GPUs or mobile-grade processors like the Mali-G31. Such computational efficiency and broad hardware compatibility ensure that HFMoE can be integrated into existing clinical CT reconstruction workflows without introducing significant latency.
Figure 6. Runtime vs. PSNR with model complexity. Horizontal axis: running time (s); vertical axis: PSNR (dB); parameter count is noted below each method.

4.5. Ablation Study

We conducted a comprehensive ablation study on the Synthesized DeepLesion dataset to validate the contributions of three core components: the Hybrid-Frequency Interaction Module (HFIM), the Frequency-Aware Mixture-of-Experts (FAME), and the multi-scale loss function.

4.5.1. Efficacy of HFIM

The HFIM leverages complementary information from the wavelet and Fourier domains. To analyze its impact, we constructed a baseline model (Model-01) based on the Transformer architecture, where the self-attention mechanism was replaced by channel attention as the primary interaction mechanism, and we progressively integrated frequency modulation blocks.
As shown in Figure 7, introducing wavelet-domain modulation (Model-02) yields performance gains by capturing local high-frequency details. While Fourier-domain modulation alone (Model-03) and cascaded wavelet-Fourier modulation (Model-04) prove less stable due to spectral sensitivity, their parallel combination with wavelet modulation (Model-05) produces a notable improvement, suggesting that the two domains offer complementary representations.
Figure 7. Contribution analysis of HFIM variants: PSNR (left) and SSIM (right) versus implant size (large to small). Curves represent different model configurations. Note: The black curve sequences Model-1 to Model-7 (left to right), not implant size.
Moreover, we provide visualizations of feature maps across different branches in Figure 8. These visualizations demonstrate that wavelet-domain modulation selectively processes local details, while Fourier-domain modulation captures global structural information. The cascaded wavelet–Fourier modulation effectively integrates both capabilities, handling cases where high- and low-frequency components are not distinctly separated. The final configuration (Model-06), which incorporates a cascaded wavelet-Fourier strategy, achieves the optimal performance, confirming the synergy of the proposed triple-branch mechanism. Ablating the channel attention module (Model-07) leads to a performance drop, indicating that such a mechanism plays a crucial role in regulating the fusion of diverse frequency components.
Figure 8. Visualization of feature inputs and outputs across frequency-aware modeling branches. Red denotes higher values, whereas blue signifies lower values. The left half of the first row compares wavelet-domain subband inputs and outputs, while the right half presents FFT magnitude spectra of Fourier-domain features. The second row illustrates the inputs and outputs of the cascaded wavelet–Fourier branch, showing wavelet subbands alongside their corresponding spectral energy distributions.

4.5.2. Role of FAME

The FAME module utilizes a gating mechanism to dynamically route features based on local frequency content. We compared our frequency-aware design against standard MoE and single-domain variants (Figure 9). While replacing standard MoE experts (Model-08) or multi-scale convolutions (Model-09) with a standard feedforward network (Model-06) offers baseline results, employing domain-specific experts, either wavelet-based (Model-10) or Fourier-based (Model-11), leads to noticeable improvements. The full FAME module, which adaptively selects between local wavelet experts for stationary regions and global Fourier experts for complex textures, achieves optimal performance.
Figure 9. Performance of FAME variants across implant sizes (large to small). (Left): PSNR; (Right): SSIM. Each curve denotes a model configuration. Note: The black curve sequences Model-6 to Model-11 (left to right), not implant size.
This architecture enables an adaptive routing strategy, where the gating logic analyzes the input’s artifact intensity and spectral distribution to achieve input-dependent feature processing. Specifically, by prioritizing Fourier experts for globally coherent streaks and wavelet experts for localized anatomical recovery, FAME effectively disentangles artifacts from underlying anatomical structures. This indicates that the proposed FAME module can adaptively assign appropriate receptive fields and domain-specific priors to handle varying metal artifacts, ranging from subtle localized distortions to severe global artifacts.

4.5.3. Impact of Loss Function

We finally evaluate the composite loss function, which incorporates reconstruction, frequency-aware, and routing-assisted terms. Ablation results in Figure 10 indicate that reliance solely on the reconstruction loss is limited in restoring subtle structural details. The inclusion of the frequency-aware loss enhances performance by explicitly constraining artifact removal within the spectral domain. Furthermore, the routing-assisted loss refines the model by promoting balanced expert utilization, leading to superior visual fidelity.
Figure 10. Performance of loss function variants evaluated by PSNR (left) and SSIM (right). Implant size decreases along the horizontal axis; colored bars denote the metric values of each variant.

5. Discussion

In medical image processing, wavelet and Fourier transforms offer complementary frequency-domain representations: wavelets capture localized spatial-frequency features and edges, while Fourier analysis characterizes global periodic structures and spectral distributions. While each transform has been utilized independently in CNNs or Transformers for tasks such as denoising, few methods combine them to leverage their inherent complementary strengths. We propose a Hybrid-Frequency-Aware Mixture-of-Experts (HFMoE) model for CT metal artifact reduction (MAR). This model employs a gating mechanism to dynamically combine wavelet and Fourier features, adapting the artifact suppression process to varying frequency content.
Our work has several limitations that suggest directions for future research. First, the model is trained exclusively on synthetic data, which may not fully capture the diverse artifact patterns in real clinical scans, potentially affecting its generalization. This reliance on simulation poses a risk where high quantitative scores on synthetic benchmarks might obscure performance gaps in handling complex, real-world physical effects. Second, our method processes 2D slices independently, without fully exploiting 3D spatial context. Third, imaging parameters (e.g., scanner settings, X-ray energy) that influence artifact formation are not yet incorporated. Future work could (i) use more realistic training data or domain adaptation techniques to improve clinical generalization; (ii) extend HFMoE to 3D to utilize volumetric context; and (iii) incorporate acquisition parameters to enhance robustness across diverse clinical protocols.

6. Conclusions

This paper presents a Hybrid-Frequency-Aware Mixture-of-Experts (HFMoE) model for metal artifact reduction in CT imaging. The core of this work lies in the systematic integration of wavelet and Fourier transforms to exploit their complementary properties: wavelet transforms capture localized spatial-frequency details, while Fourier transforms characterize global spectral structures. The proposed method consists of two key components. First, a hybrid module processes features in parallel through wavelet, Fourier, and combined branches, followed by adaptive fusion. Second, a routing mechanism dynamically selects specialized experts tailored to handle different artifact patterns based on local frequency content. Experiments on synthetic and clinical datasets demonstrate that HFMoE improves artifact reduction and preserves anatomical details more effectively than existing methods, as validated by quantitative metrics and visual inspection. Future work will extend this framework to 3D reconstruction. Furthermore, the multi-frequency approach of HFMoE offers a generalizable framework for other modalities, such as MRI denoising and PET motion correction; this would involve adapting the input encoding for k-space data or re-specializing experts to model modality-specific statistical noise.

Author Contributions

Conceptualization, P.L. and H.Z.; methodology, P.L.; software, P.L.; validation, P.L., H.Z., and F.J.; formal analysis, P.L.; investigation, P.L.; resources, P.L. and H.Z.; data curation, P.L. and C.Z.; writing—original draft preparation, P.L.; writing—review and editing, H.Z. and C.Z.; visualization, H.Z. and F.J.; supervision, H.Z. and F.J.; project administration, H.Z. and F.J.; funding acquisition, F.J. and C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the China Postdoctoral Science Foundation (Grant No. 2024M754208), the National Natural Science Foundation of China (Grant Nos. 62371164 and 62576123), and the Henan Provincial Science and Technology Research Project (Grant No. 252102210147).

Data Availability Statement

All data used in this study are publicly available and can be accessed via the following sources: Synthesized DeepLesion (https://github.com/hongwang01/SynDeepLesion, accessed on 4 July 2024), Clinical SpineWeb (https://csi-workshop.weebly.com/ or https://zenodo.org/records/7049844, accessed on 1 June 2014), and the CLINIC-metal subset from the CTPelvic1K dataset (https://github.com/MIRACLE-Center/CTPelvic1K, accessed on 8 February 2021). No new primary data were generated in this study.

Acknowledgments

All authors declare that they have no known conflicts of interest in terms of competing financial interests or personal relationships that could have an influence or are relevant to the work reported in this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, H.; Li, Y.; Zhang, H.; Chen, J.; Ma, K.; Meng, D.; Zheng, Y. InDuDoNet: An interpretable dual domain network for CT metal artifact reduction. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, Strasbourg, France, 27 September–1 October 2021; pp. 107–118. [Google Scholar]
  2. Kalender, W.A.; Hebel, R.; Ebersberger, J. Reduction of CT artifacts caused by metallic implants. Radiology 1987, 164, 576–577. [Google Scholar] [CrossRef] [PubMed]
  3. Meyer, E.; Raupach, R.; Lell, M.; Schmidt, B.; Kachelrieß, M. Normalized metal artifact reduction (NMAR) in computed tomography. Med. Phys. 2010, 37, 5482–5493. [Google Scholar] [CrossRef] [PubMed]
  4. Tian, C.; Cheng, T.; Peng, Z.; Zuo, W.; Tian, Y.; Zhang, Q.; Wang, F.Y.; Zhang, D. A survey on deep learning fundamentals. Artif. Intell. Rev. 2025, 58, 381. [Google Scholar] [CrossRef]
  5. Wang, H.; Li, Y.; Meng, D.; Zheng, Y. Adaptive Convolutional Dictionary Network for CT Metal Artifact Reduction. In Proceedings of the 31st International Joint Conference on Artificial Intelligence—IJCAI, Vienna, Austria, 23–29 July 2022; pp. 1401–1407. [Google Scholar]
  6. Wang, H.; Xie, Q.; Zeng, D.; Ma, J.; Meng, D.; Zheng, Y. OSCNet: Orientation-Shared Convolutional Network for CT Metal Artifact Learning. IEEE Trans. Med. Imaging 2023, 43, 489–502. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Yu, H. Convolutional neural network based metal artifact reduction in X-ray computed tomography. IEEE Trans. Med. Imaging 2018, 37, 1370–1381. [Google Scholar] [CrossRef]
  8. Ghani, M.U.; Karl, W.C. Fast enhanced CT metal artifact reduction using data domain deep learning. IEEE Trans. Comput. Imaging 2019, 6, 181–193. [Google Scholar] [CrossRef]
  9. Lin, W.A.; Liao, H.; Peng, C.; Sun, X.; Zhang, J.; Luo, J.; Chellappa, R.; Zhou, S.K. DuDoNet: Dual domain network for CT metal artifact reduction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 10512–10521. [Google Scholar]
  10. Zhou, B.; Chen, X.; Zhou, S.K.; Duncan, J.S.; Liu, C. DuDoDR-Net: Dual-domain data consistent recurrent network for simultaneous sparse view and metal artifact reduction in computed tomography. Med. Image Anal. 2022, 75, 102289. [Google Scholar] [CrossRef]
  11. Wang, H.; Yang, S.; Bai, X.; Wang, Z.; Wu, J.; Lv, Y.; Cao, G. IRDNet: Iterative Relation-Based Dual-Domain Network via Metal Artifact Feature Guidance for CT Metal Artifact Reduction. IEEE Trans. Radiat. Plasma Med. Sci. 2024, 8, 959–972. [Google Scholar] [CrossRef]
  12. Zheng, S.; Zhang, D.; Yu, C.; Jia, L.; Zhu, L.; Huang, Z.; Zhu, D.; Yu, H. MAReraser: Metal Artifact Reduction with Image Prior Using CNN and Transformer Together. In Proceedings of the 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Lisbon, Portugal, 3–6 December 2024; pp. 4060–4065. [Google Scholar]
  13. Yao, X.; Tan, J.; Deng, Z.; Xiong, D.; Zhao, Q.; Wu, M. MUPO-Net: A Multilevel Dual-domain Progressive Enhancement Network with Embedded Attention for CT Metal Artifact Reduction. In Proceedings of the ICASSP 2025—2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025; pp. 1–5. [Google Scholar]
  14. Tian, C.; Zheng, M.; Li, B.; Zhang, Y.; Zhang, S.; Zhang, D. Perceptive self-supervised learning network for noisy image watermark removal. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 7069–7079. [Google Scholar] [CrossRef]
  15. Tian, C.; Zheng, M.; Lin, C.W.; Li, Z.; Zhang, D. Heterogeneous window transformer for image denoising. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 6621–6632. [Google Scholar] [CrossRef]
  16. Tian, C.; Song, M.; Zuo, W.; Du, B.; Zhang, Y.; Zhang, S. Application of convolutional neural networks in image super-resolution. CAAI Trans. Intell. Syst. 2025, 20, 719–749. [Google Scholar]
  17. Tian, C.; Xie, J.; Li, L.; Zuo, W.; Zhang, Y.; Zhang, D. A Perception CNN for Facial Expression Recognition. IEEE Trans. Image Process. 2025, 34, 8101–8113. [Google Scholar] [CrossRef] [PubMed]
  18. Lyu, Y.; Lin, W.A.; Liao, H.; Lu, J.; Zhou, S.K. Encoding metal mask projection for metal artifact reduction in computed tomography. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2020, Lima, Peru, 4–8 October 2020; pp. 147–157. [Google Scholar]
  19. Li, Z.; Gao, Q.; Wu, Y.; Niu, C.; Zhang, J.; Wang, M.; Wang, G.; Shan, H. Quad-Net: Quad-domain network for CT metal artifact reduction. IEEE Trans. Med. Imaging 2024, 43, 1866–1879. [Google Scholar] [CrossRef] [PubMed]
  20. Liu, P.; Zhang, H.; Tian, C.; Jiang, F.; Zhang, Y.; Zuo, W. Tri-Domain Filtering Transformer for CT Metal Artifact Reduction. IEEE Trans. Radiat. Plasma Med. Sci. 2025. [Google Scholar] [CrossRef]
  21. Peng, C.; Qiu, B.; Li, M.; Yang, Y.; Zhang, C.; Gong, L.; Zheng, J. GPU-accelerated dynamic wavelet thresholding algorithm for X-ray CT metal artifact reduction. IEEE Trans. Radiat. Plasma Med. Sci. 2017, 2, 17–26. [Google Scholar] [CrossRef]
  22. Liu, P.; Zhang, H.; Zhang, K.; Lin, L.; Zuo, W. Multi-level wavelet-CNN for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 773–782. [Google Scholar]
  23. Cui, Y.; Ren, W.; Cao, X.; Knoll, A. Image restoration via frequency selection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 46, 1093–1108. [Google Scholar] [CrossRef]
  24. Dai, T.; Wang, J.; Guo, H.; Li, J.; Wang, J.; Zhu, Z. FreqFormer: Frequency-aware transformer for lightweight image super-resolution. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24), Jeju, Republic of Korea, 3–9 August 2024; pp. 731–739. [Google Scholar]
  25. Tian, C.; Zhang, X.; Zhang, Q.; Yang, M.; Ju, Z. Image super-resolution via dynamic network. CAAI Trans. Intell. Technol. 2024, 9, 837–849. [Google Scholar] [CrossRef]
  26. Suvorov, R.; Logacheva, E.; Mashikhin, A.; Remizova, A.; Ashukha, A.; Silvestrov, A.; Kong, N.; Goka, H.; Park, K.; Lempitsky, V. Resolution-Robust Large Mask Inpainting With Fourier Convolutions. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 2149–2159. [Google Scholar]
  27. Liu, T.; Li, B.; Du, X.; Jiang, B.; Geng, L.; Wang, F.; Zhao, Z. FAIR: Frequency-aware image restoration for industrial visual anomaly detection. arXiv 2023, arXiv:2309.07068. [Google Scholar] [CrossRef]
  28. Tian, C.; Liu, K.; Zhang, B.; Huang, Z.; Lin, C.W.; Zhang, D. A Dynamic Transformer Network for Vehicle Detection. IEEE Trans. Consum. Electron. 2025, 71, 2387–2394. [Google Scholar] [CrossRef]
  29. Deeba, F.; Kun, S.; Ali Dharejo, F.; Zhou, Y. Wavelet-Based Enhanced Medical Image Super Resolution. IEEE Access 2020, 8, 37035–37044. [Google Scholar] [CrossRef]
  30. Dharejo, F.A.; Zawish, M.; Deeba, F.; Zhou, Y.; Dev, K.; Khowaja, S.A.; Qureshi, N.M.F. Multimodal-boost: Multimodal medical image super-resolution using multi-attention network with wavelet transform. IEEE/ACM Trans. Comput. Biol. Bioinform. 2022, 20, 2420–2433. [Google Scholar] [CrossRef] [PubMed]
  31. Jiang, X.; Zhang, X.; Gao, N.; Deng, Y. When fast fourier transform meets transformer for image restoration. In Computer Vision—ECCV 2024, 18th European Conference, Milan, Italy, 29 September–4 October 2024; Springer: Cham, Switzerland, 2024; pp. 381–402. [Google Scholar]
  32. Zhao, C.; Cai, W.; Dong, C.; Hu, C. Wavelet-based fourier information interaction with frequency diffusion adjustment for underwater image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 8281–8291. [Google Scholar]
  33. Ren, Y.; Li, X.; Li, B.; Wang, X.; Guo, M.; Zhao, S.; Zhang, L.; Chen, Z. MoE-DiffIR: Task-Customized Diffusion Priors for Universal Compressed Image Restoration. In Computer Vision—ECCV 2024, 18th European Conference, Milan, Italy, 29 September–4 October 2024; Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G., Eds.; Springer: Cham, Switzerland, 2024; Volume 15067, pp. 116–134. [Google Scholar]
  34. Mandal, D.; Chattopadhyay, S.; Tong, G.; Chakravarthula, P. UniCoRN: Latent Diffusion-based Unified Controllable Image Restoration Network across Multiple Degradations. arXiv 2025, arXiv:2503.15868. [Google Scholar]
  35. Lin, J.; Zhang, Z.; Li, W.; Pei, R.; Xu, H.; Zhang, H.; Zuo, W. UniRestorer: Universal Image Restoration via Adaptively Estimating Image Degradation at Proper Granularity. arXiv 2024, arXiv:2412.20157. [Google Scholar] [CrossRef]
  36. An, T.; Gao, H.; Liu, R.; Dai, K.; Xie, T.; Li, R.; Wang, K.; Zhao, L. An MoE-Driven Unified Image Restoration Framework for Adverse Weather Conditions. IEEE Trans. Circuits Syst. Video Technol. 2025, in press. [Google Scholar] [CrossRef]
  37. Yang, Z.; Chen, H.; Qian, Z.; Yi, Y.; Zhang, H.; Zhao, D.; Wei, B.; Xu, Y. All-in-one medical image restoration via task-adaptive routing. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2024, 27th International Conference, Marrakesh, Morocco, 6–10 October 2024; Springer: Cham, Switzerland, 2024; pp. 67–77. [Google Scholar]
  38. Deng, Z.; Campbell, J. Sparse mixture-of-experts for non-uniform noise reduction in MRI images. In Proceedings of the Winter Conference on Applications of Computer Vision (WACV), Tucson, AZ, USA, 28 February–6 March 2025; pp. 297–305. [Google Scholar]
  39. Wang, Z.; Ru, Y.; Chetouani, A.; Chen, F.; Bauer, F.; Zhang, L.; Hans, D.; Jennane, R.; Jarraya, M.; Chen, Y.H. MoEDiff-SR: Mixture of Experts-Guided Diffusion Model for Region-Adaptive MRI Super-Resolution. arXiv 2025, arXiv:2504.07308. [Google Scholar]
  40. Wang, Y.; Li, Y.; Zheng, Z.; Zhang, X.P.; Wei, M. M2Restore: Mixture-of-Experts-based Mamba-CNN Fusion Framework for All-in-One Image Restoration. arXiv 2025, arXiv:2506.07814. [Google Scholar] [CrossRef]
  41. Zamfir, E.; Wu, Z.; Mehta, N.; Tan, Y.; Paudel, D.P.; Zhang, Y.; Timofte, R. Complexity experts are task-discriminative learners for any image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 11–15 June 2025; pp. 12753–12763. [Google Scholar]
  42. Tiwari, N.; Hemrajamani, N.; Goyal, D. Improved digital image watermarking algorithm based on hybrid DWT-FFT and SVD techniques. Indian J. Sci. Technol. 2017, 10, 1–7. [Google Scholar] [CrossRef][Green Version]
  43. Tian, C.; Song, M.; Fan, X.; Zheng, X.; Zhang, B.; Zhang, D. A Tree-guided CNN for image super-resolution. IEEE Trans. Consum. Electron. 2025, 71, 3631–3640. [Google Scholar] [CrossRef]
  44. Yu, L.; Zhang, Z.; Li, X.; Xing, L. Deep Sinogram Completion with Image Prior for Metal Artifact Reduction in CT Images. IEEE Trans. Med. Imaging 2020, 40, 228–238. [Google Scholar] [CrossRef]
  45. Liu, P.; Han, H.; Du, Y.; Zhu, H.; Li, Y.; Gu, F.; Xiao, H.; Li, J.; Zhao, C.; Xiao, L.; et al. Deep learning to segment pelvic bones: Large-scale CT datasets and baseline models. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 749–756. [Google Scholar] [CrossRef]
  46. Cheng, T.; Bi, T.; Ji, W.; Tian, C. Graph convolutional network for image restoration: A survey. Mathematics 2024, 12, 2020. [Google Scholar] [CrossRef]
  47. Wang, T.; Xia, W.; Huang, Y.; Sun, H.; Liu, Y.; Chen, H.; Zhou, J.; Zhang, Y. Dual-domain adaptive-scaling non-local network for CT metal artifact reduction. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Springer: Cham, Switzerland, 2021; pp. 243–253. [Google Scholar]
  48. Wang, H.; Li, Y.; He, N.; Ma, K.; Meng, D.; Zheng, Y. DICDNet: Deep Interpretable Convolutional Dictionary Network for Metal Artifact Reduction in CT Images. IEEE Trans. Med. Imaging 2021, 41, 869–880. [Google Scholar] [CrossRef]
  49. Wang, H.; Xie, Q.; Li, Y.; Huang, Y.; Meng, D.; Zheng, Y. Orientation-Shared Convolution Representation for CT Metal Artifact Learning. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Singapore, 18–22 September 2022; pp. 665–675. [Google Scholar]
  50. Wang, H.; Li, Y.; Zhang, H.; Meng, D.; Zheng, Y. InDuDoNet+: A deep unfolding dual domain network for metal artifact reduction in CT images. Med. Image Anal. 2023, 85, 102729. [Google Scholar] [CrossRef]
  51. Liu, X.; Xie, Y.; Diao, S.; Tan, S.; Liang, X. Unsupervised CT metal artifact reduction by plugging diffusion priors in dual domains. IEEE Trans. Med. Imaging 2024, 43, 3533–3545. [Google Scholar] [CrossRef]
  52. BaoShun, S.; ShaoLei, Z.; ZhaoRan, F. Artifact Region-Aware Transformer: Global Context Helps CT Metal Artifact Reduction. IEEE Signal Process. Lett. 2024, 31, 1249–1253. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.