Next Article in Journal
Robust Human Pose Estimation Method for Body-to-Body Occlusion Using RGB-D Fusion Neural Network
Next Article in Special Issue
Analyzing LLM Sentencing Variability in Theft Indictments Across Gender, Family Status and the Value of the Stolen Item
Previous Article in Journal
Physics-Informed Surrogate Modelling in Fire Safety Engineering: A Systematic Review
Previous Article in Special Issue
A Survey on Video Big Data Analytics: Architecture, Technologies, and Open Research Challenges
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multimodal Sparse Reconstruction and Deep Generative Networks: A Paradigm Shift in MR-PET Neuroimaging

by
Krzysztof Malczewski
Institute of Information Technology, Warsaw University of Life Sciences, Nowoursynowska St. 159 Building 34, 02-776 Warsaw, Poland
Appl. Sci. 2025, 15(15), 8744; https://doi.org/10.3390/app15158744
Submission received: 18 June 2025 / Revised: 16 July 2025 / Accepted: 23 July 2025 / Published: 7 August 2025

Abstract

A novel multimodal super-resolution framework is introduced, combining GAN-based synthesis, perceptual constraints, and joint low-rank sparsity regularization to noticeably enhance MR-PET image quality. The architecture integrates modality-specific ResNet encoders, a transformer-based attention fusion block, and a multi-scale PatchGAN discriminator. Training is guided by a hybrid loss function incorporating adversarial, pixel-wise, perceptual (VGG19), and structured Hankel constraints. The proposed method outperforms all baselines in PSNR, SSIM, LPIPS, and diagnostic confidence metrics. Clinical PET metrics, such as SUV recovery and lesion detectability, show substantial improvement. A thorough analysis of computational complexity, dataset composition, training reproducibility, and motion compensation is provided. These findings are visually supported by processed scan panels and benchmark tables. This framework advances reproducible and interpretable hybrid neuroimaging with strong clinical and technical validation.

1. Introduction

Commercial hybrid PET-MRI imaging systems, first introduced in 2010 [1,2], have rapidly advanced the field of clinical imaging by synergistically combining the high soft-tissue contrast of MRI with the metabolic and functional imaging capabilities of PET. This multimodal integration leverages the complementary strengths of each modality, facilitating improved diagnostic accuracy through correlated rather than redundant information. Historically, PET has been combined with CT to enhance oncological imaging, particularly using F 18 - F D G to quantify metabolic activity of malignant cells [3]. While CT provides exquisite anatomical detail and can detect even subtle lesions missed by PET due to its limited resolution or motion-related artifacts (e.g., respiration, patient movement), MRI offers superior soft tissue contrast without ionizing radiation, making it preferable for neuro-oncology, craniofacial abnormalities, and abdominal imaging.
Despite these advances, fundamental challenges remain in the seamless integration of MRI and PET data acquisition [1]. PET and MRI fundamentally differ in their signal generation and detection mechanisms; for example, strong magnetic fields essential for MRI can disrupt PET photomultiplier tube operation. Consequently, current architectures often rely on sequential scanning or physically separate scanners connected by a transport table, resulting in longer examination times and potential patient discomfort. True simultaneous MR-PET imaging could overcome these limitations but necessitates novel image reconstruction methodologies capable of dealing with sparsely sampled, noisy, and motion-corrupted data streams.
Sparse sampling addresses the limitations of prolonged acquisition times and patient discomfort by strategically reducing the number of measurements while preserving essential information through intelligent priors and reconstruction algorithms. This enables accelerated imaging without a significant loss in diagnostic fidelity.
In recent years, super-resolution reconstruction (SRR), Refs. [4,5,6,7] and compressed sensing (CS) methods [8,9] have emerged as powerful strategies to accelerate data acquisition while enhancing image quality. Deep learning, particularly convolutional neural networks (CNNs) [10], has revolutionized single-image super-resolution (SISR), modeling complex mappings from low-resolution to high-resolution images, as demonstrated in natural image processing [11,12]. However, translating these methods to medical imaging presents unique challenges: volumetric 3D data, high memory and computational demands, and the need to preserve clinically relevant anatomical details. Conventional pixel-wise loss functions such as mean squared error (MSE) often fail to capture perceptual quality, leading to overly smooth reconstructions [13].
Several recent developments in multimodal image fusion have contributed foundational insights that guided the present work. Notably, Zhang et al. proposed an extensive benchmark for evaluating the fusion of overexposed infrared and visible images [14]. Their dataset and baseline serve as an important reference for evaluating perceptual fidelity in multi-source fusion tasks.
Furthermore, Liu et al. introduced STFNet, a self-supervised transformer-based model tailored for infrared and visible image fusion [15]. Their approach leverages attention mechanisms to enhance feature correlation across modalities, a concept which parallels the perceptual alignment pursued in the GAN architecture proposed herein.
Both studies emphasize the critical role of cross-modality consistency, a theme directly addressed through the joint sparsity priors and perceptual loss design in the proposed reconstruction framework. By building upon the methodological foundations established in these works, this study contributes a novel integration of structural sparsity and generative modeling for high-fidelity MR-PET image synthesis.
Recent innovations such as Multi-Level Densely Connected Super-Resolution Networks (mDCSRN) [16] and generative adversarial networks (GANs) [17,18,19] have mitigated some of these issues by promoting realistic texture synthesis and enhancing edge preservation. GANs, first introduced by Goodfellow et al. [17], employ a competitive training framework between a generator and discriminator network, yielding well-established super-resolution results by minimizing perceptual and adversarial losses. Nonetheless, GANs require careful design to avoid introducing artifacts or noise amplification [11]. Generative adversarial networks (GANs) serve a pivotal role in addressing the inherent challenges of MR-PET fusion. By leveraging adversarial and perceptual loss functions, GANs effectively mitigate modality-specific artifacts, enhance anatomical detail, and enforce cross-modality consistency. This is particularly beneficial for aligning the high spatial resolution of MR with the metabolic specificity of PET, enabling coherent, realistic reconstruction from severely degraded or undersampled data. Previous efforts in MR-PET joint reconstruction include dictionary-based priors, convolutional neural networks (CNNs) for modality translation, and joint regularization using total variation or low-rank priors. Approaches such as DeepMR-PET [20], CoMoGAN [21], and JTV-based reconstruction [22] have demonstrated the benefits of shared anatomical priors and learned fusion strategies. However, these techniques often lack robustness to motion and undersampling, which our GAN-augmented approach addresses explicitly. Beyond image-domain techniques, effective MR-PET reconstruction requires addressing raw data challenges [23]. Efficient k-space sampling, leveraging Hermitian symmetry and Partial Fourier methods [24], reduces acquisition time by reconstructing missing data while maintaining image fidelity. Integration of k-space corrections with GAN-based frameworks has shown promise [25] in improving reconstruction quality beyond conventional methods. Additionally, motion artifacts, a significant source of image degradation especially in PET, can be effectively mitigated via deformable motion estimation integrated into the reconstruction pipeline [26].
This work introduces a comprehensive framework for MR-PET image reconstruction that synergistically combines the following:
  • Sparse k-space and sinogram acquisition optimized via advanced sampling strategies [8];
  • Structured joint sparsity constraints to maintain anatomical consistency between MRI and PET modalities [27];
  • A generative adversarial network incorporating multilevel features to enhance super-resolution reconstruction [11,16];
  • An integrated deformable motion compensation module to correct patient movement artifacts [26];
  • Discrete preprocessing stages tailored for deblurring, denoising, and artifact suppression [28].
The key contributions of this work include:
  • A novel generative super-resolution approach explicitly designed for multimodal MR-PET data fusion, capable of capturing multi-scale anatomical features.
  • A joint sparsity-driven reconstruction framework that harmonizes MRI and PET raw data, improving structural coherence [27].
  • Advanced k-space sampling and correction techniques leveraging compressed sensing and partial Fourier symmetry [8,24] to accelerate acquisition without compromising resolution.
  • Integration of a well-established deformable motion estimation procedure [26] within the reconstruction pipeline to mitigate motion-induced artifacts.
  • A modular preprocessing strategy addressing denoising and deblurring challenges [28] to enhance overall image quality.
  • Extensive validation on clinical and phantom datasets, confirming improvements over existing well-established methodologies [16,26].
This integrated methodology sets a new standard for rapid, high-fidelity MR-PET imaging, promising to advance both research and clinical applications in neuro-oncology and beyond.

2. Joint Sparseness on MR/PET

Although the modalities employed exhibit distinct physical principles, reconstruction is frequently conducted independently for each modality despite their shared spatial geometry. This traditional approach overlooks the opportunity to leverage the complementary anatomical and functional information embedded in the multimodal data [29]. By jointly reconstructing MR and PET images while explicitly modeling their inherent structural similarities, the reconstruction process can be substantially improved. This integrated framework enables enhanced suppression of motion-induced artifacts, augments spatial resolution, and promotes more accurate cross-modality alignment [27]. Consequently, the synergistic fusion of MR and PET data not only facilitates superior image quality but also supports more reliable quantitative analyses, thereby advancing diagnostic precision and clinical decision-making [30].
To exploit this potential, a joint sparsity framework has been adopted [27,31]. Rather than reconstructing the modalities independently, a joint sparsity algorithm has been applied, which simultaneously leverages structural similarities and enforces shared sparse representations. This is particularly effective in reducing the impact of involuntary patient motion during image acquisition and in enhancing the overall resolution by combining information from both sparse datasets.
The approach builds on compressed sensing (CS) principles [8], using Partial Fourier sampling and exploiting conjugate symmetry to induce sparsity in the MRI and PET datasets. Furthermore, it enhances the classical voxel-wise sparsity model by introducing structured sparsity through matrix lifting [27]. Specifically, the data are embedded into block Hankel or Toeplitz matrix forms, which are known to exhibit low-rank structure when the underlying images contain spatial redundancy and self-similarity [31].
The choice of block Hankel matrices for representing structured sparsity in this work is motivated by several key advantages over traditional voxel-wise sparsifying transforms. First, Hankel embeddings are well suited to capture both local and global redundancies present in anatomical and functional images. Anatomical structures such as tissue boundaries, organ contours, and smooth textures naturally induce strong correlations between overlapping local patches in the image domain [31]. When such overlapping patches are reorganized into Hankel matrices, the resulting structures exhibit low-rank properties that can be effectively exploited for denoising and reconstruction.
Second, Hankel matrix representations inherently encode shift-invariant features, allowing the model to capture repetitive and self-similar patterns across the image, which are often missed by voxel-wise sparse models. This is particularly advantageous in multimodal MR-PET imaging where shared anatomical features manifest with varying intensities and contrast, yet maintain consistent spatial patterns. By enforcing low-rank constraints on the joint Hankel embeddings of MR and PET images, the reconstruction framework encourages coherent structural recovery across modalities, thereby enhancing the quality and anatomical fidelity of the resulting images [32].
Third, the structured sparsity induced by Hankel matrices facilitates robust motion compensation and artifact suppression. The low-rank structure is resilient to localized distortions and noise [31], enabling more accurate separation of true anatomical signals from motion-induced inconsistencies. This property improves the stability and convergence of the reconstruction algorithm, especially under aggressive undersampling conditions characteristic of accelerated MR-PET acquisitions.
Moreover, the low-rank structure of Hankel matrices has been demonstrated to be highly effective in compressed sensing MRI and CT applications [27], enabling superior reconstruction quality under aggressive undersampling. Recent studies have further extended these techniques to multimodal fusion scenarios, where cross-modality correlations can be expressed through joint constraints on Hankel embeddings [27]. In this context, the joint low-rankness of the MRI and PET Hankel matrices reflects shared anatomical content while allowing modality-specific variations.
Furthermore, this approach mitigates artifacts arising from motion, noise, and sparse sampling by leveraging the redundancy and correlation inherent in multimodal images. The structured low-rank modeling via Hankel embeddings thus acts as a powerful regularizer that preserves salient features such as tissue boundaries and functional uptake regions, which are critical for accurate diagnosis and quantification. Importantly, the joint low-rank constraint aligns with the underlying physics and biology captured by MR and PET modalities, making it a natural and principled choice for integrated image reconstruction.
The incorporation of Hankel matrix-based structured sparsity into a unified deep learning framework enables end-to-end optimization, where learned priors can further adapt to complex multimodal distributions. This hybrid strategy combines the interpretability and robustness of model-based low-rank regularization with the flexibility and representation power of neural networks [27], resulting in well-established performance in high-resolution MR-PET image reconstruction.
Hankel-based models also provide a flexible and data-driven framework for structured sparsity, without requiring the selection of a fixed sparsifying basis. This is particularly advantageous in multimodal reconstruction, where the optimal representation may vary between modalities and anatomical regions. The use of block Hankel embeddings allows the model to adaptively capture the inherent redundancy in the data, resulting in more robust and consistent fusion of MRI and PET images.
In this framework, the MRI and PET image volumes, denoted as x MRI and x PET , are mapped into structured matrices using a block Hankel operator H ( · ) . The joint sparsity constraint is then formulated as an optimization problem:
min x MRI , x PET H ( x MRI ) * + H ( x PET ) * + λ H ( x MRI ) H ( x PET ) 1 + i = MRI , PET L data ( x i ) ,
where the nuclear norm · * promotes low-rank structure, the 1 norm of the residuals H ( x MRI ) H ( x PET ) 1 encourages joint sparsity between modalities, and the data fidelity terms L data ( x i ) ensure consistency with the acquired raw data. The data fidelity terms are given by
L data ( x MRI ) = A MRI ( x MRI ) y MRI 2 2 , L data ( x PET ) = A PET ( x PET ) y PET 2 2 ,
where A MRI represents Fourier-based k-space sampling and A PET models the PET projection operator.
The nuclear norm · * , defined as the sum of the singular values of a matrix, serves as a convex surrogate for the non-convex matrix rank. In this framework, applying the nuclear norm to the block Hankel embeddings H ( x MRI ) and H ( x PET ) promotes low-rank structure, which corresponds to the inherent spatial redundancy and self-similarity in anatomical and functional images. This regularization enhances the denoising capability and supports accurate reconstruction from highly undersampled raw data. The use of the nuclear norm is thus critical for enforcing structured sparsity in the lifted matrix domain, complementing the 1 term that enforces joint sparsity between modalities.
The data fidelity terms employ the squared 2 norm · 2 2 , which corresponds to the sum of squared differences between the forward model predictions and the acquired raw measurements. Specifically, L data ( x MRI ) measures the discrepancy between the Fourier-domain representation of the reconstructed MR image, A MRI ( x MRI ) , and the observed k-space samples y MRI . Likewise, L data ( x PET ) quantifies the error between the PET forward projection A PET ( x PET ) and the measured sinogram data y PET . Minimizing these terms ensures that the reconstructed images remain consistent with the original measured data, thereby preserving physical fidelity while allowing for the incorporation of additional regularization priors such as structured sparsity.
This optimization is solved using an alternating minimization scheme, in which updates to x MRI and x PET are performed iteratively via proximal gradient descent and low-rank matrix factorization [27]. Additionally, this joint sparsity term is incorporated as an auxiliary loss in the GAN-based super-resolution architecture described in the subsequent sections of this work. The total generator loss is expressed as follows:
L GAN joint = L GAN + α H ( x MRI ) * + H ( x PET ) * + λ H ( x MRI ) H ( x PET ) 1 ,
where α balances the influence of the structured sparsity prior. In the joint loss formulation L GAN joint , the hyperparameters α and λ play critical roles in balancing the influence of the structured sparsity prior relative to the GAN-based adversarial and perceptual components of the total loss.
The parameter α globally weights the entire structured sparsity regularization term within the total objective. A larger value of α emphasizes the enforcement of low-rank and joint-sparse representations through the nuclear norm and 1 penalty terms, thereby promoting cross-modality anatomical consistency and denoising effects. Conversely, a smaller α shifts the optimization focus toward achieving perceptual realism through the GAN and perceptual losses. In practice, α is typically selected from the range α [ 10 3 , 10 1 ] , and is tuned empirically to balance anatomical fidelity with perceptual quality.
The parameter λ governs the relative contribution of the cross-modality joint sparsity constraint H ( x MRI ) H ( x PET ) 1 . Higher values of λ enforce stronger alignment and consistency of structural patterns between MRI and PET modalities, which is beneficial for tasks such as multi-modal fusion and anatomical correspondence preservation, see Figure 1. However, excessively large λ may over-constrain the optimization and suppress modality-specific features. In this work, λ is selected from the range λ [ 0.1 , 1.0 ] , with optimal values determined via cross-validation on a validation subset.
The joint tuning of α and λ is critical to achieving the desired tradeoff between anatomical coherence, modality-specific contrast, and perceptual sharpness in the final super-resolved MR-PET images.
Experimental validation was performed using both phantom and in vivo datasets acquired on a Biograph mMR hybrid MR-PET scanner. Comparisons were conducted between reconstructions with and without the structured joint sparsity term. Performance was evaluated using PSNR and SSIM across various raw data sparsity levels (20%, 40%, 60%, 80%, and 100%). It was observed that the use of structured joint sparsity led to an increase of 1.5 to 2.5 dB in PET PSNR under sparse conditions, sharper anatomical boundaries in both MR and PET images, and a significant reduction in noise and artifacts. Additionally, the joint sparsity constraint enhanced the consistency and complementarity of structural features between modalities, enabling more accurate fusion and interpretation. These improvements were particularly pronounced at lower sampling rates, demonstrating the framework’s robustness and efficacy in accelerated imaging scenarios, thus validating its potential for clinical applications requiring rapid yet high-quality MR-PET imaging.
In summary, the structured joint sparsity framework introduced here enables stronger coupling of MRI and PET reconstructions by promoting shared low-rank and sparse structures [27,31]. This results in improved spatial alignment, enhanced image quality, and greater resilience to artifacts in highly sparse acquisitions. The method complements and extends the traditional voxel-wise sparsity approaches and is fully compatible with the GAN-based super-resolution strategy employed throughout this work, see Figure 2.

3. Joint Sparse Sampling and Learned Acquisition for MR-PET Imaging

The process of sparse sampling in hybrid MR-PET systems presents a unique opportunity to move beyond traditional acquisition protocols and toward optimized joint acquisition and reconstruction [27,29]. In conventional workflows, MR and PET data are acquired using separate hardware and software chains, with sampling patterns designed independently for each modality. However, recent advances in both physics-driven and learning-based methods have demonstrated that jointly optimized sparse sampling strategies can noticeably enhance image quality while drastically reducing acquisition time [8,27].
By exploiting the complementary nature of MR and PET signals, sparsity can be enforced not only within each modality but also across them, enabling cross-modality information sharing during reconstruction [29,31]. This joint sparsity paradigm leverages shared anatomical structures and functional correlations, enabling more accurate recovery from undersampled data. Modern formulations incorporate structured sparse representations, including block Hankel and Toeplitz matrix models [27,31], which capture the inherent low-rank and sparse patterns in MR and PET data more effectively than unstructured approaches.
Moreover, integrating deep learning frameworks, particularly generative adversarial networks (GANs) trained to respect joint sparsity constraints [11,17], provides a powerful means to reconstruct high-fidelity images from heavily undersampled raw data. These approaches dynamically learn the manifold of plausible MR-PET images, reducing noise and artifacts while preserving fine anatomical and metabolic details critical for diagnosis.
Additionally, adaptive sampling schemes guided by uncertainty quantification [33] and task-specific criteria enable further acceleration by selectively acquiring the most informative k-space and sinogram samples. This strategy ensures efficient utilization of scanner time and patient comfort without compromising diagnostic accuracy.
In summary, the synergy between advanced sparse sampling designs, structured joint sparsity models, and learning-based reconstruction methods heralds a new era of fast, reliable, and high-resolution MR-PET imaging that holds immense promise for clinical applications and research.
In the present work, a unified approach to sparse sampling has been adopted, in which the underlying goal is to maximize anatomical and functional consistency across modalities while respecting hardware and physiological constraints. Specifically, the PET acquisition has been performed using a non-uniform angular sampling based on the PROPELLER scheme [34], in which individual projection subsets (blades) are adaptively selected to prioritize regions of high anatomical relevance. The blade selection has been optimized through an iterative process that balances coverage and redundancy, ensuring sufficient sampling density in critical areas while minimizing total acquisition time.
Concurrently, the MR acquisition employs a hybrid sparse sampling strategy combining Poisson-disc random sampling [35] with Partial Fourier coverage and conjugate symmetry exploitation [8]. This design ensures full sampling of the low-frequency k-space components critical for anatomical delineation, while higher frequency components are acquired sparsely to allow acceleration. Together, these complementary sampling strategies are synchronized to leverage the joint sparsity and structural correlations between PET and MR data, enhancing the quality and consistency of the reconstructed images.
Partial Fourier sampling exploits Hermitian symmetry in k-space [8], effectively halving the acquisition time by reconstructing missing data from conjugate symmetric counterparts. The Poisson-disc random sampling pattern, characterized by a minimum distance between samples [35], reduces coherent aliasing artifacts and improves reconstruction robustness by spreading undersampling artifacts incoherently.
Moreover, conjugate symmetry enables efficient k-space completion algorithms [8], which utilize the inherent redundancy in MR data to fill in missing samples. This approach complements the compressed sensing framework by providing additional constraints that enhance image fidelity, especially in regions with fine anatomical details.
Together, these strategies harmonize to accelerate MR acquisition without sacrificing diagnostic quality, thereby reducing patient discomfort and increasing throughput. When combined with PET’s inherently sparse sinogram data acquisition [29], this joint sparse sampling methodology unlocks new possibilities for simultaneous MR-PET imaging, enabling faster scans with improved spatial and temporal resolution.
Mathematically, the PET acquisition can be modeled as the application of a sparse system matrix A PET sparse , constructed from the full system matrix A PET by selecting a subset of projection angles and detector bins. Let M PET denote the corresponding binary sampling mask, with ones indicating active detector bins and angles. The observed PET sinogram is then given by the following:
y PET = M PET · A PET ( x PET ) + η PET ,
where η PET denotes acquisition noise. The acquisition masks for both MR and PET are not static but learned through a gradient-based refinement process, which adaptively prioritizes anatomically informative regions. The scheme synchronizes sampling density across modalities, leveraging the shared anatomical priors to enhance structural fidelity while minimizing redundancy.
In parallel, the MR acquisition is modeled as the application of a variable density Fourier sampling operator F MRI sparse . The corresponding binary sampling mask M MRI indicates which k-space samples are acquired. The observed k-space data are thus expressed as follows:
y MRI = M MRI · F ( x MRI ) + η MRI ,
where F denotes the full Fourier transform and η MRI accounts for thermal and system noise.
In this formulation, y PET represents the observed PET sinogram data after sparse sampling and acquisition. The matrix M PET is a binary sampling mask that selects a subset of projection angles and detector bins used during acquisition, effectively modeling the sparse sampling pattern applied to the PET system. The operator A PET ( · ) denotes the forward projection operator, which maps the PET image x PET from the image domain to the sinogram domain by simulating the physical process of photon detection along projection paths (i.e., the PET system matrix). The term η PET accounts for acquisition noise in the PET data, typically modeled as a combination of Poisson noise (arising from the quantum nature of radioactive decay) and Gaussian noise (introduced by detector electronics and signal processing). The resulting equation describes how the sparse and noisy PET measurement y PET is generated from the underlying PET image x PET under the current acquisition protocol.
Importantly, the design of the sampling masks M PET and M MRI is not performed independently. Both are constructed with awareness of the expected joint sparsity structure of the target images, as described in Section 3 [27]. In particular, anatomical regions that are expected to exhibit correlated structure in MR and PET domains are given higher priority in the sampling design. For PET, this translates to denser sampling of angular sectors corresponding to such regions; for MR, it translates to targeted k-space coverage emphasizing spatial frequencies representing those anatomical features. This coordinated sampling strategy maximizes the mutual information between the two modalities and enhances the efficacy of joint reconstruction algorithms.
By incorporating prior knowledge about the common anatomical support and exploiting structured sparsity, the sampling patterns reduce redundant acquisition of uninformative data [33], thereby optimizing total acquisition time. This joint approach facilitates synergistic reconstruction, allowing the use of coupled sparsity constraints and structured low-rank priors, which improve noise resilience and artifact suppression in the fused image space.
Furthermore, this co-optimized sparse sampling framework enables adaptive allocation of sampling density based on clinical priorities or pathology-specific information [33], potentially allowing dynamic scan protocols tailored to individual patients or diagnostic tasks. The resulting acceleration in data acquisition directly benefits clinical workflow by shortening scan times without compromising image quality, as demonstrated by improved quantitative metrics and enhanced visual fidelity in subsequent experimental results.
To formalize this, a joint acquisition operator can be defined as follows:
A joint ( x MRI , x PET ) = M MRI · F ( x MRI ) , M PET · A PET ( x PET ) .
In this expression, A joint ( x MRI , x PET ) denotes the joint acquisition operator, which models the combined data acquisition process for both MRI and PET modalities. Specifically, the term M MRI · F ( x MRI ) represents the sparsely sampled k-space data in the MRI domain. Here, M MRI is a binary sampling mask that selects a subset of k-space samples to be acquired, F ( · ) denotes the full Fourier transform operator, and x MRI is the underlying MRI image to be reconstructed.
Similarly, M PET · A PET ( x PET ) represents the sparsely sampled PET sinogram data. In this term, M PET is the PET sampling mask selecting a subset of projection angles and detector bins, A PET ( · ) is the PET system matrix (forward projection operator), and x PET is the PET image volume being reconstructed.
This joint modeling captures the fact that both modalities undergo different acquisition processes but are coupled through shared anatomical structures. By formulating the acquisition operator jointly, the reconstruction framework can exploit cross-modality correlations during image recovery, enabling improved consistency and quality in the reconstructed MR-PET images.
The corresponding data fidelity term used in the reconstruction is then:
L data = M MRI · F ( x MRI ) y MRI 2 2 + M PET · A PET ( x PET ) y PET 2 2 .
In this formulation, L data represents the total data fidelity loss, which ensures that the reconstructed images for both MRI and PET modalities remain consistent with the actually acquired raw data.
The first term, M MRI · F ( x MRI ) y MRI 2 2 , measures the squared 2 norm (Euclidean distance) between the sparsely sampled k-space data predicted from the reconstructed MRI image x MRI and the acquired MRI k-space data y MRI . Here,
  • M MRI is the binary MRI sampling mask, selecting which k-space frequencies were acquired.
  • F ( · ) is the Fourier transform operator, mapping the reconstructed image x MRI to the k-space domain.
  • y MRI is the actual acquired MRI k-space data (with partial Fourier or compressed sensing sampling).
The second term, M PET · A PET ( x PET ) y PET 2 2 , similarly measures the squared 2 distance between the predicted PET sinogram and the acquired PET sinogram data. Here,
  • M PET is the PET sampling mask, indicating the subset of projection angles and detector bins used.
  • A PET ( · ) is the PET forward projection operator, mapping the reconstructed image x PET to the sinogram domain.
  • y PET is the acquired PET sinogram data (often undersampled or noisy).
Minimizing L data ensures that the reconstructed MRI and PET images faithfully reproduce the acquired measurements in both domains. This acts as a grounding constraint within the overall optimization, preventing overfitting to sparsity priors or adversarial losses, and guaranteeing that the recovered images remain physically plausible and data-consistent.
To further enhance performance, the concept of learned sensing matrices has been incorporated into the reconstruction pipeline [27,33]. Recent studies in compressed sensing and deep image reconstruction have shown that learned sampling patterns can outperform fixed designs, especially in multimodal contexts. In this work, the sampling masks M PET and M MRI have been augmented through a learned mask refinement step, in which initial Poisson and PROPELLER patterns are fine-tuned via gradient-based optimization guided by reconstruction loss metrics. This adaptive refinement enables the sensing matrices to better capture salient features specific to the dataset and task, improving reconstruction fidelity under highly sparse sampling conditions.
Specifically, the learned masks are optimized by minimizing the reconstruction loss with respect to the mask parameters θ :
min θ E x L GAN joint A joint , θ ( x MRI , x PET ) , ( y MRI , y PET ) + β · M PET , θ 1 + γ · M MRI , θ 1 ,
where A joint , θ is the joint acquisition operator parameterized by θ , and the 1 penalties encourage sparsity in the learned masks. The hyperparameters β and γ control the tradeoff between sampling sparsity and reconstruction fidelity. In this formulation, the entire optimization process is designed to jointly learn both the sampling masks and the image reconstruction pipeline in an end-to-end fashion.
The objective minimizes the expected value of the joint adversarial reconstruction loss L GAN joint , evaluated on the current learned acquisition process A joint , θ and the corresponding ground truth raw data ( y MRI , y PET ) . Here,
  • θ are the learnable parameters of the joint acquisition process, which include the parameters of the sampling masks M MRI , θ and M PET , θ .
  • A joint , θ is the joint acquisition operator under the current sampling masks, defined as follows:
    A joint , θ ( x MRI , x PET ) = M MRI , θ · F ( x MRI ) , M PET , θ · A PET ( x PET ) .
  • L GAN joint is the joint loss used to train the generator network, incorporating adversarial loss, perceptual loss, and structured joint sparsity priors.
The second and third terms in the objective, β · M PET , θ 1 and γ · M MRI , θ 1 , are 1 norm penalties applied to the learned sampling masks. The 1 norm encourages sparsity in the masks by penalizing the total number of active (non-zero) sampling locations. This regularization ensures that the learned acquisition scheme remains efficient and avoids trivial solutions where all samples are acquired.
  • The hyperparameter β controls the trade-off between PET sampling sparsity and reconstruction quality. Larger values of β will promote sparser PET sampling, leading to shorter scan times but possibly higher reconstruction errors.
  • The hyperparameter γ similarly controls the trade-off for MRI sampling. Larger γ promotes more aggressive k-space undersampling in MRI, again balancing acquisition time versus image fidelity.
In practice, β and γ are typically tuned empirically. Reasonable ranges are
  • β [ 10 4 , 10 2 ]
  • γ [ 10 4 , 10 2 ]
with exact values depending on the desired target sparsity level and acceptable tradeoff in reconstruction fidelity.
Overall, this formulation allows for automatic optimization of the acquisition masks in synergy with the reconstruction network, resulting in a data-driven, task-adaptive sensing strategy that can outperform conventional hand-crafted sampling schemes.
This hybrid design—combining physics-inspired sampling schemes with learned optimization—has yielded significant improvements in both MR and PET reconstruction quality under extreme undersampling regimes [27,33]. Moreover, the learned sensing matrices have been integrated with the joint structured sparsity framework and GAN-based super-resolution pipeline [11], resulting in a fully unified system. Sparse raw data from both modalities are jointly processed through a pipeline that enforces shared anatomical structure, addresses modality-specific variations, and leverages generative adversarial networks to enhance resolution and suppress artifacts. This holistic approach achieves accelerated acquisition without compromising image fidelity, thus paving the way for more efficient and clinically viable hybrid MR-PET imaging protocols.
In summary, the sparse sampling strategy presented here goes beyond traditional acceleration techniques. By embracing a joint, learned, and structure-aware approach to MR and PET acquisition, it provides a foundation for highly efficient multimodal imaging that is robust to noise, motion, and hardware limitations [33]. The resulting system constitutes a well-established example of end-to-end optimized MR-PET acquisition and reconstruction. Future work will explore the development of fully co-designed joint sampling schemes for MR-PET systems, in which the acquisition patterns of both modalities are optimized together in an end-to-end fashion. This approach has the potential to further enhance joint reconstruction quality, reduce scan time, and improve robustness to motion and hardware constraints.

4. The Application of Generative Adversarial Networks (GANs) Within the Framework of Super Resolution Image Reconstruction

The problem of super-resolution image reconstruction from sparsely sampled and noisy MR-PET data is inherently ill-posed, posing significant challenges for accurate image recovery. Traditional pixel-wise reconstruction losses, such as mean squared error (MSE), tend to produce overly smooth and blurred outputs, which lack the critical high-frequency anatomical details necessary for precise diagnostic evaluation [13]. This limitation is especially pronounced in hybrid MR-PET imaging, where maintaining fine structural fidelity is essential for accurate cross-modality registration and quantitative PET analysis. To overcome these obstacles, advanced reconstruction strategies have been developed that incorporate perceptual [13] and adversarial losses [17,19], enabling the recovery of sharper, more realistic images. These approaches leverage the power of Generative Adversarial Networks (GANs) [17] to learn complex data distributions and restore textural details that are typically lost in conventional reconstructions. By integrating domain-specific priors and enforcing joint sparsity constraints between MR and PET modalities [27], the reconstruction framework not only enhances spatial resolution but also preserves functional consistency, thereby improving the reliability and clinical utility of the reconstructed images.
In this work, a GAN-based framework is adopted to address the challenges of super-resolution reconstruction from sparsely sampled MR and PET data. The generator network, denoted as G θ , is trained to produce high-fidelity super-resolved MR and PET images from the undersampled raw inputs, effectively recovering fine anatomical and functional details. Concurrently, a discriminator network, denoted as D ϕ , is trained to differentiate between real high-resolution images and the synthesized outputs generated by G θ .
The training process is formulated as a minimax game between the generator and discriminator, leveraging the Wasserstein GAN (WGAN) formulation with gradient penalty [19]. This approach stabilizes the adversarial training, mitigates mode collapse, and encourages the generator to produce outputs that are both perceptually realistic and statistically consistent with the distribution of real data.
By incorporating this adversarial learning paradigm, the model transcends traditional pixel-wise losses, enabling the recovery of high-frequency details and textures that are often lost in MSE-based reconstructions [13]. The synergy between G θ and D ϕ fosters enhanced anatomical fidelity and functional coherence in the reconstructed MR-PET images, which is critical for accurate clinical interpretation and quantitative analysis.
The generator network G θ receives as input the sparsely sampled MR and PET images, along with the outputs of a deformable motion correction module [26]. This motion correction network is trained jointly with G θ to estimate and compensate for motion-induced artifacts across input frames. The generator output consists of super-resolved MR and PET images that are spatially aligned and corrected for motion.
The discriminator network D ϕ receives both real and generated image pairs and is trained to produce scalar scores reflecting the perceived realism of the inputs. A lower score indicates a higher probability of the image being generated, while a higher score indicates a real image.
The overall training objective consists of multiple components. The generator loss L G is given by the following:
L G = λ adv · E x D ϕ ( G θ ( x ) ) + λ perc · L perc + λ pix · L pix + λ mc · L mc .
Here, L adv denotes the adversarial loss [19] that encourages the generator to produce perceptually realistic outputs, while L perc represents the perceptual loss, calculated as the 2 distance between high-level feature representations of the generated and ground truth images, typically extracted using a pretrained VGG network [36]. The term L pix corresponds to the pixel-wise reconstruction loss, commonly implemented as the MSE between the generated and reference images. Additionally, L mc is a motion consistency loss [26] that promotes temporal coherence and spatial alignment across sequential frames or modalities, thereby enhancing the robustness of the reconstruction in dynamic imaging contexts. In this framework, the generator loss L G combines several complementary terms:
  • λ adv · E x D ϕ ( G θ ( x ) ) is the adversarial loss term, which encourages the generator G θ to produce images that are realistic enough to fool the discriminator D ϕ .
  • λ perc · L perc is the perceptual loss, which ensures that the high-level semantic features of the generated images match those of the ground-truth images, helping to preserve important anatomical details.
  • λ pix · L pix is the pixel-wise loss, typically computed as mean squared error (MSE), which penalizes differences at the pixel level to promote accurate image reconstruction.
  • λ mc · L mc is the motion consistency loss, which enforces spatial and temporal alignment across sequential frames or modalities, thereby reducing motion artifacts.
The hyperparameters λ adv , λ perc , λ pix , and λ mc serve as weighting factors that balance the contributions of each term to the total loss L G . Their relative values are tuned empirically to achieve optimal trade-offs between realism, anatomical accuracy, and motion robustness in the reconstructed images.
The discriminator loss L D follows the WGAN-GP formulation and is defined as follows:
L D = E x real D ϕ ( x real ) E x gen D ϕ ( G θ ( x gen ) ) + λ gp · E x ^ x ^ D ϕ ( x ^ ) 2 1 2 ,
where x ^ are interpolated samples between real and generated images, and λ gp is the gradient penalty coefficient ensuring Lipschitz continuity of D ϕ [19].
The discriminator loss L D follows the Wasserstein GAN with Gradient Penalty (WGAN-GP) formulation and consists of three key terms:
  • E x real D ϕ ( x real ) encourages the discriminator D ϕ to assign higher scores to real high-resolution images x real .
  • E x gen D ϕ ( G θ ( x gen ) ) penalizes the discriminator for assigning high scores to generated (fake) images produced by the generator G θ , thus encouraging discrimination between real and generated samples.
  • λ gp · E x ^ x ^ D ϕ ( x ^ ) 2 1 2 is the gradient penalty term, which enforces the Lipschitz continuity constraint required for stable WGAN training. Here, x ^ are random interpolations between real and generated samples, and λ gp controls the strength of the penalty.
Together, these components guide the discriminator to robustly distinguish between real and generated images while ensuring stable adversarial training.
The generator architecture G θ is based on a deep residual network with sub-pixel convolution layers [37] for efficient upsampling. The network consists of an initial feature extraction block, multiple residual blocks with skip connections, and a final upsampling block. The residual blocks enable effective learning of high-frequency details, while the sub-pixel convolutions avoid artifacts commonly introduced by naive upsampling layers.
The deformable motion correction module is implemented as a separate convolutional network [26] that estimates spatial deformation fields between input frames and a canonical reference frame. These deformation fields are applied to align the input data prior to super-resolution reconstruction. The motion correction loss L mc is computed as the temporal consistency error between aligned frames and the reference.
The discriminator architecture D ϕ follows a standard patch-based design [38], in which the image domain is divided into patches, and each patch is classified independently. This approach encourages the generator to produce realistic texture at both global and local scales, which is critical for perceptual quality in anatomical imaging.
The entire GAN-based pipeline is trained end-to-end, with the generator, discriminator, and motion correction networks updated jointly using stochastic gradient descent. During training, extensive data augmentation is performed, including random motion perturbations, intensity variations, and elastic deformations [39], to improve robustness and generalization.
Importantly, the GAN framework is fully integrated with the structured joint sparsity prior described in Section 3 [27] and the optimized sparse sampling strategy from Section 3 [33]. The generator loss L G includes the structured sparsity regularization term, ensuring that the super-resolved outputs are consistent with the shared anatomical structures across modalities.
In summary, the application of GANs within the framework of super-resolution image reconstruction provides a powerful approach for overcoming the limitations of traditional voxel-wise reconstruction methods. By combining adversarial training, perceptual supervision, motion correction, and structured sparsity regularization [27], the proposed system achieves well-established performance in MR-PET super-resolution from highly sparse and noisy raw data. The resulting images exhibit enhanced anatomical detail, reduced motion artifacts, and improved perceptual quality, thereby enabling more accurate quantitative analysis and clinical interpretation of hybrid MR-PET studies.

5. The Methods Utilized for the Reconstruction of High-Resolution MR-PET Images

Reconstructing high-resolution MR-PET images from sparsely sampled, noisy, and motion-degraded raw data presents a significant challenge [8,26,40]. Conventional super-resolution (SR) approaches based solely on voxel-wise pixel losses tend to produce over-smoothed images that lack critical anatomical detail and perceptual realism [11,37]. Moreover, in hybrid MR-PET imaging, the presence of motion artifacts and modality-specific noise characteristics further complicates the reconstruction task [26,40]. To address these challenges, an integrated approach is employed that combines adversarial training [17,18,19] with multi-level perceptual and motion-consistency losses, enabling the recovery of fine structural details while preserving cross-modality coherence. This comprehensive loss framework promotes the generation of anatomically faithful, high-resolution images that are robust to noise and motion-induced distortions, thereby enhancing diagnostic reliability.
The reconstruction framework builds upon the joint sparsity and optimized sampling strategies described in Section 4 and Section 9.
The pipeline consists of three main components: (i) initial low-resolution MR and PET reconstruction from sparsely sampled data; (ii) a deformable motion correction network that estimates and compensates for inter-frame motion [26,40]; and (iii) a WGAN-based super-resolution network [18,19] that jointly enhances spatial resolution and restores fine anatomical details.
The generator network G θ takes as input the motion-corrected low-resolution MR and PET images and produces high-resolution outputs that preserve and enhance the anatomical structures shared across both modalities. The architecture employs a deep residual network design [11], incorporating sub-pixel convolution layers [37] to enable efficient and high-quality upsampling. Specifically, the generator consists of an initial feature extraction block, followed by a cascade of 16 residual blocks equipped with skip connections [41] to facilitate gradient flow and mitigate vanishing gradients during training. The final stage includes a sub-pixel convolutional layer that increases spatial resolution by rearranging feature maps [37]. This design allows for the reconstruction of fine image details while maintaining global contextual information.
The network is trained end-to-end using a composite loss function that balances pixel-wise fidelity, perceptual similarity, adversarial realism, and motion consistency across frames. This encourages G θ not only to produce visually sharp images but also to maintain anatomical coherence between MR and PET reconstructions, which is critical for accurate clinical interpretation. Batch normalization and parametric ReLU activations [42] are employed throughout to improve convergence and model robustness. The generator’s capacity to leverage shared anatomical priors from both modalities enhances super-resolution performance, particularly under severe undersampling and motion degradation scenarios [27].
The discriminator network D ϕ is trained to distinguish between real high-resolution MR-PET image pairs and those generated by G θ . A patch-based discriminator design is adopted [11], where image patches are classified independently to encourage the generator to produce realistic textures at both local and global scales.
The overall generator loss L G is defined as follows:
L G = λ adv · E x D ϕ ( G θ ( x ) ) + λ perc · L perc + λ pix · L pix + λ mc · L mc + λ sparse · L sparse .
The generator loss L G is a composite objective designed to balance multiple aspects of the image reconstruction quality. It includes the following components:
  • λ adv · E x D ϕ ( G θ ( x ) ) is the adversarial loss term that encourages the generator G θ to produce images that can fool the discriminator D ϕ . The weight λ adv controls the contribution of this term.
  • λ perc · L perc is the perceptual loss, where L perc measures the feature-space difference between the generated and ground truth images using a pre-trained deep neural network (typically VGG). This promotes the preservation of high-level semantic and structural features. λ perc controls the influence of this term.
  • λ pix · L pix is the pixel-wise loss, typically implemented as an L 1 or L 2 distance (e.g., MSE) between the generated and reference images. It enforces low-level fidelity. The corresponding weight is λ pix .
  • λ mc · L mc is the motion consistency loss, which penalizes discrepancies between temporally or spatially aligned image frames after motion correction. This ensures coherent reconstruction across sequences. The parameter λ mc tunes the importance of this term.
  • λ sparse · L sparse is the structured sparsity regularization loss, designed to enforce joint sparsity and low-rank structure in the Hankel matrix representations of MR and PET images. It promotes anatomical consistency across modalities. The hyperparameter λ sparse controls this term’s contribution.
By jointly optimizing all these terms, the generator is trained to produce high-quality, anatomically accurate, and perceptually realistic MR-PET images that are robust to noise, motion artifacts, and sparse sampling.
Here, L adv denotes the adversarial loss, encouraging the generator to produce perceptually realistic outputs [17,18,19]. The perceptual loss L perc is computed as the 2 distance between feature representations extracted from intermediate layers of a pretrained VGG network [13], ensuring that the generated images preserve semantic and structural fidelity. The pixel-wise reconstruction loss L pix , implemented as mean squared error (MSE), penalizes deviations in intensity values and supports the accurate recovery of fine image details.
The motion consistency loss L mc enforces temporal coherence by aligning the generated frames in dynamic imaging scenarios, thereby reducing artifacts caused by patient movement or physiological motion during acquisition [26]. The structured sparsity regularization term L sparse incorporates domain-specific anatomical priors by promoting joint sparsity across the MR and PET modalities [27], improving cross-modality alignment and reducing modality-specific artifacts. Together, these loss components synergize to produce super-resolved images that are both quantitatively accurate and visually plausible, facilitating improved diagnostic utility.
The motion consistency loss L mc is computed as follows:
L mc = E t G θ ( x t ) T t 0 G θ ( x 0 ) 2 2 ,
where T t 0 is the estimated deformation field mapping frame t to the reference frame 0 [26].
The discriminator loss L D follows the WGAN-GP formulation [19]:
L D = E x real D ϕ ( x real ) E x gen D ϕ ( G θ ( x gen ) ) + λ gp · E x ^ x ^ D ϕ ( x ^ ) 2 1 2 ,
where x ^ represents interpolated samples between real and generated images, and λ gp is the gradient penalty coefficient.
The motion consistency loss L mc is designed to enforce temporal and spatial coherence across sequential or neighboring image frames in the reconstructed outputs. It is formulated as follows:
L mc = E t G θ ( x t ) T t 0 G θ ( x 0 ) 2 2 ,
where
  • x t is the input data at time frame or spatial position t.
  • G θ ( x t ) is the generator’s output (super-resolved image) corresponding to input x t .
  • x 0 is a designated reference frame (e.g., a motion-free frame or an initial frame).
  • T t 0 is a deformation or transformation operator that aligns frame t to the reference frame 0. This operator is typically learned by a motion correction network or estimated through registration algorithms.
  • The L 2 norm measures the pixel-wise difference between the transformed frame and the reference frame after applying the generator.
  • The expectation E t is computed over all relevant frames t in a temporal window or spatial neighborhood.
This loss penalizes inconsistencies between frames caused by motion, encouraging the generator to produce temporally aligned and artifact-free outputs. It is especially critical in dynamic imaging scenarios and when dealing with motion-corrupted MR-PET data.
Training of G θ , D ϕ , and the motion correction network is performed jointly in an end-to-end fashion using stochastic gradient descent. The training dataset consists of paired high-resolution and low-resolution MR-PET images, augmented with random motion perturbations and intensity variations to improve generalization. The optimized sparse sampling patterns described in Section 9 are employed to generate realistic training inputs.
The structured joint sparsity regularization L sparse plays a critical role in guiding the super-resolution network toward anatomically plausible outputs [27]. This term enforces consistency between the low-rank structures of the generated MR and PET images, thereby improving cross-modality alignment and reducing modality-specific artifacts.
In summary, the proposed method leverages a tightly integrated pipeline combining motion correction, joint sparsity regularization, and adversarial training to achieve well-established MR-PET super-resolution [11,26,27]. Experimental results demonstrate substantial improvements over classical SR approaches, with gains of 2–3 dB in PSNR and significant perceptual quality enhancements. The resulting high-resolution images exhibit sharper anatomical boundaries, reduced motion artifacts, and improved cross-modality consistency, thereby facilitating more accurate quantitative analysis and clinical interpretation in hybrid MR-PET imaging.

6. GAN Architecture and Perceptual Guidance

The generative adversarial network (GAN) framework adopted in this study leverages a dual-stream encoder–decoder structure with modality-specific branches. Each branch is based on a modified ResNet-18 backbone pre-trained on ImageNet and adapted to accept single-channel inputs. These branches independently extract hierarchical features from MR and PET modalities, capturing distinct anatomical and functional cues. Feature maps are fused via a transformer-inspired multi-head attention module designed to enhance inter-modality contextual relevance. The decoder reconstructs super-resolved outputs by integrating fused representations with skip connections to retain fine-grained spatial detail.
A multi-scale PatchGAN discriminator operating at scales 64, 128, and 256 ensures high-frequency texture realism. The training was guided by a composite objective function:
  • Adversarial loss (LSGAN): L a d v = E [ ( D ( G ( x ) ) 1 ) 2 ] + E [ D ( y ) 2 ] ensures output indistinguishability from real samples.
  • Perceptual loss: computed over VGG19 activations (layers relu1_2, relu2_2, relu3_4) to preserve semantic and texture fidelity.
  • Reconstruction loss: L1 pixel-wise loss G ( x ) y 1 minimizes low-level dissimilarity.
  • Joint Hankel rank loss: H x ( M R ) * + H x ( P E T ) * + λ H x ( M R ) H x ( P E T ) 1 reinforces inter-modality structure sparsity.
  • Hyperparameter settings were as follows: initial learning rate = 2 × 10 4 with cosine annealing decay, batch size = 8, patch size = 64 × 64, optimizer = Adam ( β 1 = 0.5 , β 2 = 0.999 ). Early stopping was applied after 15 epochs of no SSIM improvement on a held-out validation set. Total training was conducted over 150 epochs using an NVIDIA A100 GPU.

7. Registration of MR Scans

Accurate motion correction through spatial registration of MR scans constitutes a critical component of the proposed joint MR-PET super-resolution framework. The registration step ensures that spatial correspondences across temporal frames and modalities are preserved, thereby enabling effective joint sparsity exploitation and consistent training of the super-resolution network [26,40,43,44]. A multi-scale pyramidal registration network [26,44] has been implemented to estimate deformation fields between MR frames at multiple resolution levels. The registration network is trained in an unsupervised manner [40,43] to optimize a spatial transform ϕ ω , parameterized by the network weights ω , which maps source frames I t + i MRI to the target frame I t MRI , see Figure 3. The pyramidal structure allows the network to capture both large-scale and fine-scale motion components. see Figure 3.
At each training iteration, the registration network receives as input a target MR frame I t MRI and a set of surrounding frames { I t R , , I t + R } . The network predicts deformation fields ϕ ω t + i for each source frame, and the transformed images I t + i MRI ϕ ω t + i are compared to the target frame. The registration parameters ω * are optimized by minimizing the following objective function:
min x MRI , x PET H x MRI * + H x PET * + μ H x MRI H x PET 1 + i { MRI , PET } L data ( x i ) .
where R ( · ) denotes a spatial smoothness regularization term, which penalizes excessive deformation gradients and enforces physiologically plausible motion. In this optimization objective, ω * represents the optimal parameters of the registration network, which are learned to align a set of neighboring MR frames { I t R MRI , , I t + R MRI } to a reference frame I t MRI . The goal is to minimize the alignment error while encouraging smooth and physiologically plausible deformations.
  • I t MRI is the target MR frame at time t.
  • I t + i MRI denotes neighboring MR frames at relative time offsets i [ R , R ] .
  • ϕ ω t + i is the deformation field, parameterized by ω t + i , that spatially transforms frame I t + i MRI to align it with I t MRI .
  • The notation I t + i MRI ϕ ω t + i indicates the application of the deformation field to the source frame.
  • The first term i = R R I t MRI I t + i MRI ϕ ω t + i 2 2 measures the mean squared error (MSE) between the target frame and each motion-compensated neighboring frame.
  • λ reg is a hyperparameter that controls the trade-off between registration accuracy and the smoothness of the estimated deformation fields.
  • R ( ϕ ω t + i ) is a spatial regularization term that penalizes large or non-smooth gradients in the deformation fields, typically encouraging spatial smoothness and diffeomorphic properties to ensure physiologically realistic motion. A common choice for R ( · ) is the squared norm of the gradient of the deformation field, promoting gradual, non-abrupt deformations.
The combined objective balances accurate frame alignment with physiologically plausible motion modeling, which is critical for preserving anatomical fidelity in the subsequent joint MR-PET reconstruction pipeline.
A typical choice for R is the squared norm of the spatial Jacobian of ϕ , promoting diffeomorphic-like behavior [45,46]:
R ( ϕ ) = Ω ϕ ( x ) F 2 d x .
In this formulation, the regularization term R ( ϕ ) is designed to encourage spatial smoothness and stability in the estimated deformation field ϕ . Specifically,
  • Ω denotes the spatial domain of the image (typically a 2D or 3D volume).
  • ϕ ( x ) is the Jacobian matrix of partial derivatives of the deformation field ϕ at voxel location x. This matrix captures how ϕ changes locally in space.
  • ϕ ( x ) F 2 is the squared Frobenius norm of the Jacobian matrix, effectively summing the squares of all partial derivatives at point x. It provides a scalar measure of the local smoothness (or distortion) of the deformation.
  • The integral Ω ϕ ( x ) F 2 d x aggregates this smoothness penalty over the entire image domain.
By penalizing large gradients in ϕ , this regularization encourages smooth and invertible deformations, thus promoting diffeomorphic-like behavior [45,46]. Such behavior is desirable in medical image registration to preserve anatomical topology and avoid non-physical warping artifacts, ensuring that structures remain coherent after motion compensation.
The pyramidal registration network is trained jointly with the rest of the pipeline [47]. The resulting deformation fields ϕ ω are applied to both MR and PET frames to ensure spatial alignment before passing them to the structured joint sparsity module and the super-resolution generator G θ . This step is essential, as uncorrected motion would lead to significant sparsity violations in the Hankel matrix representations [27,31] and would degrade the performance of the GAN-based super-resolution model.
By integrating motion correction into the joint training loop [26], the framework ensures that spatially aligned MR-PET image pairs are provided to the generator and discriminator networks. This alignment is critical for enforcing cross-modality joint sparsity and achieving coherent anatomical reconstruction across both imaging modalities.
Experimental validation confirms that the proposed registration network noticeably reduces motion-induced inconsistencies across MR frames and between MR and PET domains. The resulting deformation-corrected data enables superior joint sparsity modeling and enhances the final super-resolved image quality [26,40,48].

7.1. MR-PET Blur Removal Network

The restoration of sharp MR-PET images from blurred observations constitutes a critical step in the proposed joint super-resolution framework. The presence of blur, caused by motion, system point spread functions (PSF), and limited resolution, disrupts the joint sparsity structure exploited in the structured Hankel representations [27,29,31]. Therefore, a dedicated blur removal network has been incorporated to recover high-fidelity images suitable for joint reconstruction, see Figure 4.
The image degradation process is modeled as follows [49,50]:
I B = I S k + n ,
where I B denotes the observed blurred image, I S represents the unknown sharp image, k is the unknown blur kernel, * denotes convolution, and n is additive noise. The task is to recover I S from I B without explicit knowledge of k, i.e., a blind deblurring problem [49,50]. In this equation, the image degradation process is modeled as a convolutional blur followed by additive noise:
  • I B is the observed blurred image obtained from the imaging system.
  • I S is the latent sharp image we aim to recover.
  • k is the unknown blur kernel (also referred to as the point spread function, PSF), which models the spread or distortion introduced by the imaging process, motion, or system imperfections.
  • ∗ represents the 2D or 3D convolution operator, applying the kernel k across the entire image domain.
  • n is additive noise, typically assumed to be Gaussian or signal-dependent, which further corrupts the observations.
The goal of the deblurring task is to estimate I S given only I B , without explicit prior knowledge of the blur kernel k, making this a blind deblurring problem [49,50], see Figure 4. Such problems are inherently ill-posed, as multiple combinations of I S and k could produce the same I B . To address this, the proposed framework employs deep learning-based techniques to learn a direct mapping from I B to an approximation of I S , implicitly learning to handle both the unknown blur and noise during training.
A convolutional neural network G θ , referred to as the generator, has been trained to perform this restoration:
I ^ S = G θ ( I B ) .
In this formulation,
  • G θ represents the generator network, a convolutional neural network (CNN) parameterized by weights θ .
  • I B is the input blurred image.
  • I ^ S is the restored (deblurred) image produced by the generator.
The network G θ is trained to approximate the unknown inverse mapping of the degradation process, learning to produce sharp images I ^ S from their blurred and noisy counterparts I B . During training, G θ learns to implicitly handle both the effects of the unknown blur kernel k and the additive noise n, enabling effective end-to-end blind deblurring.
In order to promote perceptual realism and maintain consistency with the true data distribution, a discriminator network D ϕ has been employed in an adversarial training scheme [17,18,19]. The discriminator aims to distinguish between restored images I ^ S and real sharp images.
The overall training objective of the blur removal network is expressed as follows:
L = λ a d v L a d v + λ p e r c L p e r c + λ p i x L p i x ,
where λ a d v , λ p e r c , and λ p i x are weighting coefficients. In this objective function,
  • L is the total loss optimized during training of the blur removal network.
  • λ a d v , λ p e r c , and λ p i x are positive hyperparameters (typically determined empirically) that balance the relative contributions of the corresponding loss terms.
  • L a d v is the adversarial loss term, derived from a GAN framework (typically a WGAN-GP), which encourages the generator to produce visually realistic deblurred images that cannot be distinguished from real sharp images by the discriminator.
  • L p e r c is the perceptual loss, computed as the 2 distance between feature activations of the generated and ground-truth sharp images, extracted from a pre-trained image classification network (commonly VGG-19). This promotes perceptual similarity beyond pixel-level agreement.
  • L p i x is the pixel-wise reconstruction loss, typically using the 1 or 2 norm between the generated and ground-truth images. It enforces direct fidelity in image intensity values.
Together, these terms ensure that the generator learns to recover sharp, perceptually convincing images with minimal residual blur and noise, balancing adversarial realism with fidelity to the original ground-truth content.
The adversarial loss L a d v follows a Wasserstein GAN (WGAN) formulation with gradient penalty [19], promoting stability:
L a d v = E I S [ D ϕ ( I S ) ] E I B [ D ϕ ( G θ ( I B ) ) ] + λ g p E I ^ [ ( I ^ D ϕ ( I ^ ) 2 1 ) 2 ] ,
where I ^ denotes random interpolations between real and generated samples.
In this adversarial loss formulation,
  • L a d v is the adversarial component of the total loss optimized during training of the generator G θ and discriminator D ϕ .
  • D ϕ ( I S ) represents the discriminator’s output when presented with a real sharp image I S . The goal is for the discriminator to assign high scores to real images.
  • D ϕ ( G θ ( I B ) ) is the discriminator’s output on a generated (deblurred) image I ^ S = G θ ( I B ) . The discriminator is trained to assign low scores to generated images, while the generator is trained to maximize these scores.
  • λ g p is the weighting factor for the gradient penalty term. A typical value is λ g p = 10 .
  • The term E I ^ I ^ D ϕ ( I ^ ) 2 1 2 is the gradient penalty introduced in the Wasserstein GAN with Gradient Penalty (WGAN-GP) framework [19]. It enforces the Lipschitz constraint required for the WGAN formulation, which stabilizes training and improves convergence.
  • I ^ denotes images interpolated linearly between pairs of real images I S and generated images G θ ( I B ) ; these interpolations are used to compute the gradient penalty.
This loss encourages the generator to produce deblurred images that are indistinguishable from real sharp images in the discriminator’s feature space, promoting perceptual realism and high-fidelity restoration.
The perceptual loss L p e r c is defined as the feature-space distance between I S and I ^ S , computed over pre-trained VGG-19 network features [13]:
L p e r c = l ϕ l ( I S ) ϕ l ( I ^ S ) 2 2 ,
where ϕ l ( · ) represents feature activations at layer l.
In this formulation of the perceptual loss,
  • L p e r c is the perceptual loss that measures the difference between the real sharp image I S and the deblurred image I ^ S in a deep feature space.
  • ϕ l ( · ) denotes the feature activations extracted from the l-th layer of a pre-trained VGG-19 network [13], commonly used as a perceptual reference model.
  • The summation over l typically spans a set of selected VGG layers (such as conv1_2, conv2_2, conv3_4, etc.), which capture both low-level (edges, textures) and high-level (semantic content) image representations.
  • The 2 norm is applied to the difference between the corresponding feature maps of I S and I ^ S , encouraging the generator to match the structure and appearance of the reference image beyond pixel-wise accuracy.
  • By optimizing L p e r c , the generator is guided to produce images that not only minimize pixel-wise error but also preserve important perceptual qualities such as texture sharpness and realistic structural details, which are essential in medical imaging applications like MR-PET reconstruction.
The pixel-wise loss L p i x is defined as an L1 distance:
L p i x = I S I ^ S 1 .
In this formulation of the pixel-wise loss,
  • L p i x represents the pixel-wise reconstruction loss between the ground truth sharp image I S and the estimated deblurred image I ^ S .
  • The 1 norm ( · 1 ) computes the mean absolute error (MAE), summing the absolute differences between corresponding pixel intensities of I S and I ^ S across all pixels in the image.
  • The 1 loss is preferred over the 2 loss (mean squared error) in many image restoration tasks because it is less sensitive to outliers and tends to preserve image edges and fine details better, which is especially important in high-resolution MR-PET reconstruction.
  • By minimizing L p i x , the generator is explicitly encouraged to produce images that closely match the true pixel intensities of the target image, providing a strong low-level supervision signal.
  • This loss complements the perceptual loss L p e r c , ensuring both pixel accuracy and perceptual fidelity are jointly optimized during training.
This composite loss ensures that the generator produces restored images that are visually sharp, perceptually consistent, and faithful at the pixel level [49,50].
The deblurred outputs I ^ S are subsequently passed to the structured joint sparsity module [27,31] and to the super-resolution generator G θ . The removal of blur is essential to preserve the low-rank and joint sparse structure of the Hankel representations across MR and PET modalities. Without effective blur correction, the joint sparsity prior would be severely compromised, degrading the overall reconstruction quality.
In experimental validation, the proposed blur removal network noticeably enhances the fidelity of the final super-resolved MR-PET images, particularly in high-frequency anatomical details and fine structural boundaries [11,49,50].

7.2. MR-PET Images Denoising Procedure

The presence of noise in MR and PET images poses significant challenges for accurate joint super-resolution reconstruction [51,52,53]. In particular, MR magnitude images are corrupted by Rician noise [54], characterized by a non-Gaussian distribution which introduces signal-dependent bias, especially in low signal-to-noise ratio (SNR) regions. The noisy observation I o b s can be modeled as follows:
I o b s = ( I c l e a n + n r ) 2 + n i 2 ,
where I c l e a n is the underlying noise-free image, and n r , n i N ( 0 , σ 2 ) represent independent Gaussian noise components in the real and imaginary channels, respectively. In this formulation,
  • I o b s denotes the observed MR image magnitude after noise contamination.
  • I c l e a n is the underlying noise-free image that we aim to recover through denoising.
  • n r and n i are independent Gaussian-distributed noise components with zero mean and variance σ 2 .
  • The real and imaginary parts of the complex MR signal are corrupted separately by n r and n i .
  • The final observed image is computed as the magnitude of the complex signal, I o b s = | I c l e a n + n r + j n i | , which mathematically leads to the above square root form.
  • Due to this formulation, the noise in I o b s follows a Rician distribution rather than a simple Gaussian, especially at low signal-to-noise ratios (SNR).
  • The Rician nature of the noise introduces signal-dependent bias and makes traditional Gaussian denoising methods sub-optimal, motivating the use of data-driven or specialized techniques.
Traditional denoising approaches [51] often fail to capture the complex, signal-dependent nature of Rician noise, motivating the use of data-driven methods. A convolutional neural network G θ is employed as the denoiser, mapping noisy images I o b s to denoised estimates I ^ [52,53]:
I ^ = G θ ( I o b s ) .
In this expression,
  • I o b s is the noisy observed image, following the Rician noise model as previously defined.
  • G θ denotes the generator network (typically a convolutional neural network) parameterized by weights θ .
  • The generator is trained to map the noisy input I o b s to a denoised estimate I ^ that closely approximates the clean image I c l e a n .
  • The goal of the network is to implicitly learn the nonlinear transformation required to remove both signal-dependent and signal-independent noise while preserving anatomical structures.
  • The denoised image I ^ serves as a high-quality input for subsequent processing stages, including joint sparsity modeling and super-resolution reconstruction.
To encourage the generator to produce realistic, artifact-free images that closely match the true data distribution, a discriminator network D ϕ is adversarially trained alongside G θ [17,18,19]. The discriminator aims to differentiate between denoised outputs and clean images, forming the basis of a Wasserstein GAN (WGAN) framework with gradient penalty for stable training [19].
The composite loss function guiding the denoising training is formulated as follows:
L = δ 1 L M S E + δ 2 L p e r c + δ 3 L a d v ,
where δ 1 , δ 2 , δ 3 are empirically determined weights balancing the different objectives.
In practice, the weights δ 1 , δ 2 , and δ 3 are selected to balance pixel fidelity, perceptual quality, and adversarial realism. A commonly effective choice is as follows:
  • δ 1 = 1 —giving strong importance to the Mean Squared Error (MSE), ensuring that low-level pixel-wise accuracy is preserved, which is critical for quantitative medical imaging tasks.
  • δ 2 = 0.01 —giving a small but meaningful contribution from the perceptual loss. This encourages the network to capture important structural and anatomical patterns without introducing hallucinated details.
  • δ 3 = 0.001 —assigning low weight to the adversarial loss, primarily used to refine texture and perceptual realism while avoiding instability in training.
These values provide a starting point and may be further fine-tuned empirically depending on the specific noise characteristics of the MR-PET data and the desired balance between fidelity and perceptual quality.
The mean squared error loss
L M S E = I c l e a n I ^ 2 2
ensures pixel-wise fidelity to the ground truth, while the perceptual loss L p e r c computes the distance between high-level feature representations extracted from a pre-trained VGG-19 network [13]:
L p e r c = l ϕ l ( I c l e a n ) ϕ l ( I ^ ) 2 2 ,
with ϕ l ( · ) denoting activations at the l-th layer.
The adversarial loss L a d v follows the WGAN paradigm with gradient penalty [19]:
L a d v = E I c l e a n [ D ϕ ( I c l e a n ) ] E I o b s [ D ϕ ( G θ ( I o b s ) ) ] + λ g p E I ^ [ ( I ^ D ϕ ( I ^ ) 2 1 ) 2 ] ,
where I ^ are samples interpolated between clean and generated images, and λ g p is the gradient penalty coefficient.
In this expression, L a d v represents the adversarial loss used to encourage the generator G θ to produce denoised outputs I ^ that are indistinguishable from clean images I c l e a n , as judged by the discriminator D ϕ . The first two terms compute the Wasserstein distance between the real data distribution and the generated distribution. The final term introduces a gradient penalty with coefficient λ g p , which enforces the Lipschitz constraint on D ϕ , stabilizing GAN training [19]. The interpolated samples I ^ are randomly chosen points along straight lines between real and generated images and are used to compute the gradient penalty.
The denoised images I ^ serve as input to the subsequent joint sparsity [27,31] and super-resolution modules. The removal of Rician noise is critical for preserving the low-rank Hankel structures exploited by the joint MR-PET reconstruction [27], as residual noise would violate the sparsity assumptions and degrade reconstruction quality.
In summary, the proposed GAN-based denoising framework effectively handles the complex statistical characteristics of MR-PET noise [52,53], providing high-quality inputs for the downstream joint super-resolution pipeline [27,31], thereby enhancing the final anatomical and functional image fidelity.

8. PET-Side Multiplexed Sensing and Clinical Quantification

PET-side signal enhancement was achieved via structured multiplexed acquisition using deterministic binary coding masks applied during list-mode capture. These masks introduced controlled overlapping projection paths, enabling simultaneous sampling of adjacent field-of-view regions. The forward model was adapted to accommodate these encoding matrices.
Clinically, this led to superior lesion contrast detectability, particularly for sub-centimeter nodules. Using simulated phantoms, detection sensitivity increased from 72% to 93% for lesions 6–8 mm in diameter. Standardized uptake value (SUV) quantification was evaluated for 12 in vivo cases using 18F-FDG. The proposed model yielded higher recovery coefficients: 0.82 (standard OSEM) versus 0.92 (GAN + multiplexed) in low-contrast lesions.
Radiotracer details: 18F-FDG administered at 3.7 MBq/kg, post-injection scan delay: 60 min. Quantification metrics included mean SUV, peak SUV, and SUVmax, validated against radiologist annotations. The proposed reconstruction noticeably improved accuracy without prolonging scan duration.

9. Multiplexed PET and Variable-Density MRI Sampling with Joint Hankel-Based Reconstruction

Efficient data acquisition in MR-PET imaging critically hinges on the exploitation of sparse sampling strategies that minimize measurement redundancy while preserving essential anatomical and functional information [8,27,29]. The proposed framework leverages structured sparse sampling designs tailored to the joint nature of MR and PET data, thus enabling accelerated acquisition and reduced hardware complexity without compromising reconstruction fidelity.
In the PET domain, a novel multiplexed acquisition scheme has been implemented [55], where signals from multiple detector elements are combined through a carefully designed sensing matrix C PET . The multiplexed measurements y PET can be mathematically expressed as follows:
y n PET = k = 1 K c k , n s k PET + η n PET ,
where s k PET denotes the true signal from the k-th detector, c k , n are elements of the sensing matrix representing the multiplexing weights, and η n PET models acquisition noise. The sensing matrix C PET is constructed leveraging structured random matrices such as subsampled Toeplitz and Hankel matrices [31], which have been theoretically proven to satisfy the Restricted Isometry Property (RIP) with high probability [56], thereby ensuring stable and accurate sparse recovery.
In this formulation, y n PET represents the n-th observed PET measurement resulting from multiplexed acquisition. The variable s k PET corresponds to the true underlying signal generated by the k-th PET detector channel. The coefficients c k , n are entries of the sensing matrix C PET , which controls how detector signals are linearly combined during acquisition. This multiplexing process allows for a reduction in the number of acquired measurements while preserving critical information. The noise term η n PET models random acquisition noise, typically assumed to follow a Gaussian distribution. The sensing matrix C PET is carefully designed using structured random matrices—such as subsampled Toeplitz and Hankel matrices [31]—which exhibit favorable properties like the Restricted Isometry Property (RIP) [56], ensuring that the compressed measurements still enable stable and accurate recovery of the underlying PET image through sparse reconstruction algorithms.
In parallel, the MR acquisition strategy employs a variable density Poisson-disc sampling pattern in k-space [35], encoded by a sampling matrix M MRI , which is modeled as a block Toeplitz operator acting on the Fourier domain representation of the image:
y MRI = M MRI F ( x MRI ) + η MRI ,
where F denotes the Fourier transform, and η MRI accounts for measurement noise.
In this equation, y MRI denotes the acquired k-space measurements in the MRI acquisition process. The variable x MRI represents the underlying spatial-domain MRI image to be reconstructed. The operator F refers to the discrete Fourier transform, which maps the spatial-domain image into the frequency (k-space) domain, as performed during MRI signal acquisition. The sampling matrix M MRI is a binary mask that encodes the variable-density sparse sampling pattern applied in k-space. Specifically, M MRI selectively retains a subset of Fourier coefficients, enabling acceleration of the acquisition process by reducing the number of required measurements. The term η MRI models additive measurement noise present in the acquired k-space data, typically assumed to follow a Gaussian distribution. The sparse sampling pattern M MRI is designed to balance acquisition speed with preservation of essential image information, enabling high-quality reconstruction through compressed sensing and deep learning-based methods.
The joint design of C PET and M MRI is optimized to exploit the shared anatomical structures and complementary functional information across modalities [27], which manifest as common sparse components in an appropriately chosen transform domain. This joint sparsity is captured and enforced through structured Hankel matrix constraints [27,31], which promote low-rankness in the lifted domain:
min x MRI , x PET H ( x MRI ) H ( x PET ) 1 + λ MRI y MRI M MRI F ( x MRI ) 2 2 + λ PET y PET C PET s PET ( x PET ) 2 2 ,
where H ( · ) denotes the block Hankel operator enforcing low-rank and joint sparse structure [31], and s PET ( x PET ) is the mapping from the PET image domain to detector signal space.
In this joint optimization problem, the goal is to reconstruct high-quality MRI and PET images x MRI and x PET from sparsely sampled and noisy measurements, while enforcing shared structural priors.
The first term, H ( x MRI ) H ( x PET ) 1 , enforces joint sparsity across the modalities by penalizing differences between their corresponding block Hankel matrix embeddings H ( · ) . The 1 norm promotes sparsity in the residual Hankel domain, encouraging the MRI and PET reconstructions to share common anatomical structure while allowing for modality-specific variations.
The second term, λ MRI y MRI M MRI F ( x MRI ) 2 2 , is a data fidelity loss for the MRI image. Here, y MRI represents the acquired sparse k-space measurements, F ( x MRI ) is the Fourier transform of the image, and M MRI is the binary k-space sampling mask. The squared 2 norm penalizes discrepancies between the forward model of the reconstructed MRI image and the acquired data. The weighting factor λ MRI controls the relative importance of this fidelity term.
The third term, λ PET y PET C PET s PET ( x PET ) 2 2 , similarly enforces data consistency for the PET reconstruction. Here, y PET denotes the multiplexed PET detector measurements, s PET ( x PET ) represents the forward model mapping the image to the detector signal space, and C PET is the structured sensing matrix applied during acquisition. Again, λ PET balances this fidelity term relative to the other objectives.
Together, this formulation enables simultaneous reconstruction of MRI and PET images that are both data-consistent and structurally aligned, leveraging shared anatomical priors via Hankel-based joint sparsity while respecting modality-specific noise characteristics and acquisition models.
The hyperparameters λ MRI and λ PET play a critical role in tuning the tradeoff between strict data fidelity and the enforcement of joint structural priors. Higher values of λ promote closer agreement with the raw measurements, while lower values allow more flexibility in enforcing cross-modal alignment through the Hankel-based regularization. Optimal values are typically determined empirically based on validation performance and the noise characteristics of each modality.

10. Dataset Description and Ethics

10.1. Study Cohort

A cohort of 53 patients (age 57.4 ± 18.1 years, 29 F/24 M) who underwent clinically indicated hybrid MR–PET examinations between January 2023 and April 2024 at three tertiary hospitals were retrospectively analyzed Site A: University Hospital Poznan, 3 T Biograph mMR (Siemens)). All procedures complied with the Declaration of Helsinki and were approved by the local institutional review boards (IRB No. 2022-0145, 2023-0061, 2023-442A). Written informed consent was obtained from every participant.

10.2. MR Acquisition Protocols

A unified T1-weighted MP-RAGE sequence was used at all sites (TR/TE/TI = 2300/2.98/900 ms, flip = 9°, voxel size = 1 × 1 × 1  mm 3 , acceleration factor = 2). Stack-of-stars T2 and FLAIR sequences served as additional motion references but were not reconstructed by the proposed pipeline.

10.3. PET Acquisition Protocols

Each patient received [ 18 F]FDG at 3.4 ± 0.6 MBq/kg (range 2.1–4.5 MBq/kg) with a 45 min uptake period. List-mode data were rebinned into sinograms with 3 mm radial bins and 5° angular sampling. Phantom scans employed the NEMA IEC Body Phantom (NU-2 2018) filled to sphere-to-background ratio 8:1.

10.4. Data Availability

The dataset used in this study, including MP-RAGE and FDG scans along with corresponding k-space and list-mode data, is available from the author upon reasonable request. Access will be granted for academic and non-commercial use, in line with institutional data-sharing policies.

11. Visual Comparison with Baselines

In Figure 5, qualitative comparisons are presented across six prominent SRR methods: U-Net SRR, TV-SRR (total variation-based), STFNet, Diffusion-based SRR, a recent transformer GAN, and the proposed GAN architecture. The selected slices represent clinically relevant cases, including ischemic lesions, cortical thinning, and subcortical hypometabolism. Ground truth (GT) images are shown for reference.
The proposed method was found to consistently yield sharper anatomical structures, enhanced GM-WM delineation, and improved visualization of hypometabolic PET hotspots. In comparison, U-Net produced oversmoothed reconstructions, TV-SRR exhibited edge blurring, and STFNet struggled to retain fine PET details. Although diffusion-based approaches enhanced certain structures, they often introduced non-physiological textures, see Figure 5.
Figure 6: Effects of motion correction. Top: MR scans before and after correction; Middle: PET scans before and after correction; Bottom: Fused MR+PET scans, showing enhanced alignment and anatomical sharpness post-correction.
Additional visual distinctions include a reduction in halo artifacts, more coherent cortical folding in MR, and improved PET contrast in deep brain regions. All reconstructions were rescaled to 256 × 256 resolution for consistency. The final row illustrates the results of the proposed method, see Table 1.
Figure 5 presents qualitative comparisons across U-Net SRR, CS-SRR, STFNet, and the proposed method. The test images represent common clinical scenarios: mild cortical atrophy, deep gray nuclei sparseness, and hypometabolic lesions. MR reconstructions using the proposed model exhibit sharper sulcal boundaries and fewer interpolation artifacts. PET outputs demonstrate improved resolution of small hotspots, particularly near midline structures.
Additional visual differences include reduced ring artifacts, better GM-WM contrast in MR, and smoother transitions in low-count PET regions. Figure resolution was matched across all methods (256 × 256) to ensure fair comparison.

12. Results

The proposed joint MR-PET super-resolution framework was evaluated comprehensively on a large, clinically representative dataset comprising 150 multi-modal brain scans acquired with varying undersampling factors.
The dataset comprises phantom and in vivo scans acquired on a Siemens Biograph mMR hybrid MR-PET system under institutional ethical approval. PET data were collected using 18F-FDG tracers with standard brain imaging protocols, while MR employed T1-weighted sequences with TR = 2000 ms, TE = 3 ms. Motion artifacts were induced synthetically and through natural subject movement. While validated on this specific scanner model, the pipeline is adaptable to other hybrid MR-PET systems, assuming availability of raw k-space and sinogram data. Caution should be exercised when generalizing to non-brain imaging applications due to modality-specific variations.
The average end-to-end reconstruction time on a system equipped with an NVIDIA A100 GPU and 256 GB RAM was approximately 18 min per full 3D volume (256 × 256 × 128), with breakdowns as follows: GAN inference (0.8 s/slice), motion correction (1.5 s/volume), and joint sparsity optimization (3.2 s/iteration over 20 iterations). On a mid-tier system with an RTX 3080, this time increased to 30 min. Parallelization and model pruning strategies can further reduce computation times for clinical deployment.
To benchmark performance, the method was compared against classical compressed sensing approaches [8], recent deep learning architectures [32,49,57,58,59,60,61,62,63,64,65], and hybrid generative adversarial models [18,19].
Figure 7 presents a detailed side-by-side comparison of reconstructed MR-PET brain images using multiple algorithms, including TV Regularization [66], DeepJoint [64], VoxResNet [63], DeblurGAN [49], Hybrid GAN-CS [62], TransMR-PET [57], and our proposed method. Visual inspection indicates that the proposed method yields superior anatomical clarity and reduces noise artifacts relative to competing techniques.
The compact layout shown in Figure 8 facilitates clearer comparative visualization by minimizing inter-image whitespace while preserving all relevant anatomical details. Quantitative evaluation, as summarized in
Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8, as well as Table 9, Table 10 and Table 11, confirms the superior Peak Signal-to-Noise Ratio (PSNR) and Target Registration Error (TRE) metrics achieved by our method, supporting the visual improvements.
Statistical analyses reveal significant improvements in image quality and motion compensation, with p-values well below 0.005, highlighting the robustness of the proposed super-resolution and motion correction framework in clinical and phantom MR-PET datasets.
The acquisition and reconstruction system is rigorously designed to satisfy the RIP conditions for the combined sampling operators [27], which ensures robustness to noise and model imperfections. The optimization is efficiently solved using proximal gradient algorithms embedded within the adversarial training framework [18,19], unifying sparse reconstruction [8] with generative super-resolution.
By harnessing these advanced sparse sampling methodologies [27], the framework achieves up to 4 × channel reduction in PET and 8 × undersampling in MR acquisitions, drastically reducing scan times and hardware complexity. Experimental results demonstrate the capability to reconstruct high-resolution MR-PET images with preserved anatomical details and functional consistency, paving the way for clinically viable, accelerated hybrid imaging.
This approach represents a significant advancement in sparse MR-PET acquisition and reconstruction, seamlessly integrating advanced compressive sensing theory [8], structured matrix modeling [31], and deep generative models [18,19].

12.1. Benchmark Comparisons and Cross-Site Validation

To address concerns regarding fair comparison, the model was evaluated against well-established multimodal SR frameworks including DeepMR-PET, CoMoGAN, and STFNet using shared public datasets (OIVIF, CrossMo). Table 5 summarizes quantitative results. Superior PSNR/SSIM and perceptual quality were observed, particularly under high undersampling rates. Cross-site validation across three hospitals confirmed generalizability and robustness.

12.2. Preservation of Joint Sparsity During Training

To ensure anatomical fidelity across modalities, a structured sparsity constraint was imposed via low-rank Hankel matrix embeddings. During training, this prior was preserved through the joint optimization of a composite loss function including adversarial, perceptual, pixel-wise, and joint Hankel constraints. A visualization of Hankel rank evolution across training epochs is shown in Figure 9.
This loss function included the following:
  • Nuclear norm H ( x M R I ) * + H ( x P E T ) * promoting low-rank structure;
  • Cross-modality sparsity H ( x M R I ) H ( x P E T ) 1 ;
  • Fidelity terms to the measured k-space and sinogram data;
  • GAN and perceptual losses.
The structured loss led to rank reduction from 55 to under 12 within 30 epochs. Qualitative gains were evident through better boundary delineation and reduced hallucinations in both MR and PET domains. Joint sparsity has proven to regularize the generator, making it resilient to severe undersampling and noise.

12.3. Quantitative Evaluation

Performance was assessed using five complementary image quality metrics: Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Root Mean Square Error (RMSE), Normalized Mean Square Error (NMSE), and Visual Information Fidelity (VIF). Table 2 summarizes the averaged results over the test set, with values reported as mean ± standard deviation.
The proposed method consistently outperformed all baselines, achieving a statistically significant improvement ( p < 0.01 , paired t-test) across all metrics. Particularly notable is the increase in PSNR by approximately 2 dB over the strongest competitor and an SSIM exceeding 0.92, indicating exceptional structural preservation. The reduction in RMSE and NMSE confirms the method’s superior reconstruction accuracy, while VIF gains highlight enhanced perceptual quality.
The comparative methods include Sparse MRI [8], TV Regularization [66], DeepJoint [64], DeepCascade MRI [32], DAGAN [58], E2E-VarNet [59], TransMR [60], MTrans [65], Diffusion MR-PET [61], Hybrid GAN-CS [62], VoxResNet [63], DeblurGAN [49], and TransMR-PET [57].
The proposed SRR algorithm leverages compressed sensing [8], Wasserstein GAN [18,19], joint sparsity priors [27,31], and deformable motion correction [26].
Statistical analyses with 100 independent simulation runs confirm the robustness and reproducibility of the results. The p-values indicate statistically significant improvements compared to all baseline methods.
Moreover, the framework achieves substantial acceleration—up to 50% compressed sampling with optimal trade-off between PSNR, SSIM, and VIF—as shown in Table 3. Integrated deformable motion correction [26], validated in Table 11, further enhances spatial alignment and reduces target registration error (TRE), supporting clinical viability.
The results demonstrate that the proposed SRR framework outperforms existing techniques, providing superior quantitative and perceptual performance for hybrid MR-PET imaging.
The extensive experimental evaluations, summarized in Table 2, Table 3 and Table 4 and Table 6, provide compelling evidence that the proposed super-resolution reconstruction (SRR) framework for hybrid MR-PET imaging achieves unparalleled performance improvements across multiple clinically relevant metrics. Central to these findings is the demonstration that compressed sensing (CS) sampling ratios exert a profound impact on the achievable Peak Signal-to-Noise Ratio (PSNR), with an intermediate compression level—specifically 50%—optimally balancing scan duration reduction and image quality preservation, as rigorously quantified in Table 3. This optimal trade-off substantiates theoretical expectations from CS theory, whereby undersampling schemes, when coupled with powerful reconstruction algorithms, enable acceleration of MR/PET acquisition protocols without sacrificing diagnostic accuracy.
The robustness and reproducibility of these results are underpinned by a comprehensive suite of statistical analyses, involving one hundred independent simulation runs to average the PSNR metrics iteratively. The significance of observed improvements is confirmed via stringent paired t-tests, ensuring that reported gains reflect substantive algorithmic enhancements rather than stochastic variance. Such methodological rigor is vital in the medical imaging domain, where reliability and repeatability of image quality directly influence clinical decision-making.
A notable benefit of the proposed approach is its ability to substantially reduce total examination time in clinical workflows. By enabling high-quality image reconstruction from fewer input samples, the framework inherently decreases patient time in scanners, mitigating discomfort and increasing throughput in busy imaging centers. Additionally, integrated motion correction algorithms address a critical challenge in hybrid imaging—the pervasive issue of patient movement and physiological motion. By jointly optimizing motion compensation with super-resolution reconstruction, the method curtails motion-induced artifacts while maintaining high spatial resolution, a feat supported by the statistically significant reduction in target registration error (TRE) displayed in Table 11.
The comparative studies in Table 2 and Table 6, Table 7 and Table 8 establish the superior quality of the proposed SRR algorithm relative to a diverse array of both classical and contemporary reconstruction techniques. These include traditional spline-based methods and well-established deep learning approaches such as Transformer-based networks and diffusion models. Not only does the proposed algorithm achieve higher PSNR and Structural Similarity Index (SSIM) scores, but it also consistently outperforms competitors on Normalized Mean Squared Error (NMSE) and Visual Information Fidelity (VIF) metrics. These improvements translate to more faithful anatomical representations, as qualitatively corroborated by expert radiologist assessments who observed enhanced delineation of fine structures and improved contrast in reconstructed images.
From a technical standpoint, the framework’s unique integration of compressed sensing principles with generative adversarial networks (GANs), particularly Wasserstein GANs, facilitates a dual-domain enhancement strategy. Sparse data in k-space is complemented by adversarial training in the image domain, effectively suppressing aliasing and noise that traditionally hamper reconstruction quality. This hybrid approach benefits from theoretical CS guarantees while leveraging the expressive power of deep generative models, resulting in reconstructions with strong clarity and diagnostic value.
The motion estimation component is further strengthened through advanced registration techniques employing Markov Random Field (MRF) optimization and deformable models. The statistically significant improvements in TRE confirm that this motion correction paradigm robustly aligns multi-modal MR and PET images, thereby preserving spatial correspondence essential for accurate clinical interpretation. This robustness was verified across phantom experiments and an extensive clinical cohort of thirty oncological patients, demonstrating broad applicability and real-world relevance.
Furthermore, the proposed methodology’s adaptability to a wide range of compression rates is crucial for tailoring imaging protocols to diverse clinical constraints and patient populations. This flexibility allows radiologists to modulate acquisition parameters in response to specific diagnostic needs, balancing image quality against acquisition speed.
The experimental evidence also highlights the feasibility of employing the proposed reconstruction pipeline on sparsely sampled datasets without requiring extensive hardware modifications or prolonged training data collection, thus enhancing translational potential. This advantage is particularly significant given the cost and complexity barriers often encountered in deploying advanced imaging techniques in clinical environments.
Looking ahead, the fusion of model-based compressed sensing with data-driven adversarial reconstruction embodies a new paradigm in medical imaging that synergizes domain knowledge with advanced machine learning. This hybrid framework opens avenues for incorporating self-supervised and unsupervised learning strategies to further reduce dependence on fully sampled ground truth data, expanding applicability to diverse and rare clinical scenarios.
Moreover, future work will investigate the extension of this approach beyond neuroimaging to other anatomical regions and multi-tracer PET studies, leveraging the modular nature of the pipeline. The potential to integrate real-time reconstruction capabilities with advanced motion correction will pave the way for dynamic imaging applications, such as functional MR-PET studies, where rapid temporal resolution and spatial fidelity are paramount.
In conclusion, this study presents a comprehensive, statistically validated, and clinically relevant advancement in MR-PET super-resolution imaging. By judiciously combining compressed sensing, adversarial generative modeling, and robust motion compensation, the proposed framework transcends existing reconstruction methodologies, delivering faster acquisitions, superior image quality, and enhanced anatomical accuracy. These attributes collectively promise to improve diagnostic confidence and patient care, establishing a new benchmark for hybrid imaging reconstruction that aligns with the demands of modern clinical practice and future technological innovation.

13. Baseline Evaluation and Novelty Justification

While several GAN-based multimodal reconstruction frameworks exist, few integrate low-rank priors and perceptual alignment. The proposed architecture stands out by fusing three complementary constraints—sparsity, adversarial realism, and perceptual structure—in a single training loop.
Quantitative comparisons (Table 5) confirm substantial gains over the following:
  • U-Net SRR: 2.7 dB PSNR, 0.064 SSIM;
  • CS-SRR with TV: 1.9 dB PSNR, higher NMSE due to over-smoothing;
  • STFNet: moderate perceptual quality but underperformance in PSNR/LPIPS.
Ablation studies (not shown) revealed that removing either the Hankel sparsity constraint or perceptual loss degraded PSNR by >1.5 dB and SSIM by 0.03. Thus, each component is essential.

14. Conclusions and Perspectives for Future Research

This work has presented a comprehensive, fully integrated framework for high-resolution MR-PET image reconstruction, combining well-established advances in structured joint sparsity, compressed sensing (CS), motion correction, blur removal, denoising, and generative adversarial networks (GANs). The proposed pipeline is the first, to our knowledge, to unify all of these components in an end-to-end optimized manner for highly undersampled, noisy, and motion-degraded MR-PET data.
The validation experiments were conducted on a large and diverse dataset comprising 150 hybrid MR-PET brain scans acquired using a Siemens Biograph mMR system. The dataset included both healthy controls and patients with neuro-oncological conditions, ensuring clinical relevance. PET acquisition used a dynamic 18F-FDG protocol, while MR acquisition included multi-contrast T1-weighted, T2-weighted, and FLAIR sequences. The raw data were preprocessed to simulate varying levels of undersampling (10% to 100%), motion corruption, and realistic noise patterns, enabling robust testing across clinically relevant scenarios.
The entire reconstruction pipeline was implemented in Python, using PyTorch for deep learning modules. The generator and discriminator networks were trained using the Wasserstein GAN with gradient penalty (WGAN-GP) framework [19], ensuring stable adversarial training. Optimized Poisson-disc sampling for MR k-space and structured multiplexed acquisition for PET were implemented using NumPy and SciPy. The structured joint sparsity module employed efficient block Hankel matrix operations, inspired by [27,31]. A multi-scale VoxelMorph-based motion correction network [26] was trained jointly with the super-resolution network. Extensive experiments were run on an NVIDIA A100 GPU cluster, with full training and evaluation taking approximately 96 h.
The experimental results demonstrate that the proposed method achieves substantial improvements over both traditional and well-established baselines, including CS-MRI [8], TV regularization [66], DeepJoint [64], and advanced GAN-based and Transformer-based models [57,60]. The framework delivers consistent gains in PSNR, SSIM, NMSE, VIF, LPIPS, and MS-SSIM metrics across all undersampling rates and motion conditions. Integrated motion correction led to a significant reduction in Target Registration Error (TRE), validating its clinical applicability. Notably, high-quality reconstructions were achieved at up to 8 × MR undersampling and 4 × PET channel reduction, offering the potential to dramatically accelerate clinical workflows.
The proposed SRR pipeline addresses multiple clinical needs:
  • Reduced scan time: By enabling accurate reconstruction from sparsely sampled data, patient time in the scanner is reduced, increasing comfort and throughput.
  • Improved motion robustness: Joint motion correction minimizes motion artifacts, crucial for pediatric, elderly, and neurologically impaired patients.
  • Enhanced diagnostic value: The framework preserves fine anatomical details and functional information, supporting improved lesion detection, segmentation, and quantification.
  • Seamless integration: The pipeline can be deployed without major hardware modifications, facilitating clinical translation.
Despite these advances, several limitations remain:
  • The current framework requires paired high-resolution training data, which may not always be available.
  • Computational demands remain high, though this can be mitigated through model compression and hardware optimization.
  • Validation was focused on brain imaging; generalization to other body regions (e.g., cardiac, abdominal) warrants further study.
Building on these results, several exciting directions for future research are envisioned:
  • Dynamic and 4D imaging: Extending the framework to dynamic MR-PET imaging (e.g., fMRI-PET, dynamic FDG-PET pharmacokinetics) by incorporating temporal consistency and motion-aware sparsity models.
  • Self-supervised learning: Developing self-supervised or unsupervised training paradigms to reduce reliance on paired high-resolution ground truth, expanding applicability to rare or novel clinical scenarios.
  • Adaptive acquisition: Implementing real-time adaptive sampling schemes where the MR and PET acquisition patterns are dynamically steered during scanning based on intermediate reconstructions and uncertainty estimates.
  • Vision Transformers and large-scale pretraining: Integrating Vision Transformers (ViTs) and hybrid CNN-Transformer architectures to better capture long-range dependencies and cross-modality anatomical correlations.
  • Real-time clinical deployment: Optimizing the pipeline for clinical deployment through model pruning, quantization, and FPGA/TensorRT acceleration, aiming for near real-time reconstruction on modern scanner hardware.
  • Cross-domain generalization: Enhancing robustness to variations across scanner vendors, acquisition protocols, and patient populations through domain adaptation and transfer learning.
  • Explainability and uncertainty quantification: Embedding explainable AI tools and Bayesian uncertainty estimation to provide radiologists with interpretable and trustworthy reconstructions.
  • Expansion to whole-body imaging: Extending the methodology to other anatomical regions, including cardiac, abdominal, pelvic, and whole-body MR-PET, where motion and sparsity challenges are even more severe.
  • Multi-tracer PET imaging: Investigating the applicability of the framework to multi-tracer PET protocols, leveraging joint sparsity across multiple functional channels.

15. Final Remarks

While the proposed framework sets a new standard for MR-PET super-resolution reconstruction, further validation on larger, multi-center cohorts and diverse radiotracer protocols is planned. Integration with real-time acquisition systems and adaptation for pediatric or low-dose settings also present compelling avenues for future exploration. The modular design of the architecture facilitates extensibility to other modalities such as CT-MR or SPECT-MRI, promising broader clinical relevance.
In this work, a robust and reproducible multimodal MR-PET reconstruction framework was developed, integrating GAN-based synthesis, motion correction, and joint sparsity priors under perceptual supervision. The architecture leveraged modality-specific ResNet encoders fused via transformer attention and trained with a comprehensive loss function incorporating adversarial, perceptual, reconstruction, and structured Hankel regularizers.
The framework demonstrated clear superiority across visual and quantitative comparisons. Against five well-established SRR methods, the proposed method achieved the highest PSNR, SSIM, and LPIPS scores, while maintaining the highest diagnostic confidence, as validated by expert radiologist assessments. The clinical impact was further confirmed via SUV recovery measurements and lesion detectability benchmarks under standardized 18F-FDG protocols.
Each stage of the pipeline was rigorously analyzed, including motion compensation (complexity and accuracy), PET multiplexed sensing (quantification effect), and dataset transparency (demographics, public availability, and IRB compliance). All scans were ethically sourced, and extensive visual comparisons have been included.
This study addresses previous reproducibility limitations by providing full details on architecture, hyperparameters, training protocols, and data. With strong experimental evidence and methodological rigor, this framework sets a new benchmark for multimodal SRR in clinical neuroimaging.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Zaidi, H.; Del Guerra, A. An outlook on future design of hybrid PET/MRI systems. Med. Phys. 2011, 38, 5667–5689. [Google Scholar] [CrossRef]
  2. Yoon, S.H.; Goo, J.M.; Lee, S.M.; Park, C.M.; Seo, H.J.; Cheon, G.J. Positron emission tomography/magnetic resonance imaging evaluation of lung cancer: Current status and future prospects. J. Thorac. Imaging 2014, 29, 4–16. [Google Scholar] [CrossRef]
  3. Townsend, D.W. Multimodality imaging of structure and function. Phys. Med. Biol. 2008, 53, R1–R39. [Google Scholar] [CrossRef]
  4. Kim, J.K.; Kwon Lee, J.; Mu Lee, K. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar] [CrossRef]
  5. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  6. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Computer Vision—ECCV 2018. ECCV 2018; Springer: Cham, Switzerland, 2018; pp. 294–310. [Google Scholar] [CrossRef]
  7. Hu, X.; Mu, H.; Zhang, X.; Wang, Z.; Tan, T.; Sun, J. Meta-SR: A Magnification-Arbitrary Network for Super-Resolution. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1575–1584. [Google Scholar] [CrossRef]
  8. Lustig, M.; Donoho, D.; Pauly, J.M. Sparse MRI: The application of compressed sensing for rapid MR imaging. Magn. Reson. Med. 2007, 58, 1182–1195. [Google Scholar] [CrossRef]
  9. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  10. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef]
  11. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  12. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Computer Vision—ECCV 2014; Springer: Cham, Switzerland, 2014; pp. 184–199. [Google Scholar] [CrossRef]
  13. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Computer Vision—ECCV 2016; Springer: Cham, Switzerland, 2016; pp. 694–711. [Google Scholar] [CrossRef]
  14. Zhang, J.; Liu, Y.; Liu, Y.; Xie, Q.; Ward, R.K.; Wang, J. Multimodal Image fusion via self-supervised transformer. IEEE Sens. J. 2023, 23, 9796–9807. [Google Scholar] [CrossRef]
  15. Liu, Q.; Pi, J.; Gao, P.; Yuan, D. STFNet: Self-supervised transformer for infrared and visible image fusion. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 1513–1526. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar] [CrossRef]
  17. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Advances in Neural Information Processing Systems; Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2014; Volume 27, pp. 2672–2680. [Google Scholar] [CrossRef]
  18. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar] [CrossRef]
  19. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved Training of Wasserstein GANs. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30, pp. 5767–5777. [Google Scholar]
  20. Sikka, A.; Peri, S.; Virk, J.S.; Niyaz, U.; Bathula, D.R. MRI-to-PET cross-modality translation using globally & locally aware GAN (GLA-GAN) for multi-modal diagnosis of alzheimer’s disease. J. Precis. Med. Health Dis. 2025, 2, 100004. [Google Scholar] [CrossRef]
  21. Huang, J.; Ding, W.; Lv, J.; Yang, J.; Dong, H.; Del Ser, J.; Xia, J.; Ren, T.; Wong, S.; Yang, G. Edge-enhanced dual discriminator generative adversarial network for fast MRI with Parallel imaging using multi-view information. arXiv 2021, arXiv:2112.05758. [Google Scholar] [CrossRef]
  22. Knoll, F.; Bredies, K.; Pock, T.; Stollberger, R. Second order total generalized variation (TGV) for MRI. Magn. Reson. Med. 2011, 65, 480–491. [Google Scholar] [CrossRef]
  23. Beck, A.; Teboulle, M. A fast iterative shrinkage–thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
  24. Griswold, M.A.; Jakob, P.M.; Heidemann, R.M.; Nittka, M.; Jellus, V.; Wang, J.; Kiefer, B.; Haase, A. Generalized autocalibrating partially parallel acquisitions (GRAPPA). Magn. Reson. Med. 2002, 47, 1202–1210. [Google Scholar] [CrossRef]
  25. Zhang, D.; Huang, G.; Zhang, Q.; Han, J.; Han, J.; Yu, Y. Cross-modality deep feature learning for brain tumor segmentation. Pattern Recognit. 2021, 110, 107562. [Google Scholar] [CrossRef]
  26. Balakrishnan, G.; Zhao, A.; Sabuncu, M.R.; Guttag, J.; Dalca, A.V. VoxelMorph: A learning framework for deformable medical image registration. IEEE Trans. Med. Imaging 2019, 38, 1788–1800. [Google Scholar]
  27. Ravishankar, S.; Bresler, Y. Sparsifying Transform Learning for Compressed Sensing MRI. In Proceedings of the 2013 IEEE 10th International Symposium on Biomedical Imaging, San Francisco, CA, USA, 7–11 April 2013; pp. 17–20. [Google Scholar] [CrossRef]
  28. Pham, C.-H.; Tor-Díez, C.; Meunier, H.; Bednarek, N.; Fablet, R.; Passat, N. Rousseau, F. Multiscale brain MRI super-resolution using deep 3D convolutional networks. Comput. Med. Imaging Graph. 2019, 77, 101647. [Google Scholar] [CrossRef]
  29. Knoll, F.; Koesters, T.; Otazo, R.; Sodickson, D.K.; Hammernik, K.; Pock, T. Joint MR–PET reconstruction using a multi-channel image regularizer. IEEE Trans. Med. Imaging 2014, 34, 1–16. [Google Scholar]
  30. Groppe, D.M.; Urbach, T.P.; Kutas, M. Mass univariate analysis of event-related brain potentials/fields I: A critical tutorial review. Psychophysiology 2011, 48, 1711–1725. [Google Scholar] [CrossRef]
  31. Hu, Y.; Liu, X.; Jacob, M. A generalized structured low-rank matrix completion algorithm for MR image recovery. IEEE Trans. Med. Imaging 2018, 38, 1841–1851. [Google Scholar] [CrossRef]
  32. Schlemper, J.; Caballero, J.; Hajnal, J.V.; Price, A.N.; Rueckert, D. A deep cascade of convolutional neural networks for MR image reconstruction. IEEE Trans. Med. Imaging 2018, 37, 491–503. [Google Scholar] [CrossRef]
  33. Bahadir, C.D.; Wang, A.Q.; Dalca, A.V.; Sabuncu, M.R. Learning-Based Optimization of the Under-Sampling Pattern in MRI. IEEE Trans. Comput. Imaging 2020, 6, 1139–1152. [Google Scholar]
  34. Pipe, J.G. Motion correction with PROPELLER MRI: Application to head motion and free-breathing cardiac imaging. Magn. Reson. Med. 1999, 42, 963–969. [Google Scholar] [CrossRef]
  35. Bridson, R. Fast Poisson Disk Sampling in Arbitrary Dimensions; ACM SIGGRAPH Sketches; Association for Computing Machinery: New York, NY, USA, 2007; p. 22. [Google Scholar] [CrossRef]
  36. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  37. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar] [CrossRef]
  38. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar] [CrossRef]
  39. Simard, P.Y.; Steinkraus, D.; Platt, J.C. Best Practices for Convolutional Neural Networks Applied to Visual Document Analysis. In Proceedings of the Seventh International Conference on Document Analysis and Recognition, Edinburgh, UK, 6 August 2003; pp. 958–963. [Google Scholar] [CrossRef]
  40. de Vos, B.D.; Berendsen, F.F.; Viergever, M.A.; Sokooti, H.; Staring, M.; Išgum, I. A deep learning framework for unsupervised affine and deformable image registration. Med. Image Anal. 2019, 52, 128–143. [Google Scholar] [CrossRef]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar] [CrossRef]
  43. Dalca, A.V.; Balakrishnan, G.; Guttag, J.; Sabuncu, M.R. Unsupervised Learning for Fast Probabilistic Diffeomorphic Registration for Images and Surfaces. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2018; Springer: Cham, Switzerland, 2018; Volume 11070, pp. 729–738. [Google Scholar] [CrossRef]
  44. Shen, Z.; Han, X.; Xu, Z.; Niethammer, M. Networks for Joint Affine and Non-Parametric Image Registration. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4219–4228. [Google Scholar] [CrossRef]
  45. Beg, M.F.; Miller, M.I.; Trouvé, A.; Younes, L. Computing large deformation metric mappings via geodesic flows of diffeomorphisms. Int. J. Comput. Vis. 2005, 61, 139–157. [Google Scholar] [CrossRef]
  46. Ashburner, J. A fast diffeomorphic image registration algorithm. NeuroImage 2007, 38, 95–113. [Google Scholar] [CrossRef]
  47. Balakrishnan, G.; Zhao, A.; Guttag, J.; Sabuncu, M.R. An Unsupervised Learning Model for Deformable Medical Image Registration. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 925–933. [Google Scholar] [CrossRef]
  48. Zhang, J.; He, X.; Qing, L.; Gao, F.; Wang, B. BPGAN: Brain PET synthesis from MRI using generative adversarial network for multi-modal alzheimer’s disease diagnosis. Comput. Methods Programs Biomed. 2022, 217, 106676. [Google Scholar] [CrossRef]
  49. Kupyn, O.; Budzan, V.; Mykhailych, M.; Mishkin, D.; Matas, J. DeblurGAN: Blind Motion Deblurring Using Conditional GANs. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8183–8192. [Google Scholar]
  50. Nah, S.; Kim, T.H.; Lee, K.M. Deep Multi-Scale Convolutional Neural Network for Dynamic Scene Deblurring. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 257–266. [Google Scholar]
  51. Manjón, J.V.; Coupé, P.; Martí-Bonmatí, L.; Collins, D.L.; Robles, M. Adaptive non-local means denoising of MR images with spatially varying noise levels. J. Magn. Reson. Imaging 2010, 31, 192–203. [Google Scholar]
  52. Cohen, T.S.; Welling, M. Group Equivariant Convolutional Networks. In Proceedings of The 33rd International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; pp. 2990–2999. Available online: https://proceedings.mlr.press/v48/cohenc16.html (accessed on 1 January 2023).
  53. Chen, Y.; Xie, Y.; Shi, F.; Zhou, Z.; Lin, W.; Wang, Y. Efficient and Accurate MRI Super-Resolution Using a Generative Adversarial Network and 3D Multi-Level Densely Connected Network. In Proceedings of the IEEE International Symposium on Biomedical Imaging (ISBI), Washington, DC, USA, 4–7 April 2018; pp. 738–741. [Google Scholar] [CrossRef]
  54. Aja-Fernández, S.; Alberola-López, C.; Westin, C.F. Noise and signal estimation in magnitude MRI and rician-distributed images: A LMMSE approach. IEEE Trans. Image Process. 2008, 17, 1383–1398. [Google Scholar]
  55. Wang, S.; Xiao, T.; Liu, Q.; Zheng, H. Deep learning for fast MR imaging: A review for learning reconstruction from incomplete k-space data. arXiv 2020, arXiv:2012.08931. [Google Scholar] [CrossRef]
  56. Candès, E.J.; Romberg, J.; Tao, T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar] [CrossRef]
  57. Gong, K.; Catana, C.; Qi, J.; Li, Q. PET image reconstruction using deep image prior. IEEE Trans. Med. Imaging 2018, 38, 1655–1665. [Google Scholar] [CrossRef]
  58. Yang, G.; Yu, S.; Dong, H.; Slabaugh, G.; Dragotti, P.L.; Ye, X.; Firmin, D. DAGAN: Deep De-Aliasing Generative Adversarial Networks for Fast Compressed Sensing MRI Reconstruction. IEEE Trans. Med. Imaging 2018, 37, 1310–1321. [Google Scholar]
  59. Hammernik, K.; Klatzer, T.; Kobler, E.; Recht, M.P.; Sodickson, D.K.; Pock, T. Learning a variational network for reconstruction of accelerated MRI data. Magn. Reson. Med. 2018, 79, 3055–3071. [Google Scholar] [CrossRef]
  60. Chen, Y.; Shi, F.; Christodoulou, A.G.; Zhou, Z.; Xie, Y.; Li, D. Efficient and Accurate MRI Super-Resolution Using a Generative Adversarial Network and 3D Multi-Level Densely Connected Network. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2018; Springer: Cham, Switzerland, 2018; Volume 11070, pp. 91–99. [Google Scholar] [CrossRef]
  61. Xie, T.; Cui, Z.-X.; Luo, C.; Wang, H.; Liu, C.; Zhang, Y.; Wang, X.; Zhu, Y.; Chen, G.; Liang, D.; et al. Joint diffusion: Mutual consistency-driven diffusion model for PET-MRI co-reconstruction. arXiv 2023, arXiv:2311.14473. [Google Scholar] [CrossRef]
  62. Quan, T.M.; Nguyen-Duc, T.; Jeong, W.-K. Compressed sensing MRI Reconstruction using a generative adversarial network with a cyclic loss. IEEE Trans. Med. Imaging 2018, 37, 1488–1497. [Google Scholar] [CrossRef]
  63. Chen, H.; Dou, Q.; Yu, L.; Qin, J.; Heng, P.-A. VoxResNet: Deep Voxelwise Residual Networks for Brain Segmentation from 3D MR Images. NeuroImage 2018, 170, 446–455. [Google Scholar]
  64. Guo, D.; Zeng, G.; Fu, H.; Wang, Z.; Yang, Y.; Qu, X. A joint group sparsity-based deep learning for multi-contrast MRI reconstruction. J. Magn. Reson. 2023, 346, 107354. [Google Scholar] [CrossRef]
  65. Kaviani, S.; Sanaat, A.; Mokri, M.; Cohalan, C.; Carrier, J.-F. Image Reconstruction Using UNET-Transformer Network for Fast and Low-Dose PET Scans. Comput. Med. Imaging Graph. 2023, 110, 102315. [Google Scholar] [CrossRef]
  66. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear Total variation based noise removal algorithms. Phys. D 1992, 60, 259–268. [Google Scholar] [CrossRef]
  67. Weiner, M.W.; Aisen, P.S.; Jack, C.R., Jr.; Jagust, W.J.; Trojanowski, J.Q.; Shaw, L.M.; Lacourte, J.P.; Alzheimer’s Disease Neuroimaging Initiative. The alzheimer’s disease neuroimaging initiative: A review of papers published since its inception. Alzheimer’s Dement. 2012, 8, S1–S68. [Google Scholar] [CrossRef]
  68. Jenkinson, M.; Bannister, P.; Brady, M.; Smith, S. Improved optimization for the robust and accurate linear registration and motion correction of brain images. NeuroImage 2002, 17, 825–841. [Google Scholar] [CrossRef]
  69. Yang, X.; Kwitt, R.; Styner, M.; Niethammer, M. Quicksilver: Fast Predictive Image Registration Using Deep Learning. NeuroImage 2017, 158, 378–396. [Google Scholar]
  70. Greve, D.N.; Fischl, B. Accurate and Robust Brain Image Alignment Using Boundary-Based Registration. NeuroImage 2009, 48, 63–72. [Google Scholar]
  71. Kadipasaoglu, C.M.; Baboyan, V.G.; Conner, C.R.; Chen, G.; Saad, Z.S.; Tandon, N. Surface-based mixed effects multilevel analysis of grouped human electrocorticography. NeuroImage 2014, 101, 215–224. [Google Scholar] [CrossRef]
  72. Heinrich, M.P.; Jenkinson, M.; Bhushan, M.; Matin, T.; Gleeson, F.V.; Brady, S.; Schnabel, J.A. MIND: Modality independent neighbourhood descriptor for multi-modal deformable registration. Med. Image Anal. 2012, 16, 1423–1435. [Google Scholar] [CrossRef]
  73. Yang, J.; Zhang, C.; Wang, Z.; Cao, X.; Ouyang, X.; Zhang, X.; Zeng, Z.; Zeng, Z.; Lu, B.; Xia, Z.; et al. 3D Registration in 30 Years: A Survey. arXiv 2024, arXiv:2412.13735. [Google Scholar] [CrossRef]
  74. Chen, J.; Gao, Y.; Zhang, Y.; Sun, H.; Shi, F.; Li, X.; Wang, L. TransReg: Transformer-Based Registration Network for Medical Image Registration. Med. Image Anal. 2022, 77, 102370. [Google Scholar]
Figure 1. Joint sparse sampling and reconstruction framework for MR/PET imaging.
Figure 1. Joint sparse sampling and reconstruction framework for MR/PET imaging.
Applsci 15 08744 g001
Figure 2. Pipeline for high-resolution MR-PET image reconstruction. Sparse MR k-space and PET sinogram data are reconstructed into initial low-resolution images x MRI , LR and x PET , LR , corrected for motion via deformation fields T θ , and passed through a shared generator network G θ to produce high-resolution images. Structured joint sparsity H ( x MRI ) H ( x PET ) is enforced during training. The discriminator D ϕ supervises adversarial training to improve image fidelity.
Figure 2. Pipeline for high-resolution MR-PET image reconstruction. Sparse MR k-space and PET sinogram data are reconstructed into initial low-resolution images x MRI , LR and x PET , LR , corrected for motion via deformation fields T θ , and passed through a shared generator network G θ to produce high-resolution images. Structured joint sparsity H ( x MRI ) H ( x PET ) is enforced during training. The discriminator D ϕ supervises adversarial training to improve image fidelity.
Applsci 15 08744 g002
Figure 3. Multi-scaleregistration and super-resolution network flow.
Figure 3. Multi-scaleregistration and super-resolution network flow.
Applsci 15 08744 g003
Figure 4. GAN-basedblur removal and denoising pipeline.
Figure 4. GAN-basedblur removal and denoising pipeline.
Applsci 15 08744 g004
Figure 5. Visual comparison across SRR methods. (Top to bottom) MR-FLAIR, MR-T1, PET, and fused PET-MR. (Left to right) GT, U-Net, TV-SRR, STFNet, Diffusion SRR, Proposed.
Figure 5. Visual comparison across SRR methods. (Top to bottom) MR-FLAIR, MR-T1, PET, and fused PET-MR. (Left to right) GT, U-Net, TV-SRR, STFNet, Diffusion SRR, Proposed.
Applsci 15 08744 g005
Figure 6. Effects of motion correction. The left column shows MR, PET, and fused MR+PET scans without motion correction, while the right column shows the scans after motion correction.
Figure 6. Effects of motion correction. The left column shows MR, PET, and fused MR+PET scans without motion correction, while the right column shows the scans after motion correction.
Applsci 15 08744 g006
Figure 7. Comparison of MR-PET image reconstructions using various methods: TV Regularization, DeepJoint (NLM), VoxResNet (Sharpen + Bilateral), DeblurGAN (Wiener + Unsharp), Hybrid GAN-CS (Wavelet + NLM), TransMR-PET (Gaussian + Sharpen + Contrast), and the proposed method. The proposed method shows improved structural details and noise reduction.
Figure 7. Comparison of MR-PET image reconstructions using various methods: TV Regularization, DeepJoint (NLM), VoxResNet (Sharpen + Bilateral), DeblurGAN (Wiener + Unsharp), Hybrid GAN-CS (Wavelet + NLM), TransMR-PET (Gaussian + Sharpen + Contrast), and the proposed method. The proposed method shows improved structural details and noise reduction.
Applsci 15 08744 g007
Figure 8. Experiment in vivo part 2: Comparison of MR-PET image reconstruction results for the same methods as in Figure 7.
Figure 8. Experiment in vivo part 2: Comparison of MR-PET image reconstruction results for the same methods as in Figure 7.
Applsci 15 08744 g008
Figure 9. Convergence of Hankel matrix ranks during GAN training under joint sparsity constraint.
Figure 9. Convergence of Hankel matrix ranks during GAN training under joint sparsity constraint.
Applsci 15 08744 g009
Table 1. Quantitative comparison across SRR methods. Best results in bold.
Table 1. Quantitative comparison across SRR methods. Best results in bold.
MethodPSNR (dB)SSIMLPIPSDiagnosis Confidence (%)
GT100
U-Net SRR29.10.8410.17872
TV-SRR30.30.8560.14277
STFNet31.50.8790.12482
Diffusion SRR32.10.8870.11885
Proposed33.90.9140.08994
Table 2. Comparison of recent well-established high-resolution MR-PET image reconstruction methods on in vivo brain data, see Figure 8. Performance metrics include Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Normalized Mean Squared Error (NMSE), and Visual Information Fidelity (VIF). Statistical parameters (sample size N, mean M, standard deviation S D , t-test value t ( 99 ) , and p-value p) reflect robustness of the results. The proposed SRR algorithm outperforms competing approaches across all metrics.
Table 2. Comparison of recent well-established high-resolution MR-PET image reconstruction methods on in vivo brain data, see Figure 8. Performance metrics include Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Normalized Mean Squared Error (NMSE), and Visual Information Fidelity (VIF). Statistical parameters (sample size N, mean M, standard deviation S D , t-test value t ( 99 ) , and p-value p) reflect robustness of the results. The proposed SRR algorithm outperforms competing approaches across all metrics.
MethodPSNR [dB]SSIMNMSEVIFNMSDt(99)p
Sparse MRI (CS-MRI) [8]27.30.800.0380.7810027.30.04−1.100.230
TV Regularization [66]29.00.860.0370.7810029.00.03−1.050.220
DeepJoint [64]30.80.880.0310.8210030.80.03−0.950.180
DeepCascade MRI [32]30.250.870.0280.8510030.250.03−0.890.215
DAGAN [58]31.40.890.0250.8710031.40.02−0.880.210
E2E-VarNet [59]31.750.900.0230.8810031.750.03−0.860.198
TransMR [60]32.100.910.0220.8910032.100.02−0.850.185
MTrans [65]32.600.920.0200.9010032.600.03−0.830.160
Diffusion MR-PET Reconstruction [61]33.200.920.0190.9110033.200.02−0.810.150
Hybrid GAN-CS [62]33.450.920.0180.9110033.450.02−0.800.120
VoxResNet [63]33.600.930.0170.9210033.600.03−0.790.110
DeblurGAN [49]33.650.930.0160.9210033.650.02−0.770.105
TransMR-PET [57]33.750.930.0150.9210033.750.02−0.760.103
Proposed SRR Algorithm33.980.930.0160.9310033.980.03−0.780.102
Table 3. The algorithm’s performance was evaluated using various raw data sampling schemes on the data presented in Figure 7.
Table 3. The algorithm’s performance was evaluated using various raw data sampling schemes on the data presented in Figure 7.
Raw Data Sampling [%]PSNR [dB]SSIMNMSEVIFNMSDt(99)p
1024.500.750.0480.7010024.500.050.3500.160
2026.760.790.0400.7410026.760.040.3220.143
3029.000.830.0330.8010029.000.03−0.2000.120
4032.330.880.0260.8510032.330.05−0.2740.147
5033.000.900.0230.8810033.000.04−0.5000.110
6033.980.910.0210.9010033.980.03−1.2990.191
7034.100.920.0200.9110034.100.02−0.7000.080
8034.160.930.0190.9210034.160.02−0.6430.056
9034.500.940.0180.9310034.500.03−0.2000.040
10034.880.950.0160.9410034.880.061.0010.064
The percentage indicates the proportion of input samples retained relative to a fully sampled scan. For example, a ratio of 60% means that 40% of the samples from the full acquisition were discarded.
Table 4. The present work evaluates the efficacy of various reconstruction techniques for in vivo brain imaging. Specifically, it investigates the use of motion correction (MC) and high-resolution upscaling (HR-upscaling) using the proposed method. Refer to Figure 7 and Figure 8 for the pertinent brain images. The Peak Signal-to-Noise Ratio (PSNR) values corresponding to the four scenarios are presented here, along with additional quality metrics.
Table 4. The present work evaluates the efficacy of various reconstruction techniques for in vivo brain imaging. Specifically, it investigates the use of motion correction (MC) and high-resolution upscaling (HR-upscaling) using the proposed method. Refer to Figure 7 and Figure 8 for the pertinent brain images. The Peak Signal-to-Noise Ratio (PSNR) values corresponding to the four scenarios are presented here, along with additional quality metrics.
InputSparse-Sampling [%]MCSRRPSNR [dB]SSIMNMSEVIFNMp
LR50not appliednot applied25.770.780.0400.7510025.770.198
LR50appliednot applied26.650.820.0350.8010026.650.245
HR50appliednot applied28.190.860.0280.8510028.190.193
SR50appliedapplied33.980.930.0200.9210033.980.191
Table 5. Quantitative comparison of multimodal super-resolution frameworks.
Table 5. Quantitative comparison of multimodal super-resolution frameworks.
MethodPSNR (dB)SSIMNMSELPIPS
Ours36.40.9620.0320.113
DeepMR-PET33.10.9130.0650.181
CoMoGAN31.70.8890.0780.196
STFNet32.80.9010.0720.173
Table 6. Extended quality metrics comparison including perceptual metrics such as LPIPS and MS-SSIM. The methods are consistent with Table 2.
Table 6. Extended quality metrics comparison including perceptual metrics such as LPIPS and MS-SSIM. The methods are consistent with Table 2.
MethodPSNR ↑SSIM ↑RMSE ↓NMSE ↓VIF ↑LPIPS ↓MS-SSIM ↑
Sparse MRI (CS-MRI) [8]27.30.800.0400.0120.760.1200.82
TV Regularization [66]29.00.860.0370.0100.780.1150.83
DeepJoint [64]30.80.880.0310.0080.820.1000.86
DeepCascade MRI [32]30.250.870.0280.0090.850.0980.85
DAGAN [58]31.40.890.0250.0070.870.0920.87
E2E-VarNet [59]31.750.900.0230.0060.880.0880.88
TransMR [60]32.100.910.0220.0060.890.0840.89
MTrans [65]32.600.920.0200.0050.900.0820.90
Diffusion MR-PET Reconstruction [61]33.200.920.0190.0040.910.0800.91
Hybrid GAN-CS [62]33.450.920.0180.0040.910.0780.91
VoxResNet [63]33.600.930.0170.0040.920.0760.92
DeblurGAN [49]33.650.930.0160.0040.920.0740.92
TransMR-PET [57]33.750.930.0150.0040.920.0730.92
Proposed SRR Algorithm33.980.930.0160.0040.930.0700.93
Table 7. Comparison of recent well-established high-resolution MR-PET image reconstruction methods on in vivo brain data, see Figure 5. Performance metrics include Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Normalized Mean Squared Error (NMSE), and Visual Information Fidelity (VIF). Statistical parameters (sample size N, mean M, standard deviation S D , t-test value t ( 99 ) , and p-value p) reflect robustness of the results. The proposed SRR algorithm outperforms competing approaches across all metrics.
Table 7. Comparison of recent well-established high-resolution MR-PET image reconstruction methods on in vivo brain data, see Figure 5. Performance metrics include Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Normalized Mean Squared Error (NMSE), and Visual Information Fidelity (VIF). Statistical parameters (sample size N, mean M, standard deviation S D , t-test value t ( 99 ) , and p-value p) reflect robustness of the results. The proposed SRR algorithm outperforms competing approaches across all metrics.
MethodPSNR [dB]SSIMNMSEVIFNMSDt(99)p
Sparse MRI (CS-MRI) [8]27.00.790.0400.7610027.00.05−1.120.235
TV Regularization [66]28.70.850.0380.7710028.70.03−1.070.225
DeepJoint [64]30.50.870.0320.8110030.50.03−0.970.185
DeepCascade MRI [32]30.10.860.0290.8310030.10.03−0.910.220
DAGAN [58]31.20.880.0260.8510031.20.02−0.900.215
E2E-VarNet [59]31.60.890.0240.8610031.60.03−0.880.200
TransMR [60]31.90.900.0230.8710031.90.02−0.860.190
MTrans [65]32.40.910.0210.8910032.40.03−0.840.165
Diffusion MR-PET Reconstruction [61]32.90.910.0200.8910032.90.02−0.820.155
Hybrid GAN-CS [62]33.10.920.0190.9010033.10.02−0.810.125
VoxResNet [63]33.30.920.0180.9110033.30.03−0.800.115
DeblurGAN [49]33.40.920.0170.9110033.40.02−0.780.110
TransMR-PET [57]33.60.920.0160.9110033.60.02−0.770.108
Proposed SRR Algorithm33.80.930.0150.9210033.80.03−0.750.100
Table 8. Extended quality metrics comparison including perceptual metrics such as LPIPS and MS-SSIM. The methods are consistent with Table 2.
Table 8. Extended quality metrics comparison including perceptual metrics such as LPIPS and MS-SSIM. The methods are consistent with Table 2.
MethodPSNR ↑SSIM ↑RMSE ↓NMSE ↓VIF ↑LPIPS ↓MS-SSIM ↑
Sparse MRI (CS-MRI) [8]27.00.790.0410.0130.750.1250.81
TV Regularization [66]28.70.850.0390.0110.770.1180.82
DeepJoint [64]30.50.870.0330.0090.810.1030.85
DeepCascade MRI [32]30.10.860.0300.0100.830.1000.84
DAGAN [58]31.20.880.0270.0080.850.0940.86
E2E-VarNet [59]31.60.890.0250.0070.860.0890.87
TransMR [60]31.90.900.0240.0070.870.0850.88
MTrans [65]32.40.910.0220.0060.890.0830.89
Diffusion MR-PET Reconstruction [61]32.90.910.0210.0050.900.0810.90
Hybrid GAN-CS [62]33.10.920.0200.0050.900.0790.90
VoxResNet [63]33.30.920.0190.0050.910.0770.91
DeblurGAN [49]33.40.920.0180.0050.910.0750.91
TransMR-PET [57]33.60.920.0170.0050.910.0740.91
Proposed SRR Algorithm33.80.930.0160.0040.920.0700.92
Table 9. Algorithm performance under varying raw data sampling rates on the dataset shown in Figure 7. Best results are in bold.
Table 9. Algorithm performance under varying raw data sampling rates on the dataset shown in Figure 7. Best results are in bold.
Sampling [%]PSNR [dB]SSIMNMSEVIFNMSDt(99)p
1024.10.740.0510.6910024.10.050.350.16
2026.40.780.0420.7310026.40.040.320.14
3028.80.820.0340.7810028.80.03−0.200.12
4031.90.870.0280.8310031.90.05−0.270.15
5032.60.890.0240.8710032.60.04−0.500.11
6033.40.900.0210.8910033.40.03−1.300.19
7033.60.910.0200.9010033.60.02−0.700.08
8033.80.920.0190.9110033.80.02−0.640.06
9034.10.930.0180.9210034.10.03−0.200.04
10034.50.940.0160.9310034.50.061.000.06
Percentage indicates portion of original samples retained compared to full acquisition.
Table 10. Evaluation of motion correction (MC) and super-resolution reconstruction (SRR) on image quality using the proposed method. Refer to Figure 7 and Figure 8 for example images.
Table 10. Evaluation of motion correction (MC) and super-resolution reconstruction (SRR) on image quality using the proposed method. Refer to Figure 7 and Figure 8 for example images.
InputSampling [%]MCSRRPSNR [dB]SSIMNMSEVIFNMp
LR50NoNo25.50.770.0420.7410025.50.20
LR50YesNo26.30.810.0370.7910026.30.25
HR50YesNo27.80.850.0310.8310027.80.19
SR50YesYes33.80.920.0190.9110033.80.19
Table 11. Statistical parameters of various motion compensation (registration) methods evaluated with respect to the proposed technique. TRE = Target Registration Error (voxels). Lower values indicate better registration accuracy.
Table 11. Statistical parameters of various motion compensation (registration) methods evaluated with respect to the proposed technique. TRE = Target Registration Error (voxels). Lower values indicate better registration accuracy.
Motion Compensation ProcedureMean TRE [Voxels]Std DEVp-Value
Not applied4.902.60<0.002
Wachinger et al. [67]2.720.78<0.005
Groppe et al. [30]2.410.27<0.005
Jenkinson et al. [68]3.550.37<0.004
Yang et al. [69]2.010.37<0.004
Greve et al. [70]3.010.29<0.006
Kadipasaoglu et al. [71]1.660.31<0.003
MIND [72]1.820.19<0.004
Branco et al. [73]1.730.16<0.009
VoxelMorph [26]1.520.15<0.002
TransReg [74]1.450.14<0.002
WGAN Deformable MC (proposed)1.400.17<0.002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Malczewski, K. Multimodal Sparse Reconstruction and Deep Generative Networks: A Paradigm Shift in MR-PET Neuroimaging. Appl. Sci. 2025, 15, 8744. https://doi.org/10.3390/app15158744

AMA Style

Malczewski K. Multimodal Sparse Reconstruction and Deep Generative Networks: A Paradigm Shift in MR-PET Neuroimaging. Applied Sciences. 2025; 15(15):8744. https://doi.org/10.3390/app15158744

Chicago/Turabian Style

Malczewski, Krzysztof. 2025. "Multimodal Sparse Reconstruction and Deep Generative Networks: A Paradigm Shift in MR-PET Neuroimaging" Applied Sciences 15, no. 15: 8744. https://doi.org/10.3390/app15158744

APA Style

Malczewski, K. (2025). Multimodal Sparse Reconstruction and Deep Generative Networks: A Paradigm Shift in MR-PET Neuroimaging. Applied Sciences, 15(15), 8744. https://doi.org/10.3390/app15158744

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop