1. Introduction
Multimodal medical imaging is the integration of two or more imaging modalities during a single examination, a process with both diagnostic and therapeutic implications. It provides comprehensive insights into the anatomical and functional characteristics of human tissues and has revolutionised diagnostic and treatment procedures [
1,
2,
3]. The combination of positron emission tomography (PET) and computed tomography (CT)—known as PET/CT, a technique combining PET’s functional imaging capabilities with CT’s high-resolution anatomical details—has emerged as the gold standard in clinical and research contexts [
4,
5]. This synergistic combination permits precise localisation of metabolic disorders, boosting diagnostic accuracy in oncological, neurological, and cardiovascular applications [
6]. The PET/CT imaging technique has transformed nuclear medicine through enhancing diagnostic accuracy, presenting clinical applicability, and offering personalised treatment strategies [
7,
8]. However, the correct fusion of PET and CT datasets necessitates reliable and precise image registration techniques, which align the images spatially to ensure coherence across modalities [
9,
10].
Medical image registration, defined as the process of aligning datasets such that they form a single coordinate system for multimodality imaging, is crucial. This approach is used to determine an optimal spatial transformation that aligns the underlying anatomical structures. Registration provides a deformation field that enables morphometric analysis, such as Jacobian-based volume-change mapping, in addition to cross-modal comparison [
11]. It is utilised for a variety of clinical applications, including image guidance, motion tracking, segmentation, dose accumulation, and picture reconstruction [
12]. Despite major breakthroughs, the intrinsic discrepancies between PET and CT—such as variances in spatial resolution, intensity profiles, and imaging artefacts—present substantial challenges for accurate registration. Traditional techniques, which frequently rely on rigid transformations, have proven inadequate for handling complicated deformations and mismatches in multimodal imaging. As a result, there has been increased emphasis on designing strong non-rigid and affine registration methods that can account for these complexities [
11,
13].
The Demons algorithm (also known as non-rigid registration) has received a lot of attention for its versatility and computing efficiency. It works on the basis of optical flow, using a fluid-like transformation model to align images. Despite its capabilities, the algorithm’s performance is influenced by several variables, including sigma fluid and interpolation methods [
14,
15,
16]. Parameter tweaking is critical for improving accuracy, especially in cases involving substantial anatomical variability. Furthermore, the histogram bin sizes chosen in a given application have a significant impact on the algorithm’s capacity to capture intensity correlations between modalities, emphasising the importance of regular evaluations of these aspects [
17]. Another popular image registration approach is free-form deformation (FFD), a technique used for efficient, smooth, and precise geometrical parametrisation; examples include the approaches implemented in the Medical Image Registration Toolbox (MIRT). In this technique, B-spline transformations are used to permit highly flexible deformations, making it ideal for correcting complex mismatches in multimodal datasets [
18,
19]. The MIRT has exhibited strong performance in a variety of therapeutic applications, particularly those requiring high precision. However, because of its processing requirements and sensitivity to parameter configurations such as sigma fluid values and histogram bins, careful optimisation is required to enhance its effectiveness [
20,
21]. Affine transformation is a geometric transformation, also known as an affinity, that is distinguished by its ability to preserve lines and parallelism but not necessarily Euclidean distances and angles. It is also commonly employed in medical imaging to correct for geometric imperfections. An example of this method is the MATLAB Intensity-Based Registration tool, which provides an efficient platform for performing affine transformations with user-defined parameters [
22]. While affine approaches are less flexible than non-rigid techniques, their simplicity and computing efficiency make them desirable for some applications. Enhancements such as contrast stretching with functions like imadjust have been demonstrated to greatly enhance performance, especially with respect to datasets with difficult intensity distributions [
23].
The incorporation of preprocessing techniques into registration processes has been recognised as a significant component of improving accuracy and dependability. Adapthisteq and imadjust have been found to efficiently eliminate intensity disparities and improve feature recognition [
24]. These preprocessing processes are especially useful in multimodal imaging, where discrepancies in intensity profiles can conceal crucial features and prevent accurate registration [
11,
25]. Despite advances in registration algorithms and preparation techniques, there is still a paucity of complete comparative evaluations. Individual studies have shown the usefulness of several techniques in specific circumstances, but little is known about their relative performance across various parameter configurations and clinical scenarios [
2,
26,
27,
28]. This knowledge gap highlights the necessity of systematic evaluations that can help guide the selection and optimisation of registration procedures depending on application requirements.
In this study, we sought to address these gaps by conducting a comprehensive direct head-to-head comparison of three well-known registration techniques—Demons Image Registration with Modality Transformation, MIRT-based Free-Form Deformation, and MATLAB Intensity-Based Registration—in the context of PET/CT imaging. We aimed to identify the best configurations for every technique by carefully adjusting critical factors, including sigma fluid, histogram bins, and interpolation algorithms. Moreover, we investigated the impact of preprocessing processes, such as contrast enhancement and histogram equalisation, on registration performance. The findings are intended to provide useful insights into the relative strengths and limitations of various techniques, laying the groundwork for their use in both clinical and research settings.
4. Discussion
Multimodal image registration remains a cornerstone of medical imaging, with PET/CT integration being particularly important for clinical diagnoses and research. This study evaluates three popular registration techniques—Demons Image Registration, MIRT-Based Free-Form Deformation, and MATLAB Intensity-Based Registration—across a range of parameter settings, preprocessing approaches, and assessment criteria.
The Demons algorithm demonstrated a solid balance between computational efficiency and registration accuracy, consistent with previous findings that highlighted its suitability for clinical applications requiring both speed and reliability.
Table 1 shows that this approach performed optimally for a sigma fluid value = 6, with an RMSE of 0.1529 and an MAE of 0.0543. Histogram bin selection was also important, with lower bin sizes, such as 100, producing higher accuracy metrics than larger bins. This shows Demons registration’s sensitivity to intensity granularity. However, this approach encountered difficulties when dealing with complex non-rigid deformations, which could be an important factor for clinical situations featuring considerable anatomical variability. Recent studies confirm this observation, noting that Demons registration performs well in simpler anatomical regions but struggles with high non-rigid deformation, such as respiratory or cardiac motion [
16].
The MIRT’s ability to capture complex anatomical deformations was demonstrated by its constant performance across multiple parameter settings. The MIRT demonstrated resilience, with an RMSE of 0.1725 and a PCC of 4.0008 × 10
4. However, its computational needs were significantly higher, with processing times almost 1.5 times longer than those of the Demons algorithm. This trade-off between precision and efficiency makes the MIRT an excellent choice for research-oriented applications that require high accuracy, but less so for time-critical clinical processes. These findings are consistent with prior work reporting that free-form deformation approaches, while computationally intensive, facilitate superior alignment for organs subject to complex deformation, such as the liver and thoracic structures [
27,
28].
The MATLAB-based technique stood out because of its simplicity and adaptability, aligning with earlier studies that emphasised the robustness of affine registration strategies in hybrid workflows and their utility in diverse clinical datasets [
23]. The use of preprocessing techniques, particularly contrast augmentation with imadjust, led to much better registration results. At an alpha = 6, this approach had an RMSE of 0.1317 and an MAE of 0.0402, demonstrating its usefulness for datasets with complex intensity profiles. Affine registration strategies, although less flexible, continue to be applied in hybrid workflows due to their robustness and simplicity, particularly when combined with intensity-based preprocessing [
18,
23]. A comparison of linear and cubic interpolation methods found that the former regularly outperformed the latter in terms of RMSE and MAE, as shown in
Table 6.
Preprocessing approaches have emerged as a key factor in registration performance. Histogram equalisation and contrast enhancement effectively mitigated intensity disparities, especially in low-contrast PET/CT datasets, lowering the RMSE by as much as 16%, as evidenced by comparisons using imadjust and adapthisteq, demonstrating the importance of integrated preprocessing pipelines in multimodal registration workflows.
Clinically, the Demons algorithm demonstrated a strong balance between computational efficiency and registration accuracy, making it well suited for time-sensitive applications such as adaptive radiotherapy, where fast yet reliable performance is essential. In contrast, the MIRT-based free-form deformation approach, although more computationally intensive, has been shown to allow superior alignment in anatomically complex regions such as the thorax and abdomen, supporting its utility in high-precision applications like neuroimaging research [
27]. The MATLAB-based affine registration technique stood out in this study for its simplicity and adaptability, particularly when combined with preprocessing methods, a finding consistent with previous reports highlighting the robustness of affine strategies in multimodal registration tasks [
23]. By comparison, deep-learning-based methods such as VoxelMorph and Quicksilver offer excellent accuracy and speed but require large, annotated datasets and extensive computational resources and often lack transparency, which limits their current clinical translation [
33,
34]. Beyond these approaches, recent work has extended affine and non-rigid registration methods using deep learning. Chen et al. (2025) provided a comprehensive survey of new technologies, uncertainty modelling, and evaluation metrics in deep learning registration, underscoring both progress and ongoing challenges [
35]. Similarly, Trotter et al. (2023) and Hussain et al. (2024) emphasised deep learning’s increasing role in PET/CT imaging for segmentation, enhancement, and registration-related applications while also noting persistent barriers such as data scarcity, computational demands, and limited interpretability [
1,
5]. Taken together, these findings highlight that while deep-learning approaches continue to advance rapidly, classical registration algorithms remain highly relevant in clinical workflows because of their interpretability, reproducibility, and practical feasibility.
Several commercial PET/CT registration and visualisation platforms are available, including Mirada XD (Mirada Medical,
https://mirada-medical.com), Hermia (Hermes Medical Solutions,
https://www.hermesmedical.com), AIDAN (Siemens Healthineers,
https://www.siemens-healthineers.com), and Medicalholodeck (
https://www.medicalholodeck.com). Open-source options such as 3D Slicer (
https://www.slicer.org) and AMIDE (
http://amide.sourceforge.net) are also widely used in clinical and research settings. Unlike these established tools, our approach focuses on transparent and reproducible MATLAB-based implementations of classical methods (Demons, MIRT, and affine), allowing detailed parameter tuning and benchmarking, thereby providing methodological insights that complement the functionality of commercial platforms, which often operate as closed systems.
This study has several limitations. The dataset was relatively small (100 slices from 15 oncologic patients), which may limit generalisability. Although 3D volumetric registration is standard in clinical practice, only 2D slices were analysed in this study; this choice was made to reduce computational demand and allow controlled parameter comparisons across algorithms. The dataset was also relatively homogeneous, consisting primarily of thoracic PET/CT slices, improving internal consistency but restricting applicability to other anatomical regions and more diverse patient populations. MATLAB 2012a was used to ensure compatibility with the employed toolboxes, although newer platforms may offer enhanced functionality. Additionally, the figures primarily illustrated thoracic slices, limiting the demonstration of anatomical variability. Furthermore, this study focused exclusively on technical registration performance and neglected diagnostic accuracy. Deep-learning-based registration was not directly implemented, though it was discussed as a future direction. Future investigations should incorporate larger and more heterogeneous datasets, extend analyses to 3D volumetric registration, and explore hybrid classical deep-learning approaches. Similar recommendations have been made in recent reviews by Darzi and Bocklitz [
11] and Ramadan et al. [
12]. Finally, as this study relied exclusively on institutional data, external validation using publicly available repositories such as The Cancer Imaging Archive (TCIA) [
36] will be an important step to improve generalisability across populations and imaging platforms.
5. Conclusions and Clinical Implications
This comprehensive evaluation of multimodal image registration techniques highlights the unique strengths and limitations of Demons Image Registration, MIRT-Based Free-Form Deformation, and MATLAB Intensity-Based Registration. By systematically varying parameters such as sigma fluid, histogram bins, and interpolation methods, we identified optimal configurations for each technique, tailored to specific applications. Demon image registration demonstrated significant computational efficiency, exhibiting optimal performance at a sigma fluid value of 6, with an RMSE of 0.1529. The MIRT was effective at managing complicated anatomical deformations, achieving an RMSE of 0.1725 at a sigma fluid = 6, albeit with higher computational demands, whereas MATLAB-based image registration improved with preprocessing approaches such as contrast enhancement, achieving an RMSE of 0.1317 at an alpha = 6 and highlighting its adaptability regarding datasets with challenging intensity profiles.
This study’s findings have direct implications for clinical workflows. Demons registration, with its computational efficiency, is ideal for applications like radiation therapy planning, where speed is critical. In contrast, the MIRT’s outstanding precision makes it appropriate for applications that require precise deformations, such as neuroimaging research. MATLAB-based image registration is a versatile option, especially when preprocessing is incorporated into the workflow to handle datasets with complex intensity distributions. The findings of our research illustrate each technique’s strengths and weaknesses, emphasising the importance of parameter adjustment and preprocessing for achieving optimal outcomes. Ultimately, this study provides a core paradigm for optimising multimodal image registration algorithms, providing important insights for clinical and research applications. By addressing significant challenges in PET/CT integration, this study’s outcomes enhance diagnostic accuracy in medical imaging and contribute to advancing precision medicine. Future research studies should focus on hybrid approaches and machine learning-based optimisation techniques to improve the precision and efficiency of multimodality image registration.