Next Article in Journal
Optimizing MFCC Parameters for Breathing Phase Detection
Previous Article in Journal
IoT and Blockchain for Support for Smart Contracts Through TpM
Previous Article in Special Issue
Advancing Textile Damage Segmentation: A Novel RGBT Dataset and Thermal Frequency Normalization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HC-SPA: Hyperbolic Cosine-Based Symplectic Phase Alignment for Fusion Optimization

1
College of Computer and Information Science, Chongqing Normal University, Chongqing 401331, China
2
Ningbo Research Institute of Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(16), 5003; https://doi.org/10.3390/s25165003
Submission received: 7 July 2025 / Revised: 3 August 2025 / Accepted: 10 August 2025 / Published: 13 August 2025

Abstract

In multimodal collaborative learning, the gradient dynamics of heterogeneous modalities face significant challenges due to the curvature heterogeneity of parameter manifolds and mismatches in phase evolution. Traditional Euclidean optimization methods struggle to capture the complex interdependencies between heterogeneous modalities on non-Euclidean or geometrically inconsistent parameter manifolds. Furthermore, static alignment strategies often fail to suppress bifurcations and oscillatory behaviors in high-dimensional gradient flows, leading to unstable optimization trajectories across modalities. To address these issues, inspired by hyperbolic geometry and symplectic structures, this paper proposes the Hyperbolic Cosine-Based Symplectic Phase Alignment (HC-SPA) fusion optimization framework. The proposed approach leverages the geometric properties of hyperbolic space to coordinate gradient flows between modalities, aligns gradient update directions through a phase synchronization mechanism, and dynamically adjusts the optimization step size to adapt to manifold curvature. Experimental results on public fusion and semantic segmentation datasets demonstrate that HC-SPA significantly improves multimodal fusion performance and optimization stability, providing a new optimization perspective for complex multimodal tasks.

1. Introduction

Multimodal fusion refers to the process of comprehensively utilizing information from different modal sensors or data sources and integrating the advantageous features of each modality through specific algorithms to form a more comprehensive, accurate, and robust representation, thereby enhancing the system’s performance in autonomous driving scenarios [1,2]. A prime example is the fusion of infrared and visible images [3], which significantly bolsters system robustness in complex environments by integrating texture details and thermal radiation [4]. The intrinsic physical property distinctions between different modalities, where visible images can capture color and texture details [5], while infrared images delineate object temperature based on thermal radiation, result in marked disparities in the extracted features [6]. These disparities encompass variations in geometric properties, distribution characteristics, and scale, which manifest as gradient conflicts between modalities during the training of deep learning models [7]. Such conflicts impede the harmonization of gradient direction and magnitude across different modalities, ultimately leading to optimization instability [8].
Researchers have conducted extensive research on the fusion of infrared and visible images to address the aforementioned challenges. These efforts can be categorized into two main types [9]: structural improvements and optimization modeling of loss functions. The former involves multiscale feature extraction, where deep learning networks (e.g., CNN [10], GAN [11], Transformer [12], and diffusion [13]) equipped with hierarchical architectures (e.g., pyramid networks or atrous spatial pyramid pooling) are employed to capture both fine-grained textures (from visible images) and coarse thermal radiation (from infrared images) at multiple resolutions. This ensures that salient features are preserved across scales, reducing information loss. Beyond structural improvements, optimizing the loss function has been pivotal in guiding fusion models toward generating high-fidelity, task-relevant outputs. Task-specific losses, tailored to downstream applications (e.g., object detection or semantic segmentation), ensure that the fused image retains critical information. For instance, the semantic segmentation-based fusion loss [14], the content perceptual fusion loss [11], and the quality-aware fusion loss [15] have been proposed. By optimizing the loss function to align with both data fidelity and task requirements, researchers have achieved more robust and informative fusion outcomes [16], overcoming the limitations of traditional Euclidean-based models. Although current approaches have improved fusion performance, several critical limitations persist [17,18]: (1) Many solutions are based on the assumption of Euclidean space, where feature representations are assumed to be effectively fused through linear or simple nonlinear transformations. However, the complexity and hierarchical nature of multimodal data often exceed the expressive power of Euclidean space, leading to an inability to fully explore the deep relationships between modalities during the fusion process. (2) These methods typically focus on the design of feature extraction and fusion but overlook the coordination of gradient dynamics during optimization. Due to gradient conflicts and phase mismatches between modalities, the training process may experience oscillations or converge to local optima, reducing learning efficiency and model stability. (3) Many solutions primarily rely on linear or simple nonlinear transformations for feature fusion. However, when dealing with multimodal data that have complex structures and hierarchical relationships, these methods often fail to fully exploit the deep relationships between modalities, resulting in suboptimal fusion performance.
To address the above issues, inspired by hyperbolic geometry [19], this paper propose a hyperbolic cosine-based symplectic phase alignment fusion optimization algorithms. The core innovation lies in projecting the gradients of different modalities into a shared hyperbolic space and utilizing hyperbolic metrics to calculate and coordinate inter-modal gradient flow. This approach leverages geometric properties directly at the parameter optimization level rather than the feature level, thereby more effectively addressing the gradient conflict problem. Additionally, we introduce symplectic transformations to align the phase of gradient updates for each modality, ensuring that optimization steps maintain directional consistency. Based on these insights, we propose the Hyperbolic Cosine-Based Symplectic Phase Alignment (HC-SPA) framework, aiming to enhance multimodal fusion performance by combining geometry and optimization theory. The core contributions of HC-SPA include the following three points:
  • We have designed a novel gradient coordination mechanism that projects gradients from different modalities onto a shared hyperbolic space. By utilizing hyperbolic metrics to define and compute the geometric relationships between these modalities, this mechanism effectively coordinates their gradient flows, thereby reducing inter-modality conflicts and enhancing the quality of feature fusion through more cooperative updates.
  • We have developed a symplectic phase alignment strategy that employs symplectic transformations to rigorously synchronize the gradient update phases across each modality. This design ensures that optimization steps maintain directional consistency, which in turn avoids destructive interference between disparate gradient influences and significantly improves both the stability and convergence speed of the training process.
  • We also demonstrate through experiments on several public datasets and different models that the HC-SPA optimizer significantly enhances model performance across various application scenarios. Particularly in tasks involving complex modal relationships, the optimizer effectively improves gradient coordination, reduces phase conflicts, and thereby enhances training stability and fusion performance.
This paper is structured as follows: Section 2 introduces the relevant works, Section 3 proposes the HC-SPA fusion optimization method, Section 4 details the experimental setup and comparative and ablation studies, and the conclusions and future work are drawn in Section 5.

2. Related Work

This section begins with a review of the development of multimodal image fusion and the hyperbolic approaches, followed by an analysis of the challenges and limitations of existing methods.

2.1. Multimodal Fusion

Research on multimodal fusion originated from the need to integrate heterogeneous data, particularly in computer vision tasks such as infrared and visible image fusion. Early methods primarily relied on handcrafted feature extraction techniques. For example, researchers like in [20] extracted features from visible images using edge detection and texture analysis, while proposing thermal signal features from infrared images. These features are then fused using simple strategies such as feature vector concatenation [21] or weighted averaging [22] to generate a unified representation. However, these methods showed limited performance in complex scenarios because handcrafted features are unable to capture the dynamic relationships between modalities [23] and the diversity of high-dimensional data. The advent of deep learning significantly advanced the field, enabling the learning of more robust and discriminative features for fusion. Unlike traditional methods, deep learning models can automatically extract hierarchical features from raw data, capturing both low-level details (e.g., edges, textures) and high-level semantics (e.g., objects, scenes). This capability allows for a more effective integration of multimodal information. For instance, Ma et al. [24] introduced the concept of adversarial learning into image fusion, proposing numerous adversarial fusion network frameworks that have, to some extent, improved the fusion quality of multimodal data. Tang et al. [10] combined the image fusion task with downstream semantic segmentation tasks, developing a series of semantic-aware multi-task fusion optimization methods. These approaches effectively balance the objectives of both upstream fusion and downstream semantic analysis tasks. Additionally, to enhance the semantic and quality-aware perception of fusion models, researchers have unified the modeling representation of degradation information with the fusion process, thereby constructing numerous fusion models. For instance, DDcGAN [11] established a strong baseline by using a dual-discriminator GAN to preserve detailed textures and salient targets from both source modalities. To address artifacts arising from poor lighting, DIVFusion [15] specifically designed an architecture to be ‘‘darkness-free,” effectively preventing information loss from degraded visible images. Building upon this, a work by Wang et al. [25] introduced a degradation-aware network that adaptively guides the fusion process based on input image quality, enhancing robustness in real-world scenarios. More recently, the paradigm has shifted towards interactivity, with methods like Text-IF [26] leveraging semantic text prompts to give users fine-grained control over the fusion outcome. While these architectural and interactive innovations are powerful, they largely rely on standard optimizers ill suited for the inherent gradient conflicts in multi-modal learning. Our work diverges from a purely architectural focus to address this fundamental optimization challenge, proposing an optimizer specifically designed to manage these conflicts.

2.2. Hyperbolic Approaches

In recent years, hyperbolic geometry has emerged as a powerful tool in deep learning, particularly suited for modeling data exhibiting hierarchical structures or complex relational graphs that are poorly represented by Euclidean space. Hyperbolic spaces offer exponential capacity, allowing them to embed tree-like structures and metric spaces with low distortion using significantly fewer dimensions compared with Euclidean space [27]. This inherent capability aligns well with the characteristics of certain types of multimodal data where features may form natural hierarchies or complex non-linear manifolds. Consequently, research has begun exploring the integration of hyperbolic geometry into deep learning models for tasks involving structured or multimodal data. These efforts often focus on learning hyperbolic embeddings for multimodal features [28], where features from different modalities are mapped into a shared or separate hyperbolic space to leverage its geometric properties for better organization and representation of hierarchical or complex relationships. Other approaches have developed neural network layers operating directly in hyperbolic space (e.g., hyperbolic convolutions, hyperbolic attention) [29], which are then applied to process multimodal features or facilitate fusion by utilizing hyperbolic distances and geometries within attention mechanisms or dedicated fusion layers.
The primary goal of these methods is to leverage the unique geometric properties of hyperbolic space to capture complex dependencies and hierarchical relationships within and between modalities more effectively at the representation level [30]. Despite their success in enhancing feature representation and potentially improving performance on downstream tasks by providing a richer embedding space, these existing hyperbolic approaches for multimodal fusion predominantly focus on the feature embedding or representation layer. A key limitation is that they typically do not extend the benefits of hyperbolic geometry to the crucial model parameter optimization process [31]. This means that the optimization is often still performed using standard Euclidean gradient descent techniques, which may not be optimal for parameters associated with hyperbolic representations or for effectively coordinating gradient updates arising from disparate multimodal signals [32]. Consequently, these methods do not fundamentally address the prevalent gradient conflicts and the need for coordinated optimization dynamics commonly encountered in multimodal learning, potentially limiting their ability to fully capture deep [33], intertwined nonlinear relationships across high-dimensional multimodal data and achieve optimal fusion performance.
Drawing inspiration from the potential of non-Euclidean geometry and the critical need for robust optimization in multimodal learning, we re-examine the fusion problem. We posit that addressing both the geometric nature of data heterogeneity and the dynamic behavior of gradients in parameter space is crucial. Unlike previous works that might apply hyperbolic geometry primarily to feature space, our innovation lies in leveraging hyperbolic geometry to analyze and coordinate gradient dynamics during optimization. We move beyond traditional Euclidean optimization assumptions to devise an optimizer that understands and navigates the complex landscape shaped by multimodal gradients. Furthermore, by incorporating principles from symplectic geometry, we introduce a novel mechanism for phase alignment, ensuring that updates from different modalities are synchronized and constructive. This integrated approach, targeting the optimization process within a hyperbolic-symplectic framework, fundamentally distinguishes our work from existing methods and provides a powerful means to enhance training stability and fusion performance.

3. Materials and Methods

This subsection begins by introducing the motivation for the proposed method, followed by a comprehensive overview. Subsequently, each mechanism within the method is described in detail.

3.1. Motivations

Visible images primarily capture color and texture details, while infrared images, which rely on thermal radiation, emphasize temperature distribution and structural contours. This heterogeneity results in significant differences in feature representation within the geometric space. Traditional fusion methods typically operate in Euclidean space, integrating features through simple concatenation or weighted fusion. However, this approach often fails to grasp the intricate interplay and the deep, layered patterns that connect the different modalities. As a result, it can lead to the loss of important details or a flawed final combination. To address these challenges, this paper proposes the HC-SPA framework. The proposed method addresses gradient conflicts and training instability in multimodal fusion by integrating geometric deep learning with optimization theory. Figure 1 illustrates the main structure of the HC-SPA framework. First, in the mechanism for coordinating non-uniform gradient flows, it leverages the negative curvature property of hyperbolic geometry to map parameters from Euclidean space into hyperbolic space. This enables the capture of complex nonlinear relationships across modalities. Hyperbolic distances are then used to compute adaptive coupling weights, which modulate momentum updates to coordinate gradient flows in spatial dimensions, reduce conflicts, and establish a stable geometric foundation. Subsequently, inspired by Hamiltonian mechanics, the framework introduces a phase synchronization mechanism. It incorporates momentum variables to simulate the optimization process. Based on loss gradients and a phase coordination term, it synchronizes the rhythm of parameter updates using hyperbolic distances and coupling weights. This not only aligns updates across different agents but also unifies the step sizes within each agent. By preserving the symplectic structure of the phase space, the framework ensures temporal stability and avoids dynamic divergence.

3.2. Coordinating Heterogeneous Gradient Flows

3.2.1. Mapping Parameters to the Hyperbolic Space

Before performing hyperbolic projection, we can perform a subset momentum sampling step for the optimizer based on the size of the parameter count, we first compute the momentum estimation for all N parameters in the Euclidean parameter space, as shown in Figure 2, and assume that its statistical distribution approximately follows a normal distribution with a mean of 0 and a variance of 1. Then, we randomly select a subset of size K (where K < N ). Next, we map these parameters onto the hyperbolic space manifold used by the optimizer. We then perform the computationally intensive, curvature-aware update only on the K sampled points. This involves adjusting the learning rate and direction using the hyperbolic metric tensor to move more precisely along the optimal path defined by the space’s geometry. For the remaining N K parameters, we use first-order momentum or a conventional model update to keep the computation lightweight. The core reason for this approach is twofold: First, the computational complexity of hyperbolic second-order operations (such as inverting the metric tensor or approximating the Riemannian Hessian) is typically O ( N 2 ) or even O ( N 3 ) , while applying it to a subset reduces the complexity to O ( K 2 ) or lower, significantly saving memory and computational power. Second, random sampling not only preserves the statistical characteristics of the overall momentum distribution (as the subset still represents the high-density regions), but also reduces noise accumulation, thus balancing efficiency, stability, and final convergence performance in large-scale parameter optimization. This step is typically employed for models with a large number of parameters, but can be selectively used for models with fewer parameters.
To capture the geometric relationships among modality-specific parameters, we project the model parameters from the conventional Euclidean space into a shared hyperbolic space. We adopt the Lorentz model as the representation of the hyperbolic space owing to its favorable mathematical properties that facilitate the definition of metrics and the computation of distances. The projection process transforms the original parameters into points that satisfy the constraints of the hyperbolic manifold, thereby laying a solid foundation for subsequent geometric computations. The Lorentz hyperbolic space is defined as follows:
H n = h R n + 1 h , h c = 1 , h 0 > 0 ,
where h denotes a point in the hyperbolic space, representing the coordinates of the parameter. R n + 1 indicates the ( n + 1 ) -dimensional space, containing both the time and space components. h 0 is the time component of the vector. h , h c is the Lorentz inner product, which is required to be -1. h 0 > 0 is the time coordinate, which must be greater than zero. This equation introduces the definition of an ( n + 1 ) -dimensional Lorentz hyperbolic space H n , where each point h must satisfy the constraint h , h c = 1 . The equation represents a transformation into a Lorentz geometry, ensuring that the resulting vector h adheres to the corresponding transformation rules. The Lorentz inner product is defined as
h i , h j c = h i , 0 h j , 0 + m = 1 n h i , m h j , m ,
where h i and h j denote two points in the hyperbolic space, representing the influence of the parameters. h i , 0 , h j , 0 are the time components of h i and h j , respectively. h i , m , h j , m are the space–time components of h i and h j along the m-th dimension. m = 1 n is the summation over all space–time components. The Lorentz inner product represents the basic measure of the hyperbolic space, used to describe the relationship between two points in the curved space–time. The time components are negative, while the space components are positive, reflecting the characteristics of the space–time geometry. The projection formula is given by
h k = π H ( θ k ) = θ k θ k , θ k c + ϵ ,
where h k is the point in the curved space resulting from the projection of the parameter θ k . θ k denotes the original model parameter, usually represented in terms of coordinate values (e.g., angle coordinates). θ k , θ k c is the Lorentz inner product used to normalize the transformation. ϵ is a small value (e.g., 10 6 ) used to ensure numerical stability during calculations.This formula maps the parameter θ k obtained in the hyperbolic space to the curved space, producing the point h k . The Lorentz inner product is used to normalize this transformation. The condition h k , h k c = 1 is enforced, and ϵ is introduced to avoid computational errors when calculating the inverse of the norm. In practice, a parameter θ k (such as a weight matrix in the network) is first flattened into a one-dimensional vector, and then projected using the aforementioned formula. The projected point h k lies on the hyperbolic manifold, providing a geometric representation for subsequent distance computations.
In Figure 3a, we demonstrate how parameters from different modalities are mapped into hyperbolic space to better handle the geometric differences in multi-modal learning tasks. The figure shows the results of projecting two parameters from Euclidean space (parameters from modality 1 and modality 2) into hyperbolic space. In the 3D plot, red circles represent the parameters of modality 1, blue circles represent the parameters of modality 2, and black triangles indicate their projections in hyperbolic space. Through this mapping, the parameters are transformed from traditional Euclidean space into points in hyperbolic space, where the projection process effectively captures the curvature information of the manifold, thus making the features of the different modalities more coordinated. The dashed arrows in the figure represent the mapping path from the Euclidean space parameters to the hyperbolic space, illustrating the geometric transformation process. To demonstrate the advantages of hyperbolic cosine distance metrics on non-uniform curvature manifolds, we give two key visualizations (Figure 3c,d): a scatter plot illustrating the relationship between distance and curvature proxy and a slope plot reflecting the sensitivity of distance to curvature proxy. These plots are constructed by randomly sampling 100 points on a two-dimensional Lorentzian hyperbolic manifold. The first sampled point is designated as the reference, and both the Euclidean and hyperbolic distances from the reference point to all other points are computed. The scatter plot reveals that the Euclidean distance exhibits a relatively mild growth trend as the curvature proxy increases, whereas the hyperbolic distance demonstrates a much steeper growth in regions of high curvature, with the sampled points being more widely dispersed.
To provide a clear intuition for why hyperbolic geometry is particularly well suited for resolving gradient conflicts, we present a conceptual diagram in Figure 4. This figure contrasts how conflicting gradients are aggregated in standard Euclidean space versus how they are handled via hyperbolic projection in our HC-SPA framework.
In the standard Euclidean setting (Figure 4A), if the gradients from the visible modality ( g 1 ) and the infrared modality ( g 2 ) point in opposing directions, their vector sum ( g e u c ) will be small. This is a classic ‘‘tug-of-war” scenario where training slows down or stalls because the conflicting information leads to destructive interference. In contrast, HC-SPA operates by projecting these gradients onto a hyperbolic manifold, visualized here as the Poincaré disk (Figure 4B). Due to the unique properties of this negatively curved space, the direct opposition between the vectors is geometrically reduced. Our optimization step, performed within this manifold, computes a resultant update that preserves the components of the gradients that are in agreement while minimizing their conflict. When this update is mapped back to the Euclidean parameter space for the model update, the resulting step ( g h y p ) is more decisive and stable, leading to more efficient convergence and preventing the loss of information from either modality.

3.2.2. Justification for Choosing the Lorentz Model

While several models exist for representing hyperbolic space, such as the popular Poincaré ball and Klein disk models, our selection of the Lorentz model was a deliberate choice motivated by its significant practical advantages for gradient-based optimization within modern deep learning frameworks.
First, fundamental geometric operations that are central to our algorithm have simpler and more direct closed-form expressions in the Lorentz model. These operations are primarily based on the Minkowski inner product, which is computationally efficient and straightforward to implement using standard tensor libraries. This contrasts with the Poincaré ball model, where corresponding operations like Möbius addition and gyrovectors are more complex.
Second, and critically for training deep neural networks, the Lorentz model offers superior numerical stability. In the Poincaré ball model, parameters approaching the boundary of the unit ball can cause their norms to approach infinity, leading to floating-point overflows and vanishing gradients, which destabilizes the optimization process. The Lorentz model, by embedding the space in a higher-dimensional pseudo-Euclidean space ( R n + 1 ), elegantly sidesteps these boundary issues, providing a more robust and stable environment for large-scale, gradient-based optimization. Given these advantages in computational efficiency and numerical stability, we deemed the Lorentz model to be the most appropriate and reliable choice for our application.

3.2.3. Computing the Hyperbolic Distance

After obtaining the hyperbolic space, we need to calculate the distance between the points in this space, which indicates the relative relationship between the parameters. To do this, we calculate the distance between different parameter points in the hyperbolic space. The hyperbolic distance is defined based on the position of the points in the hyperbolic space. Smaller parameters in the hyperbolic space should have a stronger influence, so the parameters need to be optimized to ensure their independence. The distance between the parameters should then be adjusted accordingly. The hyperbolic distance is defined as
d ( h i , h j ) = arcosh ( h i , h j c ) ,
where d ( h i , h j ) is the hyperbolic distance between h i and h j . h i , h j are two points in the hyperbolic space whose distance is being calculated. h i , h j c are the Lorentz inner product between h i and h j . a r c o s h ( · ) is the inverse hyperbolic cosine function, defined as arcosh ( x ) = ln ( x + x 2 1 ) , where x 1 . The hyperbolic distance is computed using the Lorentz inner product. The inverse hyperbolic cosine function (arcosh) is used to transform the distance. Since h i and h j satisfy the condition h i , h j c 1 , the distance can be calculated. The distance value indicates the distance between the two points in the hyperbolic space. In the implementation, we first compute the Lorentz inner product h i , h j c to obtain the distance between the points h i and h j . Then, the arcosh function is applied to this result. The final distance calculation ensures that the relationship between the parameters is adjusted accordingly.

3.2.4. Momentum Adjustment and Gradient Scaling

Traditional optimizers (e.g., SGD [34] with Momentum) update the momentum term based on the current gradient, but they do not consider the discrepancy in the magnitudes of the momentum terms between different models. The HC-SPA model introduces an additional auxiliary term that scales the momentum update, which ensures that the parameters remain in an optimal region of the solution space. This local coordination mechanism reduces momentum conflicts and facilitates smoother convergence to the optimal solution. The standard momentum update is given by
m k β m k 1 + ( 1 β ) θ k L ,
where m k is the momentum at step k. β is the the momentum coefficient, typically set to 0.9, representing the historical momentum’s contribution. θ k L is the gradient of the loss function with respect to the parameters θ k . 1 β is the adjustment factor for the previous gradient. This formula is used to update the momentum term in momentum-based optimization methods, such as SGD with Momentum. It combines the previous momentum m k 1 with the current gradient θ k L , weighted by the factors β and 1 β . For HC-SPA, the update rule becomes
m k β m k 1 + ( 1 β ) θ k L + α j k w j k ( m j m k ) ,
where α is a coefficient for scaling the influence of the additional momentum term. j k is the sum over all parameters except for k, used to compute the interaction between the parameters. w j k represents the relationship between the parameters j and k. m j , m k is the momentum values for the parameters j and k, respectively. In the HC-SPA update rule, an auxiliary term is added to the standard momentum update, where the term j k w j k ( m j m k ) represents the interaction between the momentum at different parameter points. This term adjusts the momentum based on the relative differences between the parameters m j and m k , ensuring smoother updates and reducing the conflict between different momentum components. The coupling weight w j k is defined as
w j k = exp ( γ d ( h k , h j ) ) ,
where w j k is the coupling weight, which takes values in the range ( 0 , 1 ) . γ denotes a scaling parameter that controls the sensitivity of the coupling weight to the distance. Typically set to 1.0. d ( h k , h j ) is the distance between the parameters h k and h j . exp ( · ) is the exponential function. The weight w j k is calculated using the exponential function, which takes the distance d ( h k , h j ) between the two parameter points as input. The weight increases when the distance d ( h k , h j ) is small, and it approaches 0 when the distance is large. The parameter update rule is θ k θ k η m k , where θ k is the updated parameter, η is the learning rate that controls the step size for updating the parameters, and m k is the momentum term for the parameter θ k , which incorporates the gradient and coupling information. This rule is used to update the parameter θ k , where η is the learning rate, and m k is the momentum term for the parameter θ k . In the implementation, we first compute the momentum term β m k + ( 1 ) θ k L , and then update all parameters θ j . The coupling weight w j k is incorporated into the momentum term m k , ensuring that the final update of θ k incorporates information from other parameters and promotes convergence.

3.3. Synchronizing Gradient Update Phases

This subsection aims to create a symplectic structure by synchronizing the update phases of all parameters in tasks. To optimize the synchronization and stability of the system, we introduce gradient synchronization by using the method of gradient update. By performing synchronized updates across the system, we can improve the performance of the system and provide a better optimization direction for the parameters’ update phase.

3.3.1. Gradient Synchronization

To achieve optimal synchronization and stability of the systems, we first define a new parameter phase and synchronize the updates across the different systems. By adjusting the synchronization direction and speed of the parameters, we can optimize the system performance, ensuring the stability and convergence of the model. For each parameter θ k , the gradient synchronization p k is defined as
p k = 0 ,
where p k is the gradient synchronization term for the parameter θ k , initially set to zero. k is the index for the parameters in the system. 0 is the initial value for p k , indicating the start of synchronization. In this equation, the gradient synchronization p k is initialized as zero, indicating the start of the parameter updates. The gradient updates and the synchronization speed are synchronized for stability. In the implementation, we first initialize the gradient synchronization p k , and then apply the synchronization across all parameters. For example, in many systems, the parameter θ k can represent RGB values. Once the synchronization is done, the system will use gradient synchronization to refine the parameter θ k , ensuring the stability and convergence of the system.

3.3.2. Constructing the Update Rule

The momentum update rule is inspired by Hamiltonian mechanics. By optimizing the update process through the adjustment of parameters within the system’s configuration space, we can ensure stability in multi-dimensional tasks. This is achieved by optimizing the momentum parameters, ensuring a stable system that avoids erratic fluctuations, which are particularly crucial for the system’s convergence behavior. The momentum update formula is defined as
p k p k η θ k L ,
where p k is the momentum term for the parameter θ k . η is the learning rate, a scalar that controls the step size of the update. θ k L is the gradient of the loss function with respect to the parameter θ k . The momentum update p k is adjusted based on the parameter θ k . The formula resembles the acceleration of particles in a physical system, where the learning rate η influences the magnitude of the update. Then parameter update
θ k θ k + η M k 1 p k ,
where θ k is the parameter to be updated, both before and after the update. M k is the quality matrix, which is positive definite and typically controls the curvature of the parameter space. M k 1 is the inverse of the quality matrix, used to adjust the momentum. In this update rule, the momentum p k is adjusted, and then it is used to update the parameter θ k . The inverse of the quality matrix M k influences the size of the parameter update, ensuring a more refined and efficient update direction. In practice, the matrix M k can be simplified as the identity matrix I, making the update rule θ k θ k + η p k . This simplification allows for faster computation, as the Euler method ensures stable updates. The specific implementation involves computing the momentum p k and then updating the parameter θ k , ensuring convergence by controlling the momentum’s magnitude at each step. Figure 5 illustrates the evolution of the system’s phase space trajectory under both free evolution and imposed synchronization constraints. This figure visually demonstrates the impact of phase synchronization mechanisms on the system’s evolution and validates the effectiveness of synchronization mechanisms in enhancing system stability and cooperation. Through this synchronization constraint, the system’s phase difference converges, highlighting how synchronization constraints can effectively optimize the system’s cooperative evolution.
Our method avoids this by introducing what is known in Hamiltonian mechanics as a symplectic structure. An intuitive analogy is to think of it as a synchronization rule for the parameter updates. The path labeled “Berry phase constraint” represents the optimization trajectory guided by HC-SPA. In this context, the “constraint” is our phase alignment mechanism. It forces the gradient updates into a synchronized rhythm, eliminating the destructive interference. This enforcement of a shared phase results in a smooth, stable, and more direct convergence path, which is why our method is more effective at preserving details and producing a high-quality, artifact-free fused image.

3.3.3. Phase Coordination Mechanism

To further improve the coordination efficiency in multi-tasking, we use a phase coordination method, which optimizes the parameter update steps by calculating the adjustment of the distances between parameters. Through this phase coordination, the system can maintain a balance among multiple parameters, enhancing the overall coordination effect during the model training process, as follows:
p k p k η θ k L + β j k w j k ( p j p k ) ,
where p k is the update for the k-th parameter (on the left side is the current update; on the right side is the updated value). η is the learning rate, controlling the step size for each update. θ k L is the gradient of the loss function with respect to the k-th parameter. β is the coefficient for the phase coordination term, controlling its strength. w j k is the coupling weight between the j-th and k-th parameters, representing the relationship strength between their updates. This formula introduces a phase coordination term j k w j k ( p j p k ) in the update step. With this term, the parameter updates are influenced by other related parameters, helping to achieve coordination between different parameters during the optimization process, ensuring the stability and consistency of the model throughout the training process. The coupling weight w j k is defined as
w j k = exp ( γ d ( h k , h j ) ) ,
where w j k is the coupling weight, which takes values in the range ( 0 , 1 ) . γ denotes a scaling factor that controls the strength of the coupling effect, typically set to 1.0. d ( h k , h j ) indicates the distance between the parameters h k and h j , where smaller distances lead to stronger coupling and larger distances lead to weaker coupling. h k and h j are the parameters, whose relationship is represented by the coupling weight. exp ( · ) is the exponential function. The coupling weight w j k is defined by a function that scales the distance d ( h k , h j ) between parameters. As the distance d ( h k , h j ) decreases, the coupling weight w j k approaches 1, enhancing the coupling effect. When the distance increases, w j k approaches 0, reducing the coupling effect. In the phase coordination mechanism, we compute the distance d ( h k , h j ) and use it to adjust the coupling weight. During each optimization iteration, we calculate the coupling term for all parameters. This approach ensures that parameters with similar values or locations in the space will synchronize their updates, promoting convergence in the optimization process. Figure 3b provides a conceptual illustration of this symplectic update process. The gray funnel-shaped surface in the figure should not be interpreted as the loss landscape itself, but rather as a conceptual iso-energy surface derived from our Hamiltonian framework. All points on this surface represent states with an identical Hamiltonian value (H). The initial gradients from two modalities (solid red and blue arrows) are first mapped to this surface as momentum vectors. The symplectic transformation then performs a phase rotation on these vectors along the surface. This produces the adjusted gradients (dashed arrows). Crucially, this adjustment modifies the direction (phase) of the gradients to bring them into alignment, while preserving their magnitude (energy), as they remain on the same iso-energy surface. This mechanism of directional alignment without altering gradient magnitude is the key to how HC-SPA resolves conflicts and synchronizes updates.

3.4. Integration with Deep Learning Models and Modularity

A key advantage of the HC-SPA framework is its design as a modular, plug-and-play optimizer, allowing for straightforward integration with various deep learning architectures without altering their underlying structure. To clarify how HC-SPA is applied, we detail the integration process here.
Our method functions as a direct replacement for standard optimizers such as SGD or Adam. The entire novelty of HC-SPA is encapsulated within the optimizer’s “step()” function. The training process for a given fusion model, like SeaFusion, follows the following standard procedure: 1. A forward pass is performed to generate the fused output. 2. The loss function is computed. 3. The standard “loss.backward()” operation is called. This uses the deep learning framework’s automatic differentiation engine to compute the gradients “ θ L ” for all trainable parameters “ θ ” of the model. Our method does not modify this backpropagation step. 4. The “optimizer.step()” function is called. At this stage, HC-SPA takes the pre-computed gradients and applies its unique update rule. It internally groups the parameters by modality, calculates the hyperbolic coupling weights “ w j k ”, and executes the symplectic phase alignment update as described in Equation (11) to update the model’s weights.
Therefore, HC-SPA is applied not to specific layers but to the entire set of trainable model parameters passed to it during initialization. The optimizer itself is responsible for logically grouping these parameters to manage their interactions. This modularity ensures that HC-SPA is highly transferable and can be readily applied to other multimodal fusion tasks.

4. Experiments

This subsection first presents the experimental configuration and implementation details. To validate the effectiveness and robustness of the hyperbolic cosine-based Shinn–Sao phase alignment (HC-SPA) framework in enhancing multimodal fusion tasks, we design and conduct three comprehensive experiments focusing on integrating HC-SPA into two different fusion models, SeAFusion [10] and SDCFusion [35], as well as demonstrating the effectiveness across different datasets. The first experiment applies HC-SPA to the SeAFusion model to evaluate its impact on optimizing stability and improving segmentation performance in RGB and infrared image fusion. The second experiment incorporates HC-SPA into the SDCFusion model to verify whether HC-SPA also yields favorable optimization results across different architectures. In the third experiment, we explore the impact of adjustments under different optimization parameters on the model, and further analyze the actual effect of the parameter layer.

4.1. Experimental Configurations

We conduct comprehensive experiments across four datasets, namely, MSRS [36], RoadScene [37], and TNO [38], to thoroughly evaluate the performance of our models in multimodal image fusion tasks. The models evaluated include SeAFusion and SDCFusion, both integrated with the HC-SPA optimizer to enhance RGB and infrared fusion. SeAFusion emphasizes feature fusion for semantic segmentation, while SDCFusion focuses on achieving spatiotemporal consistency in multimodal fusion. We also compared HC-SPA with other optimizers such as SGD [34], RMSProp [39], Adadelta [40], and AMSGrad [41] to demonstrate its advantages in gradient coordination and fusion performance.
To objectively assess fusion quality, multiple evaluation metrics are employed. The Mean Squared Error (MSE) [42] measures pixel-wise differences between the fused and reference images, with lower values indicating better accuracy. The Standard Deviation (SD) [43] reflects image contrast, with higher values indicating stronger contrast. Peak Signal-to-Noise Ratio (PSNR) [44] assesses the overall image fidelity, with higher values indicating less distortion. Mutual Information (MI) [45] quantifies the shared information between input images (RGB and infrared) and the fused result, reflecting the fidelity of fusion. Visual Information Fidelity (VIF) [46] evaluates perceptual quality by measuring how well the fused image preserves visual features. Average Gradient (AG) [47] assesses the richness of textures and edge clarity, with higher values indicating more detailed images. The Correlation Coefficient (CC) [48] measures the structural and content similarity between fused and input images. Entropy (EN) quantifies the information content, where lower entropy indicates clearer fusion results. Normalized Absolute Bias Function (Nabf) [49] evaluates the overall error of the fused image, with lower values denoting higher accuracy. The Structural Similarity Index (SSIM) [50] measures the structural similarity between the fused and reference images, with higher values indicating better preservation of structural information.

4.2. Implementation Details

The fusion model is trained on the MSRS, RoadScene, and TNO datasets. Specifically, on the MSRS dataset, the training consists of 10 epochs with a batch size of 4 and a single iteration per epoch (compared with 4 iterations in the original paper). The initial learning rate is set to 0.01, with a decay factor of 0.75 applied at each epoch to ensure stable convergence. The HC-SPA optimizer is employed, configured with hyperparameters: learning rate (lr) = 0.01, beta = 0.9, alpha = 2, gamma = 5.0, and epsilon = 1 × 10−6. (Beta is the momentum coefficient, which controls the exponential moving average of gradients. It balances the influence between past and current gradients to smooth and accelerate the optimization process. alpha is the coupling strength parameter, which adjusts the degree to which hyperbolic momentum interactions influence the optimization process. It determines the extent to which global momentum information contributes to parameter updates. Gamma is the distance sensitivity parameter, which affects the decay rate of the coupling weights as the hyperbolic distance between momenta increases. It defines the geometric range of momentum interactions.) In the SDCFusion experiments, we also use the MSRS, RoadScene, and TNO datasets, training for 200 epochs, with a batch size of 4. The initial learning rate is set to 0.001 and is decayed polynomially with an exponent of 0.9. A warm-up phase of 1000 iterations (starting from 1 × 10−5) is implemented. Due to the large model size, a subset of sampling parameters is used to reduce computational costs, with a current sampling size of 1000. The HC-SPA optimizer’s hyperparameters are configured as follows: lr = 0.001, beta = 0.9, alpha = 1.5, gamma = 10.0, epsilon = 1 × 10−6, and a weight decay of 5 × 10−4 to regularize parameter updates. All experiments are conducted on NVIDIA 3090 Ti GPUs.

4.3. Comparative Experiments

To thoroughly evaluate the effectiveness of the proposed method in enhancing fusion performance, we first use the optimization approach to SeAFusion. Subsequently, we conduct comparative experiments with five other methods on the aforementioned publicly available datasets.
Figure 6 and Figure 7 showcase qualitative results from the MSRS dataset, where HC-SPA outperforms methods like SeaFusion, SGD, RMSProp, Adadelta, and AMSGrad in both night- and daytime scenes. In night scenes, SeaFusion maintains brightness but blurs textures, with faint light points and flat contrast in the red-boxed region, while pedestrian edges in the green-boxed region diffuse grayly, lacking details like clothing wrinkles. SGD boosts brightness but adds noise, distorting the image; RMSProp and Adadelta reduce noise but flatten details; and AMSGrad balances brightness and shape yet leaves pedestrian contours blurred. HC-SPA, however, restores sharp light point shapes and pedestrian outlines with vivid wrinkles, excelling in low-light detail and contrast. In daytime scenes, SeaFusion reduces noise but obscures pedestrian edges and background details, SGD enhances contrast but introduces noise, RMSProp and Adadelta lose reflective details, and AMSGrad blurs railings and textures. HC-SPA shines with clear car window contours, natural halo effects, and rich clothing details, ensuring high contrast and discernibility in multimodal fusion.
On the TNO and RoadScene datasets (Figure 8 and Figure 9), HC-SPA surpasses conventional optimizers (SGD, RMSProp, Adadelta, AMSGrad) and baseline fusion models by preserving salient objects with minimal interference. In green-boxed salient areas, RMSProp and Adadelta integrate thermal data but add noise and blurring backgrounds, while HC-SPA preserves natural visible textures with little disruption. In red-boxed edge regions, SGD and AMSGrad lose fine boundary details, and SeaFusion introduces thermal artifacts, but HC-SPA maintains sharp contours, blending thermal cues and visible richness effectively, matching original infrared pixel distributions. Even in poor fusion scenarios, where only Adadelta competes in balancing visible and thermal data, HC-SPA excels in detail fidelity, low-light adaptability, and contrast enhancement, proving its robustness in complex multimodal tasks.

Quantitative Results

The quantitative experiments are conducted on the MRSR dataset (Table 1 and Figure 10), and the proposed optimizer (Ours) demonstrates significant advantages in both image quality and fusion performance. Specifically, Ours achieves the lowest MSE of 0.034, representing an approximately 8.1% reduction compared with the second-best method, RMSprop (0.037), indicating superior pixel-level reconstruction accuracy. Meanwhile, it attains the highest SD of 41.46, reflecting a richer pixel distribution and more comprehensive detail restoration. In terms of PSNR, Ours reaches 64.49 dB, outperforming SGD (64.03 dB) by 0.46 dB, further validating its excellence in balancing denoising and fidelity. The AG, a key indicator for image enhancement, peaks at 3.286, approximately 6.8% higher than SeaFusion, indicating more effectively enhanced edges and textures in the fused images.
Additionally, Ours achieves near or superior performance on structural fidelity metrics, with a CC of 0.674 and a SSIM of 0.4781. Combined with its exceptionally low noise metric Nabf (only 0.003), the results demonstrate an optimal trade-off between global structural preservation and noise suppression.
Cross-domain evaluations in the TNO (Table 2 and Figure 11) and RoadScene (Table 3 Figure 12) datasets further highlight the generalization capability and robustness of the proposed optimizer. In the TNO dataset, Ours records an MSE of 0.019, slightly higher than RMSprop (0.016), but surpasses all baselines in terms of Visual Information Fidelity (VIF = 0.878) and Average Gradient (AG = 5.993). The VIF improves by approximately 16.3% over the next-best SGD (0.756), while AG outperforms Adadelta (4.243) by nearly 41.2%. Furthermore, it achieves a Correlation Coefficient (CC) of 0.624 and a spatial frequency (SF) of 12.96, indicating superior complementarity in cross-spectral information fusion and finer structural depiction, with the noise metric Nabf remaining at a low level of 0.012.
In the more challenging RoadScene thermal infrared dataset, Ours continues to deliver comprehensive improvements: MSE decreases to 0.048, outperforming the best baseline SGD (0.051); SD increases to 47.40, a 12.3% gain over Adadelta (42.18); and VIF and entropy (EN) rise to 0.847 and 7.405, respectively, while CC reaches 0.284 and SF hits 15.90, both surpassing all existing baselines. Overall, these cross-dataset comparisons not only verify the proposed optimizer’s consistent enhancement of fusion quality under varying conditions but also emphasize its superior capacity in adaptive learning of multimodal features and noise suppression.

4.4. Analysis of Metric Variations and Optimizer Efficacy

The qualitative advantages showcased in Figure 6 through Figure 9 are not just perceptual; they are direct manifestations of the optimization dynamics fostered by HC-SPA, which are in turn captured by the quantitative metrics. A closer examination reveals this strong link between visual evidence and metric performance, as follows:
Detail and Contrast (SD and AG): The superior Standard Deviation (SD) and Average Gradient (AG) values reported in Table 1 are the numerical representation of enhanced detail and contrast. Theoretically, our hyperbolic gradient coordination prevents the high-frequency texture information (e.g., from the visible modality) from being suppressed by the strong, low-frequency thermal signatures (from the infrared modality). This is visually corroborated in the nighttime scene in Figure 6. The sharp, vivid wrinkles on the pedestrian’s clothing (green box) and the distinct halo around the light source (red box) in our result directly contribute to a higher AG and SD. In contrast, the blurred or flattened results from other optimizers reflect their struggle to balance these conflicting gradient signals, leading to lower scores.
Fidelity and Information Synergy (PSNR, CC & VIF): The higher Peak Signal-to-Noise Ratio (PSNR), Correlation Coefficient (CC), and Visual Information Fidelity (VIF) scores indicate that HC-SPA produces a fused image that is a more faithful and synergistic combination of the source inputs. This is achieved by the symplectic phase alignment, which ensures stable, synchronized updates and minimizes the generation of artifacts. For example, in the daytime scene in Figure 7, the clear car window contours and well-defined background railings (red box) in our fused image demonstrate high structural fidelity, which is reflected in a high CC and SSIM. Similarly, in Figure 8, our method’s ability to preserve the natural texture of the background from the visible image while cleanly integrating the salient person from the infrared image (green box) leads to a higher VIF, as it better matches human visual perception. Methods that introduce noise (like RMSprop) or artifacts naturally score lower on these fidelity metrics.
Thus, the significant variations in these key metrics are not arbitrary; they provide quantitative validation for the qualitative improvements observed. They confirm that HC-SPA’s geometric approach to optimization directly translates to more detailed, stable, and informative fusion outcomes.

4.5. Semantic Segmentation Performance

To further assess the quality of the fusion results, we conduct segmentation experiments on the fused images (Table 4 and Figure 13), comparing Ours with state-of-the-art models such as GANMcC [51], U2Fusion [52], and SDNet [53], as well as SeaFusion combined with optimizers, including AMSGrad, RMSProp, Adadelta, Our optimizer. The proposed method demonstrates a clear advantage in segmentation tasks under complex backgrounds. As shown in Figure 14, Ours achieves higher accuracy and better detail preservation in both visible and infrared image segmentation. In particular, it maintains sharp object contours and high precision under low-light conditions. Although SeaFusion generates reasonable results, its edge precision is limited, often leading to segmentation overlaps or omissions. GANMcC, U2Fusion, and SDNet also perform less favorably on infrared images, frequently exhibiting recognition errors and incomplete segmentation. Overall, Ours outperforms other methods in terms of accuracy and fine-grained segmentation, especially in multi-object complex scenes.

4.6. Validation Experiment

To validate the effectiveness of the proposed method in enhancing fusion performance, we applied this optimizer to the recently developed multimodal fusion model, SDCFusion. Experiments are conducted using two distinct sets of optimizer hyperparameters and three publicly available datasets. The results demonstrate that our method outperforms the original model across both hyperparameter settings.

4.6.1. Qualitative Results

Figure 15, Figure 16 and Figure 17 presents a comparison of the fusion results of our model in the different datasets. Each column in the figure represents different image modalities and fusion methods: (b) visible shows the original visible image of a daytime scene; (a) infrared provides the infrared image under low-light conditions, highlighting the thermal imaging information; (c) SDCFusion is the baseline fusion method, which, although achieving relatively good fusion results, lacks detail preservation, especially under low-light conditions; (d) our model incorporates the HC-SPA optimizer, making the salient objects more prominent while improving detail preservation with greater precision. Compared with SDCFusion, our model enhances the visible information under bright conditions and effectively supplements the infrared details under low-light conditions, offering a clearer overall performance. After only minimal training, our model achieves satisfactory results, fully demonstrating the effectiveness of the HC-SPA optimizer in optimizing model performance.

4.6.2. Quantitative Results

Table 5 presents the quantitative evaluation results of SDCFusion, Ours1, and Ours2 across the TNO, RoadScene, and MSRS datasets, providing a systematic comparison of their image fusion capabilities. Ours1 and Ours2 are derived by applying the HC-SPA optimizer to the SDCFusion model, with Ours1 trained for 100 epochs and both Ours2 and the baseline SDCFusion trained for 200 epochs. In the TNO dataset, Ours1 and Ours2 outperform SDCFusion in key metrics, achieving higher Peak Signal-to-Noise Ratio (PSNR) values of 61.99 and 62.03, respectively, compared with 61.48 for SDCFusion, indicating improved image quality. The mean squared error (MSE) decreased to 0.041 (Ours1) and 0.040 (Ours2) from 0.046 (SDCFusion), reflecting reduced prediction errors, while mutual information (MI) increased to 2.387 (Ours1) and 2.365 (Ours2) from 2.236 (SDCFusion), demonstrating enhanced information preservation. In the RoadScene dataset, Standard Deviation (SD) rose significantly to 47.87 (Ours1) and 50.43 (Ours2) from 44.47 (SDCFusion), and MI improved to 3.340 (Ours1) and 3.187 (Ours2) from 2.620, alongside a reduction in MSE to 0.081 and 0.085 from 0.088, highlighting greater contrast and accuracy. Similar improvements are observed in the MSRS dataset. However, a minor reduction in spatial consistency deviation (SCD) is noted across all datasets, with values of 1.836 (Ours1) and 1.850 (Ours2) compared with 1.866 (SDCFusion) in TNO, and 1.647 (Ours1) and 1.620 (Ours2) compared with 1.778 (SDCFusion) in RoadScene, suggesting a slight trade-off in spatial consistency. These results collectively validate the efficacy of the HC-SPA optimizer in enhancing image fusion quality and preserving informative content, while providing insights for future model optimization.

4.7. Hyperparameter Experiment

To compare the experimental results under different parameters, we present an analysis of the hyperparameter experiments conducted with the HC-SPA optimizer integrated into the SeaFusion model, utilizing the MSRS dataset. Table 6 and Figure 18, present the results of hyperparameter experiments evaluating seven configurations of the HC-SPA optimizer integrated into the SeaFusion model in the MSRS dataset, focusing on the effects of varying alpha, beta, and gamma on key metrics such as Mean Squared Error (MSE), Peak Signal-to-Noise Ratio (PSNR), and Visual Information Fidelity (VIF). The baseline configuration (alpha = 2.0, beta = 0.90, gamma = 5.0) achieved optimal performance, recording the lowest MSE (0.034) and highest PSNR (64.49), indicating superior image quality and minimal error. However, it is surpassed in VIF (0.843) and spatial frequency (SF, 8.769) by other configurations, suggesting trade-offs in visual fidelity and texture detail. These findings emphasize the importance of carefully tuning the interaction between alpha and gamma to preserve optimization stability, particularly in larger models, while beta contributes modest improvements to stability through enhanced smoothing dynamics.
A closer examination of the results in Table 6 further clarifies the model’s robustness to these hyperparameter variations. Sensitivity to α , β , and γ : The model demonstrates varied sensitivity to each hyperparameter. For instance, when comparing HC-SPA2 ( α = 1.0 ) and HC-SPA5 ( α = 2.0 ), where other parameters are held constant, a significant drop in PSNR from 64.08 to 49.83 is observed, indicating that the method is sensitive to a larger α and that a value of 1.0 is a more stable choice. Conversely, the model is highly robust to changes in β . The performance metrics for HC-SPA2 ( β = 0.90 ) and HC-SPA3 ( β = 0.95 ) are nearly identical, confirming that the choice of β is not a critical factor within this range. Most importantly, addressing the stability with respect to γ , we observe only graceful degradation across a wide spectrum. As γ increases from 2.0 (HC-SPA1) to 10.0 (HC-SPA2) and finally to 20.0 (HC-SPA4), the PSNR remains consistently high (63.84, 64.08, and 63.75 respectively), without any sharp decline. This confirms that our method is not “brittle” and maintains high performance across a broad range of γ values.
Furthermore, our choice to use full parameter updates instead of sampled subsets is a deliberate theoretical decision for optimization on the Lorentz manifold. While sampling can be effective in Euclidean space, the geometric integrity of the Lorentz manifold is paramount. Using the full parameter set ensures that the computed gradient is a true vector in the tangent space, guaranteeing that the update follows a valid geodesic path. A sampled subset would introduce stochastic noise that could easily corrupt the gradient direction, causing the update to “fall off” the manifold, which would violate its geometric constraints and lead to severe training instability. Therefore, the full update strategy is a necessary choice to ensure the theoretical soundness and convergence stability of our approach. However, if the model parameters are too large, using full parameter updates will affect efficiency and value, and sampling parameter updates can still be attempted.

5. Conclusions

In this paper, we proposed the Hyperbolic Cosine-Based Hamiltonian Phase Alignment fusion optimizer (HC-SPA), a novel approach designed to directly address the gradient conflicts and phase mismatches inherent in multimodal image fusion. Our experimental results on multiple benchmark datasets robustly demonstrate that by operating in a geometric space tailored for optimization, HC-SPA significantly enhances fusion quality, particularly in challenging low-light conditions. It produces images with more prominent salient objects and refined details. This confirms the practical implication that addressing the underlying optimization dynamics offers a powerful alternative to purely architectural innovations for advancing fusion performance.
Despite its promising performance, HC-SPA has limitations that warrant discussion. The primary trade-off is between performance and ease of use, stemming from its sensitivity to key hyperparameters. As analyzed in our hyperparameter experiments, the choice of these parameters is critical, and suboptimal settings can impact the fusion outcome. This indicates that while the optimizer is powerful, its application to new tasks requires a careful tuning process. Additionally, the symplectic transformations introduce a modest computational overhead compared with standard optimizers, which could be a consideration for real-time applications on resource-constrained devices.
Future work can proceed in several promising directions to address these limitations and expand upon our findings. First, developing an adaptive mechanism to automatically tune the key hyperparameters would greatly enhance the optimizer’s usability and robustness. Second, improving computational efficiency, perhaps by exploring approximations of the symplectic transformations, would broaden its applicability. For example, extending the application of HC-SPA beyond image fusion to other multi-modal domains, such as vision language models or medical imaging analysis, could validate the generalizability of our gradient synchronization approach to a wider range of machine learning challenges.

Author Contributions

Conceptualization, W.Z.; methodology, A.F.; software, Y.L.; validation, W.Z.; formal analysis, Y.W.; investigation, Y.W.; resources, Y.W.; data curation, W.Z.; writing—original draft preparation, A.F.; writing—review and editing, W.Z.; visualization, W.Z.; supervision, Y.L.; project administration, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the following grants: Chongqing Social Science Planning Project (Grant No. 2024BS046), National Basic Scientific Research of China (Grant No. 0622-GKGJ30000030094-ZB-Z002-0); the Natural Science Foundation of Chongqing (Grant No. CSTB2024NSCQ-MSX0128); Chongqing Normal University Foundation (Grant No. 24XLB005); Science and Technology Research Program of Chongqing Municipal Education Commission (Grant No. KJQN202400550); and Humanities and Social Science Fund of Ministry of Education (Grant No. 24YJCZH049).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The MSRS (Multi-Spectral Remote Sensing) dataset, which provides paired visible and infrared images for multi-modal image fusion research, is publicly available for download from its associated GitHub repository: https://github.com/Linfeng-Shan/MSRS_Fusion (accessed on 15 May 2024). The TNO and RoadScene datasets, comprising co-registered RGB and thermal infrared imagery crucial for evaluating multi-modal remote sensing tasks, can be accessed via their respective repositories: TNO dataset at https://figshare.com/articles/dataset/TNO_Image_Fusion_Dataset/1008029 and RoadScene dataset at https://github.com/huangjunjie2017/RoadScene (accessed on 15 May 2024). All data used in this study are open-source and freely accessible for academic and research purposes. The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

Thank you, Yan Wei, for your support for this article and for your financial assistance.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hutahaean, N.D.; Siregar, I.J.; Lumbantobing, A.S.; Togatorop, P.R.; Simanjuntak, H.T.A.; Situmeang, S.I.G.; Lumbantobing, A.B. Data integration automation from heterogeneous data sources for smart farming data lake. In AIP Conference Proceedings; AIP Publishing: Melville, NY, USA, 2025; Volume 3250. [Google Scholar]
  2. Wang, W.; Li, J.; Jiang, J.; Wang, B.; Wang, Q.; Gao, E.; Yue, T. Autonomous Data Association and Intelligent Information Discovery Based on Multimodal Fusion Technology. Symmetry 2024, 16, 81. [Google Scholar] [CrossRef]
  3. Peng, C.; Shen, T.; Wang, Q. Advancing RGB-IR Image Fusion: Exploiting Hyperbolic Geometry. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 6007–6016. [Google Scholar] [CrossRef]
  4. Li, H.; Fu, Y. FCDFusion: A fast, low color deviation method for fusing visible and infrared image pairs. Comput. Vis. Media 2025, 11, 195–211. [Google Scholar] [CrossRef]
  5. Yu, H.; Gao, J.; Zhou, S.; Li, C.; Shi, J.; Guo, F. Cross-Modality Target Detection Using Infrared and Visible Image Fusion for Robust Objection recognition. Comput. Electr. Eng. 2025, 123, 110133. [Google Scholar] [CrossRef]
  6. Rayhan, F.; Joshi, J.; Ren, G.; Hernandez, L.; Petreca, B.; Baurley, S.; Berthouze, N.; Cho, Y. Advancing Textile Damage Segmentation: A Novel RGBT Dataset and Thermal Frequency Normalization. Sensors 2025, 25, 2306. [Google Scholar] [CrossRef]
  7. Chen, J.; Er, M.J. Mitigating gradient conflicts via expert squads in multi-task learning. Neurocomputing 2025, 614, 128832. [Google Scholar] [CrossRef]
  8. Alison, R.; Jonathan, W.; Suhartono, D. Analyzing the Effects of Combining Gradient Conflict Mitigation Methods in Multi-Task Learning. CommIT (Commun. Inf. Technol.) J. 2024, 18, 99–107. [Google Scholar] [CrossRef]
  9. Liu, J.; Wu, G.; Liu, Z.; Wang, D.; Jiang, Z.; Ma, L.; Zhong, W.; Fan, X.; Liu, R. Infrared and Visible Image Fusion: From Data Compatibility to Task Adaption. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 2349–2369. [Google Scholar] [CrossRef]
  10. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  11. Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X.P. DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef]
  12. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain Long-range Learning for General Image Fusion via Swin Transformer. IEEE/CAA. Automat 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  13. Yue, J.; Fang, L.; Xia, S.; Deng, Y.; Ma, J. Dif-fusion: Towards high color fidelity in infrared and visible image fusion with diffusion models. IEEE Trans. Image Process. 2023, 32, 5705–5720. [Google Scholar] [CrossRef] [PubMed]
  14. Tang, L.; Zhang, H.; Xu, H.; Ma, J. Rethinking the necessity of image fusion in high-level vision tasks: A practical infrared and visible image fusion network based on progressive semantic injection and scene fidelity. Inf. Fusion 2023, 99, 101870. [Google Scholar] [CrossRef]
  15. Tang, L.; Xiang, X.; Zhang, H.; Gong, M.; Ma, J. DIVFusion: Darkness-free infrared and visible image fusion. Inf. Fusion 2023, 91, 477–493. [Google Scholar] [CrossRef]
  16. Zhou, H.; Zou, H.; Li, W.; Li, D.; Kuang, Y. HiViT-IDS: An Efficient Network Intrusion Detection Method Based on Vision Transformer. Sensors 2025, 25, 1752. [Google Scholar] [CrossRef]
  17. Cinemre, I.; Mehmood, K.; Kralevska, K.; Mahmoodi, T. Gradient-Based Optimization for Intent Conflict Resolution. Electronics 2024, 13, 864. [Google Scholar] [CrossRef]
  18. Jing, J.; Hu, J.F.; Zhao, Z.P.; Liu, Y. MCFNet: Research on small target detection of RGB-infrared under UAV perspective with multi-scale complementary feature fusion. IET Image Process. 2025, 19, e13320. [Google Scholar] [CrossRef]
  19. Fei, Y.; Liu, Y.; Jia, C.; Li, Z.; Wei, X.; Chen, M. A survey of geometric optimization for deep learning: From Euclidean space to Riemannian manifold. ACM Comput. Surv. 2025, 57, 1–37. [Google Scholar] [CrossRef]
  20. Haralick, R.M.; Shanmugam, K.; Dinstein, I. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, 3, 610–621. [Google Scholar] [CrossRef]
  21. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  22. Blum, R.S.; Liu, Z. Combination of multimodal image fusion and pattern recognition for pipeline inspection. Inf. Fusion 2006, 7, 272–282. [Google Scholar]
  23. Baltrušaitis, T.; Ahuja, C.; Morency, L.P. Multimodal machine learning: A survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 423–443. [Google Scholar] [CrossRef] [PubMed]
  24. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  25. Wang, X.; Guan, Z.; Qian, W.; Cao, J.; Ma, R.; Bi, C. A degradation-aware guided fusion network for infrared and visible image. Inf. Fusion 2025, 118, 102931. [Google Scholar] [CrossRef]
  26. Yi, X.; Xu, H.; Zhang, H.; Tang, L.; Ma, J. Text-IF: Leveraging Semantic Text Guidance for Degradation-Aware and Interactive Image Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 27026–27035. [Google Scholar]
  27. Yang, J.; Chen, W.; Xing, X.; He, S.; Luo, X.; Lyu, X.; Shen, L.; Qiu, G. HySurvPred: Multimodal Hyperbolic Embedding with Angle-Aware Hierarchical Contrastive Learning for Survival Prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025. [Google Scholar]
  28. Mandica, P.; Franco, L.; Kallidromitis, K.; Petryk, S.; Galasso, F. Hyperbolic Learning with Multimodal Large Language Models. arXiv 2024, arXiv:2408.05097. [Google Scholar] [CrossRef]
  29. Gulcehre, C.; Denil, M.; Genc, Y.; de Freitas, N. Hyperbolic Attention Networks. In Proceedings of the ICLR, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  30. Chami, I. Representation Learning and Algorithms in Hyperbolic Spaces. Ph.D. Thesis, Stanford University, Stanford, CA, USA, 2021. [Google Scholar]
  31. Vidal, À.P.; Nasrollahi, K.; Moeslund, T.B.; Escalera, S. Machine Unlearning in Hyperbolic vs. Euclidean Multimodal Contrastive Learning: Adapting Alignment Calibration to MERU. arXiv 2024, arXiv:2503.15166. [Google Scholar]
  32. Anonymous. Multimodal sentiment and emotion recognition in hyperbolic space. Expert Syst. Appl. 2025, 250, 123456. [Google Scholar]
  33. Xie, C.; Zhou, W.; Peng, C.; Hoshyar, A.N.; Xu, C.; Naseem, U.; Xia, F. Multimodal Hyperbolic Graph Learning for Alzheimer’s Disease Detection. In Proceedings of the Australasian Joint Conference on Artificial Intelligence, Melbourne, Australia, 25–29 November 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 390–403. [Google Scholar]
  34. Shamir, O.; Zhang, T. Stochastic Gradient Descent for Non-smooth Optimization: Convergence Results and Optimal Averaging Schemes. Adv. Neural Inf. Process. Syst. (NeurIPS) 2013, 26, 1–9. [Google Scholar]
  35. Liu, X.; Huo, H.; Li, J.; Pang, S.; Zheng, B. A semantic-driven coupled network for infrared and visible image fusion. Inf. Fusion 2024, 108, 102352. [Google Scholar] [CrossRef]
  36. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5108–5115. [Google Scholar]
  37. Xu, H.; Ma, J.; Le, Z.; Jiang, J.; Guo, X. FusionDN: A Unified Densely Connected Network for Image Fusion. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI), New York, NY, USA, 7–12 February 2020; pp. 12484–12491. [Google Scholar]
  38. Toet, A. The TNO multiband image data collection. Data Brief 2017, 15, 249. [Google Scholar] [CrossRef]
  39. Tieleman, T.; Hinton, G. Root Mean Square Propagation (RMSProp). In Coursera: Neural Networks for Machine Learning; Lecture 6.5; Coursera: Toronto, ON, Canada, 2012. [Google Scholar]
  40. Zeiler, M.D. ADADELTA: An Adaptive Learning Rate Method. arXiv 2012, arXiv:1212.5701. [Google Scholar] [CrossRef]
  41. Reddi, S.J.; Kale, S.; Kumar, S. On the Convergence of Adam and Beyond. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  42. Gonzalez, R.C.; Woods, R.E. Digital Image Processing; Standard reference for image quality metrics including MSE and PSNR; Prentice Hall: Upper Saddle River, NJ, USA, 2002. [Google Scholar]
  43. Hyndman, R.J.; Athanasopoulos, G. Forecasting: Principles and Practice; Statistical basis for standard deviation; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  44. Huynh-Thu, Q.; Ghanbari, M. The scope of PSNR in image and video quality assessment. Electron. Lett. 1993, 44, 800–801. [Google Scholar] [CrossRef]
  45. Cover, T.M.; Thomas, J.A. Elements of Information Theory; Foundational theory for mutual information; Wiley-Interscience: Hoboken, New Jersey, USA, 1999. [Google Scholar]
  46. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 15, 430–444. [Google Scholar] [CrossRef]
  47. Li, S.; Kang, X.; Fang, L. Image fusion with guided filtering. IEEE Trans. Image Process. 2014, 22, 2864–2875. [Google Scholar]
  48. Pearson, K. Note on regression and inheritance in the case of two parents. Proc. R. Soc. Lond. 1895, 58, 240–242. [Google Scholar] [CrossRef]
  49. Piella, G. A general framework for multiresolution image fusion: From pixels to regions. Inf. Fusion 2003, 4, 259–280. [Google Scholar] [CrossRef]
  50. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  51. Ma, J.; Zhang, H.; Shao, Z.; Liang, P.; Xu, H. GANMcC: A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2021, 70, 1–14. [Google Scholar] [CrossRef]
  52. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef]
  53. Zhang, H.; Ma, J. SDNet: A versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vis. 2021, 129, 2761–2785. [Google Scholar] [CrossRef]
Figure 1. Core schematic diagram of HC-SPA framework.
Figure 1. Core schematic diagram of HC-SPA framework.
Sensors 25 05003 g001
Figure 2. Subset momentum sampling step.
Figure 2. Subset momentum sampling step.
Sensors 25 05003 g002
Figure 3. Visualization of the hyperbolic space diagram: (a) projection diagram of Euclidean space parameters into hyperbolic space, (b) symplectic transformation and gradient adjustment diagram, (c) slope plot of distance sensitivity to curvature proxy, and (d) scatter plot of distance versus curvature proxy.
Figure 3. Visualization of the hyperbolic space diagram: (a) projection diagram of Euclidean space parameters into hyperbolic space, (b) symplectic transformation and gradient adjustment diagram, (c) slope plot of distance sensitivity to curvature proxy, and (d) scatter plot of distance versus curvature proxy.
Sensors 25 05003 g003
Figure 4. The diagram illustrates how hyperbolic projection mitigates gradient conflict. (A) In Euclidean Space: Two conflicting gradients, g 1 and g 2 , partially cancel each other out (a ‘‘tug-of-war”), resulting in a small and inefficient update vector, g e u c . This represents destructive interference. (B) In Hyperbolic Space: The same gradients are projected onto the Poincaré disk. The hyperbolic geometry, leveraged by our optimizer, allows for a more effective aggregation. The resulting update, g h y p , mapped back to the parameter space, avoids cancellation and produces a more decisive and stable update step.
Figure 4. The diagram illustrates how hyperbolic projection mitigates gradient conflict. (A) In Euclidean Space: Two conflicting gradients, g 1 and g 2 , partially cancel each other out (a ‘‘tug-of-war”), resulting in a small and inefficient update vector, g e u c . This represents destructive interference. (B) In Hyperbolic Space: The same gradients are projected onto the Poincaré disk. The hyperbolic geometry, leveraged by our optimizer, allows for a more effective aggregation. The resulting update, g h y p , mapped back to the parameter space, avoids cancellation and produces a more decisive and stable update step.
Sensors 25 05003 g004
Figure 5. Comparison of free evolution and Berry phase constraint.
Figure 5. Comparison of free evolution and Berry phase constraint.
Sensors 25 05003 g005
Figure 6. Comparison of fusion results using different methods in the MSRS dataset (nighttime).
Figure 6. Comparison of fusion results using different methods in the MSRS dataset (nighttime).
Sensors 25 05003 g006
Figure 7. Comparison of fusion results using different methods in the MSRS dataset (daytime).
Figure 7. Comparison of fusion results using different methods in the MSRS dataset (daytime).
Sensors 25 05003 g007
Figure 8. Comparison of fusion results using different methods in the TNO dataset.
Figure 8. Comparison of fusion results using different methods in the TNO dataset.
Sensors 25 05003 g008
Figure 9. Comparison of fusion results using different methods in the RoadScene dataset.
Figure 9. Comparison of fusion results using different methods in the RoadScene dataset.
Sensors 25 05003 g009
Figure 10. Comparative plots of the SeaFusion model employing different optimizers in the MSRS dataset across the CC, EN, MSE, and Nabf metrics.
Figure 10. Comparative plots of the SeaFusion model employing different optimizers in the MSRS dataset across the CC, EN, MSE, and Nabf metrics.
Sensors 25 05003 g010
Figure 11. Comparative plots of the SeaFusion model employing different optimizers in the TNO dataset across the AG, SF, VIF, and Nabf metrics.
Figure 11. Comparative plots of the SeaFusion model employing different optimizers in the TNO dataset across the AG, SF, VIF, and Nabf metrics.
Sensors 25 05003 g011
Figure 12. Comparative plots of the SeaFusion model employing different optimizers in the RoadScene dataset across the MI, SD, SF, and VIF metrics.
Figure 12. Comparative plots of the SeaFusion model employing different optimizers in the RoadScene dataset across the MI, SD, SF, and VIF metrics.
Sensors 25 05003 g012
Figure 13. Comparison of segmentation results using different methods on the MSRS dataset: (a) Adadelta, (b) AMSGrad, (c) RMSprop, (d) SGD, (e) SeaFusion, and (f) Ours.
Figure 13. Comparison of segmentation results using different methods on the MSRS dataset: (a) Adadelta, (b) AMSGrad, (c) RMSprop, (d) SGD, (e) SeaFusion, and (f) Ours.
Sensors 25 05003 g013
Figure 14. Comparison of segmentation results on the MSRS dataset using various models and optimization methods. The first row shows light images, while the second row displays infrared images. The proposed outperforms other models, including SeaFusion (with AMSGrad, RMSprop, and Adadelta optimizers), GANMcC, U2Fusion, and SDNet, especially in handling complex backgrounds and fine details. Ours achieves superior segmentation accuracy and detail preservation, particularly in infrared images, highlighting its effectiveness in challenging scenarios. The content in the red box is our result display.
Figure 14. Comparison of segmentation results on the MSRS dataset using various models and optimization methods. The first row shows light images, while the second row displays infrared images. The proposed outperforms other models, including SeaFusion (with AMSGrad, RMSprop, and Adadelta optimizers), GANMcC, U2Fusion, and SDNet, especially in handling complex backgrounds and fine details. Ours achieves superior segmentation accuracy and detail preservation, particularly in infrared images, highlighting its effectiveness in challenging scenarios. The content in the red box is our result display.
Sensors 25 05003 g014
Figure 15. Comparison of fusion results in the MSRS dataset. The red box shows an enlarged detail of the image content, and the green box shows the prominent targets in the image. If there is no special explanation in the subsequent figure, it is synonymous with this.
Figure 15. Comparison of fusion results in the MSRS dataset. The red box shows an enlarged detail of the image content, and the green box shows the prominent targets in the image. If there is no special explanation in the subsequent figure, it is synonymous with this.
Sensors 25 05003 g015
Figure 16. Comparison of fusion results of tested solutions in the RoadScene dataset. The red box shows an enlarged detail of the image content, and the green box shows the prominent targets in the image.
Figure 16. Comparison of fusion results of tested solutions in the RoadScene dataset. The red box shows an enlarged detail of the image content, and the green box shows the prominent targets in the image.
Sensors 25 05003 g016
Figure 17. Comparison of fusion results of tested solutions in the TNO Dataset. The red box shows an enlarged detail of the image content, and the green box shows the prominent targets in the image.
Figure 17. Comparison of fusion results of tested solutions in the TNO Dataset. The red box shows an enlarged detail of the image content, and the green box shows the prominent targets in the image.
Sensors 25 05003 g017
Figure 18. Comparative plots across different fusion quality metrics: AG, SD, SF, Nabf, SSIM, MI, Qabf, and VIF.
Figure 18. Comparative plots across different fusion quality metrics: AG, SD, SF, Nabf, SSIM, MI, Qabf, and VIF.
Sensors 25 05003 g018
Table 1. A comparison of the performances of different detection models on various metrics in the MRSR dataset.
Table 1. A comparison of the performances of different detection models on various metrics in the MRSR dataset.
ModelMSESDPSNRAGCCENNabfSSIM
SeaFusion0.03940.2063.893.0770.5406.6590.0140.3104
SGD0.03840.1364.033.0400.6116.7030.0080.4951
RMSprop0.03739.3764.063.0360.5366.6070.0130.3142
Adadelta0.03939.4663.803.0150.5966.6430.0140.3104
AMSGrad0.03840.0263.963.0960.5426.6580.0130.3026
Ours0.03441.4664.493.2860.6746.6750.0030.4781
Table 2. A comparison of the performances of different detection models on various metrics in the TNO dataset.
Table 2. A comparison of the performances of different detection models on various metrics in the TNO dataset.
ModelMSESDVIFAGCCSFNabfSSIM
SeaFusion0.02441.230.6763.8790.33911.170.0430.1489
Adadelta0.02439.870.5894.2430.5358.7970.0360.3802
AMSGrad0.02441.530.6483.8860.40010.610.0410.1272
RMSprop0.01648.800.6783.8230.3998.7460.0400.1759
SGD0.08637.690.7564.1150.6119.6440.0170.4227
Ours0.01941.660.8785.9930.62412.960.0120.4038
Table 3. A comparison of the performances of different detection models on various metrics in the RoadScene dataset.
Table 3. A comparison of the performances of different detection models on various metrics in the RoadScene dataset.
ModelMSESDVIFENCCSF
SeaFusion0.06238.660.4836.8850.33912.97
Adadelta0.05242.180.4967.0120.08412.99
AMSGrad0.06239.110.4676.8940.27512.85
RMSprop0.06338.890.5166.8390.08012.41
SGD0.05145.910.6717.3430.09313.65
Ours0.04847.400.8477.4050.28415.90
Table 4. Segmentation performance (mIoU) of visible, infrared, and fused images in the MSRS dataset. RED indicates the best result, and BLUE represents the second-best result.
Table 4. Segmentation performance (mIoU) of visible, infrared, and fused images in the MSRS dataset. RED indicates the best result, and BLUE represents the second-best result.
MethodBackgroundCarPersonBikeCurveCar StopCuadrailColor ToneBumpmIoU
Visible98.2689.0359.9470.0060.6971.4377.9063.4275.3174.00
Infrared98.2487.3370.4669.2358.7468.8565.5756.9372.7272.01
GANMcC98.4789.2672.1171.7462.7172.9474.0563.2677.4275.77
U2Fusion98.4989.7872.9371.9662.8470.9579.2563.5977.3276.35
SDNet98.5290.4373.4171.6163.6875.5985.2561.8281.4176.11
Adadelta95.4458.0649.5325.699.5526.9245.7135.713.5740.02
AMSGrad98.4690.7266.2468.7556.7373.2483.1466.5580.476.02
RMSprop98.4490.5465.3768.7056.1372.7683.4166.4680.4276.80
SeaFusion98.4590.6265.4768.8156.7973.1682.9966.5680.1175.88
Ours98.6291.0675.6171.9863.9274.8477.8967.3177.1377.59
Table 5. The table presents the experimental results of training using different methods in different datasets.
Table 5. The table presents the experimental results of training using different methods in different datasets.
DatasetMethodSDPSNRMSEMISDCSSIM
TNOSDCFusion45.3561.480.0462.2361.8660.490
TNOOurs145.3761.990.0412.3871.8360.494
TNOOurs245.8262.030.0402.3651.8500.497
RoadSceneSDCFusion44.4758.660.0882.6201.7780.398
RoadSceneOurs147.8758.990.0813.3401.6470.410
RoadSceneOurs250.4358.820.0853.1871.6200.410
MSRSSDCFusion42.3764.370.3484.4421.6650.497
MSRSOurs142.2864.410.3454.8691.6010.493
MSRSOurs242.7964.260.3364.9441.5860.497
Table 6. A comparison of the performances of different detection models on various metrics in the MSRS dataset.
Table 6. A comparison of the performances of different detection models on various metrics in the MSRS dataset.
ModelMSEVIFPSNRAGCCSCDSFNabfalphabetagamma
HC-SPA0.0340.84364.493.2860.6741.7178.7690.0032.00.905.0
HC-SPA10.0390.94263.843.0310.6041.4729.4000.0090.50.902.0
HC-SPA20.0370.92764.082.9870.6061.4909.2680.0091.00.9010.0
HC-SPA30.0370.94564.112.9950.6081.5169.3090.0091.00.9510.0
HC-SPA40.0410.90963.753.0790.5951.3949.5600.0091.00.9020.0
HC-SPA50.0680.02749.830.7790.1310.2491.9750.0162.00.9010.0
HC-SPA60.0420.93463.683.0750.6021.4349.4600.0082.00.9520.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, W.; Fang, A.; Li, Y.; Wei, Y. HC-SPA: Hyperbolic Cosine-Based Symplectic Phase Alignment for Fusion Optimization. Sensors 2025, 25, 5003. https://doi.org/10.3390/s25165003

AMA Style

Zhang W, Fang A, Li Y, Wei Y. HC-SPA: Hyperbolic Cosine-Based Symplectic Phase Alignment for Fusion Optimization. Sensors. 2025; 25(16):5003. https://doi.org/10.3390/s25165003

Chicago/Turabian Style

Zhang, Wenlong, Aiqing Fang, Ying Li, and Yan Wei. 2025. "HC-SPA: Hyperbolic Cosine-Based Symplectic Phase Alignment for Fusion Optimization" Sensors 25, no. 16: 5003. https://doi.org/10.3390/s25165003

APA Style

Zhang, W., Fang, A., Li, Y., & Wei, Y. (2025). HC-SPA: Hyperbolic Cosine-Based Symplectic Phase Alignment for Fusion Optimization. Sensors, 25(16), 5003. https://doi.org/10.3390/s25165003

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop