You are currently viewing a new version of our website. To view the old version click .
Fractal and Fractional
  • Article
  • Open Access

17 November 2025

A Fractional-Order SSIM-Based Gaussian Loss with Long-Range Memory for Dense VSLAM

,
,
,
,
and
1
Laboratory of Intelligent Control, Rocket Force University of Engineering, Xi’an 710025, China
2
The School of Artificial Intelligence, Xidian University, Xi’an 710071, China
*
Authors to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Pattern Recognition—Image and Time Series Analyses—through Fractal Geometry and Complexity Theory

Abstract

In dense visual simultaneous localization and mapping VSLAM (VSLAM), a fundamental challenge lies in the inability of existing loss functions to dynamically balance luminance, contrast, and structural fidelity under photometric variations, while their underlying mechanisms, particularly the conventional Gaussian kernel in SSIM, suffer from limited receptive fields due to rapid exponential decay, preventing the capture of long-range dependencies essential for global consistency. To address this, we propose a fractional Gaussian field (FGF) that synergizes Caputo derivatives with Gaussian weighting, creating a hybrid kernel that couples power-law decay for long-range memory with local smoothness. This foundational kernel serves as the core component of FGF-SSIM, a novel loss function that adaptively recalibrates luminance, contrast, and structure using fractional-order statistics. The proposed FGF-SSIM is further integrated into a complete 3D Gaussian Splatting (3DGS)-based SLAM system, named FGF-SLAM, where it is employed across both tracking and mapping modules to enhance performance. Extensive evaluations demonstrate state-of-the-art performance across multiple benchmarks. Comprehensive analysis confirms the superior long-range dependency of the fractional kernel, dedicated illumination robustness tests validate the enhanced invariance of FGF-SSIM, and quantitative results on TUM and Replica datasets show significant improvements in reconstruction quality and trajectory estimation. Ablation studies further substantiate the contribution of each proposed component.

1. Introduction

Visual simultaneous localization and mapping (VSLAM) constitutes a computational framework that performs simultaneous position estimation and environment mapping by primarily utilizing visual sensors to acquire photometric data and scene geometry []. Owing to its cost-effectiveness and compact form factor, VSLAM has emerged as a critical enabling technology for autonomous positioning and navigation systems []. The choice of map representation remains pivotal in VSLAM, yet prevailing loss functions present a fundamental bottleneck. Their static, local computations are unable to enforce global consistency and lack the adaptability to handle varying illumination.
In contrast to established scene representations (e.g., 2.5D elevation maps [], Signed Distance Fields (SDF) [], and Neural Radiance Fields (NeRF) []), 3D Gaussian Splatting (3DGS) [] has recently demonstrated compelling performance in dense scene reconstruction. While adopting an explicit representation comprised of millions of 3D Gaussians, 3DGS achieves real-time differentiable rendering via splat-based rasterization, yet its optimization relies neither on direct geometric constraints nor on density field regularization as in prior methods. Each primitive is defined by its geometric parameters (e.g., 3D position, scale, and rotation) as well as view-dependent appearance attributes (e.g., color and opacity) []. The loss function of 3DGS is formulated as the color and depth information between the synthesized image and the actual observation, offering a straightforward and intuitive objective. Crucially, because the contribution of each Gaussian to the final rendered image is differentiable and locally confined, gradients can be efficiently propagated back to the corresponding geometric and appearance parameters []. This mechanism ensures a well-balanced optimization, simultaneously enhancing both the photometric accuracy of the rendered novel views and the geometric fidelity of the underlying 3D structure. Regarding the integration of loss functions with 3D Gaussians, numerous researchers have explored this area, yet comprehensive investigations remain limited. Existing efforts primarily revolve around variants of the L1 loss and the Structural Similarity Index (SSIM). The original 3DGS employs a composite “DSSIM + L1” loss function for rendering optimization, while derivative works such as SplaTAM [] and RGBD GS-ICP SLAM [] utilize a similar “SSIM + L1” formulation. Evidently, most methodologies adhere to the loss function framework established in the foundational 3DGS work. However, DSSIM was originally designed for meteorological floating-point data analysis, not robotic vision applications []. This metric neglects critical photometric differences inherent in SLAM environments, particularly dynamic illumination variations. Conversely, some studies omit SSIM for geometric consistency and pose estimation considerations: GS-SLAM [] relies solely on L1 minimization for joint photometric-geometric optimization, whereas MonoGS [] combines L1-based photometric loss with geometric error metrics and uniformity regularization. However, the window-based computation of SSIM renders it ineffective for global optimization, and relying solely on the L1 loss is inadequate for high-fidelity scene reconstruction. Meanwhile, fractional calculus aligns more closely with the physical essence of light transmission, facilitating adaptive regularization of Gaussian primitives and continuous-scale control of input imagery. However, its application in the construction and regulation of 3D Gaussian representations currently requires further development.
The mathematical modeling of anomalous diffusion and long-range dependence has been significantly advanced through the framework of fractional calculus. A unifying perspective is offered by combining spatial and temporal coordinate transformations within this framework. For instance, Soliman et al. [] employ the improved modified extended tanh method (IMETM) to derive exact analytical solutions for a higher-order nonlinear Schrödinger (HNLS) equation incorporating fractional derivatives in both time and space. Similarly, Shao et al. [] investigate approximate solutions for a two-dimensional space/multi-time fractional Bloch–Torrey model involving the Riesz fractional operator. Beyond these, fractional calculus has been extensively applied to telegraph equations [], the dissipative Burgers equation [], and the Korteweg–de Vries equation []. Despite these broad developments, the application of fractional calculus in image processing remains relatively underexplored. In this domain, Caputo fractional derivatives have shown particular promise. Building on the well-established Grünwald–Letnikov and Riemann–Liouville definitions, Pu et al. [] introduced six fractional differential masks and analyzed their efficacy in multiscale texture enhancement. Further extending this line of work, Pu et al. [] proposed a fractional-order variational framework for Retinex, formulated via fractional partial differential equations (FPDEs) to achieve multi-scale nonlocal contrast enhancement while preserving texture. Additional studies [] have confirmed that fractional-order differential operators outperform integer-order models in enhancing high-frequency edge information while retaining low-frequency contour content in smooth regions []. In this context, fractional calculus not only more accurately captures the physical behavior of light transport but also provides a principled methodology for modeling multi-scale and long-memory visual phenomena []. Nevertheless, the potential of fractional operators in constructing and regulating advanced 3D Gaussian representations remains largely untapped, indicating a compelling direction for future methodological innovation.
To address the challenge of uncertainty propagation in visual loss functions induced by photometric variations, we introduce a fractional Gaussian field (FGF) grounded in fractal calculus and propose a novel FGF-SSIM loss function. The core innovation lies in replacing the conventional Gaussian kernel in SSIM with a fractional-order variant that combines power-law decay with Gaussian attenuation, enabling long-range dependency modeling while preserving local smoothness. Specifically, FGF-SSIM adaptively decomposes and recalibrates luminance, contrast, and structure components through fractional-order local statistics, achieving superior illumination robustness. Furthermore, we develop FGF-SLAM, a complete 3DGS-based SLAM framework that integrates FGF-SSIM into both tracking and mapping pipelines, demonstrating that the proposed loss improves both scene reconstruction quality and trajectory estimation accuracy. The main contributions of this work are as follows:
  • To overcome the locality limitation of conventional operators that prevents them from capturing long-range dependencies essential for global consistency, we introduce a novel fractional Gaussian field formulation that integrates Caputo derivatives with Gaussian weighting kernels. This approach addresses the fundamental constraint of local receptive fields in existing methods by establishing a mathematical framework for long-range dependency modeling. The resulting kernel exhibits mathematically provable long-range memory properties while effectively mitigating boundary effects and relaxing isotropic constraints that arise in discrete image processing.
  • To address the static nature of existing loss functions that renders them incapable of adaptively re-calibrating the balance between luminance, contrast, and structural fidelity under varying illumination conditions, we propose FGF-SSIM, a fractional-order loss function that dynamically recalibrates these three fundamental components. Building upon the SSIM framework, FGF-SSIM explicitly decouples and adaptively models luminance, contrast, and structure through fractional-order statistics, achieving superior performance in challenging illumination conditions while preserving structural fidelity for high-quality 3D scene reconstruction.
  • To comprehensively validate the proposed framework and address the lack of global scene memory in conventional SLAM systems, we establish an extensive evaluation protocol that assesses both theoretical properties and practical effectiveness. We demonstrate the superiority of FGF-SSIM through rigorous comparisons under photometric variations and noise corruption, with detailed illumination invariance analysis highlighting its long-range dependency advantages. Furthermore, we integrate the loss into a complete 3DGS-based SLAM system, showing significant improvements in scene reconstruction quality, novel view synthesis fidelity, and trajectory estimation accuracy, particularly in recovering fine texture details and maintaining structural coherence under real-world operating conditions.

2. Preliminary

2.1. SSIM Loss Function

The design of loss functions is fundamentally constrained by the structural properties of the underlying scene representation. For example, the inherently unordered and permutation-invariant structure of point clouds necessitates the use of distribution-based metrics such as Chamfer distance, rather than direct point-to-point correspondences, in the loss formulation []. In VSLAM, modern scene representation methods effectively exploit the spatiotemporal coherence of sequential visual data through GPU-parallelized optimization frameworks. A widely adopted approach involves defining the loss function in image space, typically based on photometric or geometric discrepancies between observed and synthesized views. To prevent cancellation effects between positive and negative differences, the absolute value or square of pixel differences is typically adopted. Summing the absolute differences yields the L1 loss function [], which is computationally simpler and exhibits enhanced robustness to outliers in regression tasks. Conversely, summing the squared differences produces the L2 loss function []. In early neural network research, L2 loss was frequently selected as the default due to its positive definiteness, chi-square properties, and encapsulation in standard frameworks (e.g., Caffe). However, both L1 and L2 losses assume ideal Gaussian white noise characteristics and presume statistical independence between noise and local image features, neglecting human visual system (HVS) perception. To address this limitation, SSIM-based loss functions [] incorporate HVS characteristics by evaluating luminance, contrast, and structural similarities. Unlike pixel-wise methods, SSIM operates via sliding windows for computational efficiency. The Joint Video Team (JVT) formally integrated SSIM into the H.264/MPEG-4 Advanced Video Coding (AVC) standard in 2003, which is extensively deployed in video surveillance and digital broadcasting.
SSIM evaluates the similarity between two images x and y by comparing three components: luminance, contrast, and structure. The comparison is performed within local windows and the results are aggregated across the image. The functional expression of SSIM is:
S ( x , y ) = f ( l ( x , y ) , c ( x , y ) , s ( x , y ) ) = ( l ( x , y ) ) ϵ · ( c ( x , y ) ) ε · ( s ( x , y ) ) ζ ,
l ( x , y ) = 2 μ x μ y + C 1 μ x 2 + μ y 2 + C 1 , c ( x , y ) = 2 σ x σ y + C 2 σ x 2 + σ y 2 + C 2 , s ( x , y ) = σ x y + C 3 σ x σ y + C 3 .
where l ( x , y ) , c ( x , y ) , s ( x , y ) represent the luminance, contrast, and structure comparison functions respectively; ϵ , ε , ζ denote the proportional weights of each component; μ x and μ y are the local means of x and y; σ x and σ y represent the local standard deviations; and σ x y is the local covariance between x and y.
To ensure numerical stability when the comparison functions approach zero, constants ( C 1 , C 2 , C 3 ) are added to both numerator and denominator. These constants are defined as C 1 = ( k 1 L ) 2 , C 2 = ( k 2 L ) 2 , where k 1 and k 2 are constants much smaller than 1, and L is the dynamic range of the image pixel values. For practical computation, the parameters are typically set to ϵ = ε = ζ = 1 , C 3 = 0.5 C 2 , k 1 = 0.01 , k 2 = 0.03 . Therefore, the simplified SSIM expression becomes:
S ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
For pixel-level computation, let p represent the pixel index within a local region P. Define l ( p ) = 2 μ x μ y + C 1 μ x 2 + μ y 2 + C 1 and c s ( p ) = 2 σ x y + C 2 σ x 2 + σ y 2 + C 2 . The pixel-wise SSIM can then be expressed as:
S S I M ( p ) = l ( p ) · c s ( p )
In subsequent research, the MS-SSIM loss function [] incorporates a multi-scale principle, effectively integrating resolution considerations into the loss function construction. Both G-SSIM [] and MS-G-SSIM [] enhance the robustness of the loss function for image processing under blur and noise degradation. IW-SSIM [] focuses on the quality fusion stage within the full-reference image quality assessment (IQA) framework, proposing a multi-scale information-content-weighted quality evaluation method. However, existing improvements predominantly target structural distortions such as blurring and motion artifacts, while three critical issues remain understudied []. Firstly, the existing loss function for scene reconstruction using illumination-sensitive imagery remains insufficiently explored theoretically. Secondly, the existing loss function formulations are often highly scenario-specific, with limited research devoted to the development of generalizable formulations. Thirdly, the existing loss function fails to adequately account for long-range dependencies within the image and lacks a mechanistic analysis of its operation.

2.2. Fractional-Order Calculus

Let the function f ( t ) be continuously differentiable on the interval [ a , t ] , then its α -order Caputo definition can be presented as:
D t α C f ( t ) = 1 Γ ( n α ) a t f ( n ) ( τ ) ( t τ ) α n + 1 d τ
where D t α C represents the fractional operator in the Caputo definition, n = α , and Γ ( · ) is the gamma function. In image processing, we usually deal with spatial domains, and α ( 0 , 1 ) , so n = 1.
The fractional derivative of the image at pixel p (with an order of α ) is denoted as:
D α   f ( p ) = 1 Γ ( 1 α ) 0 p f ( t ) ( p t ) α d t
We adopt the Formula (6) to simplify the definition of Caputo fractional derivatives to better adapt to handling discrete two-dimensional pixels. Meanwhile, we did not directly use the function value difference, because in the local window P of image processing, the Caputo derivative does not require the pixel values outside the window, and the image boundary conditions can be handled more naturally through the integer order derivative.

2.3. Processing of Discrete Image Data

In the local window of image processing, we usually assume that the center of the window is the focus of attention, while the influence of the boundary is relatively small. When defining local statistics, to facilitate calculation, we use Gaussian weight kernels. We proposed expressing the Caputo derivative as an integer-order derivative convolution with a kernel function, and made this form compatible with the convolution operation of the Gaussian kernel, facilitating the unification of the framework:
D α x ( p ) = k Ω ω k ( α ) G σ ( k ) · x ( p k )
where Ω is the neighborhood of the convolution kernel, ω k ( α ) = ( 1 ) k α k is the coefficient of the Grunwald-Letnikov fractional derivative. G σ ( k ) is the Gaussian weight kernel, and that is G σ ( k ) = exp k 2 2 σ 2 . It is worth noting that here we actually approximate the Caputo derivative on a discrete grid and add Gaussian weights to achieve localization.
However, in the local window P, directly using Caputo’s discrete approximation requires pixel values outside the window (historical or future values). At the image boundaries, these values do not exist, resulting in severe boundary effects. Meanwhile, image processing usually requires isotropic operations. The standard fractional difference is directionally dependent (for example, typically defined in the x or y direction), which does not meet the requirement of rotational invariance. Therefore, a novel form of Caputo derivative is needed to process discrete image data.

3. Methods

3.1. Fractional Gaussian Field

The SSIM framework and its derivatives (e.g., MS-SSIM [], G-SSIM []) fundamentally rely on a conventional Gaussian kernel for computing local statistics. Specifically, the local mean μ x , standard deviation σ x , and covariance σ x y defined in the SSIM formulation are computed using a Gaussian weighting function with a fixed spatial support, normalized to unit sum. This design provides rotational invariance and mitigates blocking artifacts. However, the kernel’s rapid exponential decay [] is the source of its fundamental limitations. It strictly confines the effective receptive field, making it inherently a local operator that is myopic to long-range dependencies. Its static nature with a fixed scale prevents it from adapting to non-stationary statistics under varying illumination, and this sharp attenuation profile inherently amplifies boundary effects. These intrinsic properties of the conventional Gaussian kernel directly crystallize the three shortcomings outlined previously [], including the inability to model global context, the infeasibility of adaptive re-balancing, and unreliable statistics under photometric variations. Therefore, the quest for a more robust loss function necessitates a fundamental evolution beyond the conventional Gaussian kernel.
To solve the above problems, we propose a novel fractional Gaussian field method. We combine the idea of Caputo derivatives with Gaussian weight kernels to overcome the boundary effects and isotropic requirements in discrete image processing. The fractional Gaussian field kernel function G ( α ) [ m , n ] can be obtained:
G ( α ) [ m , n ] = 1 Γ ( 1 α ) 1 ( m 2 + n 2 ) α / 2 exp m 2 + n 2 2 σ 2
where α represents the fractional order, controlling the degree of power-law attenuation. When α = 0 , this kernel degenerates into a Gaussian kernel. Γ ( 1 α ) represents the Gamma function as the fractional-order normalization factor. σ is the standard deviation of the Gaussian term, which controls the intensity of Gaussian attenuation. ( m , n ) represents the integer displacement coordinates of pixel p relative to the central pixel p ˜ .
When calculating fractional-order local statistics, we normalize the convolution kernel in a discrete case to maintain numerical stability. We define the convolution kernel ω m , n ( α ) used to compute fractional-order local statistics after normalization as follows:
ω m , n ( α ) = 1 Γ ( 1 α ) G σ m 2 + n 2 m 2 + n 2 α = 1 Γ ( 1 α ) G ( α ) [ m , n ]
The proposed fractional Gaussian field establishes a new computational paradigm for visual data analysis, wherein the presented fractional Gaussian kernel serves as the fundamental building block. This kernel’s hybrid structure, combining a power-law component for long-range dependence with a Gaussian term for local smoothness, makes significant contributions to addressing critical challenges in VSLAM. In practical image analysis, this translates to a superior ability to model global illumination changes and maintain structural coherence across the scene, overcoming the myopia of conventional local operators. For model-prediction assessment within SLAM systems, the kernel provides a more physically-grounded similarity metric that reflects the non-local interactions inherent in light transport, leading to more reliable evaluation of reconstruction quality under photometric variations. Beyond VSLAM, this kernel holds promise for a broader class of image processing tasks requiring multi-scale analysis and robustness to non-stationary statistics, such as medical image enhancement, remote sensing image fusion, and high dynamic range (HDR) imaging, where modeling complex, long-range pixel relationships is crucial. The foundational advancement of the fractional Gaussian field, operationalized through its kernel, thus creates a natural and powerful substrate for the subsequent development of our FGF-SSIM loss function, which leverages these properties to achieve a dynamic balance between photometric invariance and structural preservation.

3.2. FGF-SSIM

For the image window P centered on pixel p ˜ , we use the fractional Gaussian field convolution kernel to calculate the local statistics. In our formulation, the local statistics are computed within a sliding window P across the entire images x and y. The pixel coordinates [ i , j ] specify the center of each local window, and the displacements [ m , n ] are used to index relative positions within that window. This approach ensures that all statistical measures are derived from local neighborhoods, distinguishing them from global image properties. Also, since the weights ω m , n ( α ) already include the normalization factor 1 Γ ( 1 α ) , there is no need to additionally divide by the window size. The fractional local mean at pixel location [ i , j ] is defined as:
μ x ( α ) [ i , j ] = m = M M n = N N ω m , n ( α ) · x [ i m , j n ]
Similarly, we can obtain that the fractional local variance is defined as:
( σ x ( α ) ) 2 [ i , j ] = m = M M n = N N ω m , n ( α ) · x [ i m , j n ] μ x ( α ) [ i , j ] 2
The fractional local covariances of the two images x and y are defined as:
σ x y ( α ) [ i , j ] = m = M M n = N N ω m , n ( α ) · x [ i m , j n ] μ x ( α ) [ i , j ] y [ i m , j n ] μ y ( α ) [ i , j ]
In the construction of the FGF-SSIM loss function, we retained the original formula of the SSIM loss function (Formula (2)). However, to simplify the calculation, we assume that ε = ζ = 1 , but ϵ 1 . Substitute the fractional statistics into the SSIM framework while retaining the luminance separation feature:
L ( α ) ( x , y ) = 2 μ x ( α ) μ y ( α ) + C 1 ( μ x ( α ) ) 2 + ( μ y ( α ) ) 2 + C 1
Contrast-Structure function:
CS ( β ) ( x , y ) = σ x y ( β ) + C 2 σ x ( β ) σ y ( β ) + C 2
Therefore, we can obtain the expression of the core formula of FGF-SSIM:
FGF - SSIM ( x , y ) = L ( α ) ( x , y ) ϵ · CS ( β ) ( x , y )

3.3. Backpropagation Formula

In order to use FGF-SSIM as the loss function in the process of restoring the three-dimensional scene from the image, we need to derive its gradient. Let L = 1 FGF - SSIM , total loss can be defined:
L total = λ 1 1 + λ 2 L
In backpropagation, we need to compute the gradient of the total loss to the input image x:
L total x = λ 1 1 x + λ 2 L x = λ 1 1 x λ 2 FGF - SSIM x
We compute the L1 gradient term by 1 x = sign ( x y ) . Compute the FGF-SSIM gradient term by decomposing into luminance components and contrast-structure components via the chain rule. Let S = FGF - SSIM = [ S L ] γ · [ S C ] δ , S L = L ( α ) , S C = CS ( β ) . This can be obtained by applying the product rule and the chain rule:
S x = S γ L ( α ) L ( α ) x + δ CS ( β ) CS ( β ) x
We then used the quotient rule and the fractional mean chain rule to solve for the luminance component gradient, and the quotient rule and the fractional covariance chain rule to solve for the contrast-structure component gradient.
L ( α ) x = m , n ω m , n ( α ) · 2 ( μ y ( α ) y μ x ( α ) x ) ( μ x ( α ) ) 2 + ( μ y ( α ) ) 2 + C 1
CS ( β ) x = m , n ω m , n ( β ) · ( y μ y ( β ) ) CS ( β ) σ y ( β ) σ x ( β ) ( x μ x ( β ) ) σ x ( β ) σ y ( β ) + C 2
Through the above derivation, the total gradient of FGF-SSIM can be expressed as:
L total x = λ 1 sign ( x y ) λ 2 · FGF - SSIM γ L ( α ) L ( α ) x + δ CS ( β ) CS ( β ) x
The gradient computation involves carefully tuned weighting parameters. Specifically, λ 1 and λ 2 balance the contributions of the L1 and FGF-SSIM components in the total loss, while γ and δ control the relative importance of luminance versus contrast-structure preservation. These parameters were optimized through multi-objective optimization to ensure stable convergence and were fixed at λ rgb = 1.0 , λ depth = 0.8 , and λ iso = 0.2 for all experiments.

4. 3D Gaussian SLAM with Fractional Gaussian Field Structural Similarity

4.1. FGF-SLAM

Figure 1 illustrates the comprehensive architecture of the proposed FGF-SLAM pipeline. Our framework aims to achieve simultaneous high-fidelity novel view synthesis and dense geometric reconstruction, enabling robust scene representation and real-time rendering capabilities. FGF-SLAM supports both RGB-D and monocular input data streams.
Figure 1. The comprehensive architecture of the proposed SLAM pipeline of FGF-SLAM.
The processing pipeline begins with the 3D Gaussian Scene Representation module, where input data undergoes scene initialization and is transformed into a collection of anisotropic 3D Gaussians through a structured parameterization process. Through differentiable projective transformation, each 3D Gaussian is projected to a 2D Gaussian on the image plane for efficient rasterization. These Gaussian primitives collectively encode both geometric and photometric properties of the environment. Simultaneously, the Camera Pose Tracking module coordinates the relationship between pose estimation and rendering optimization. This module employs a combined ‘L1+FGF-SSIM’ loss formulation for iterative camera pose refinement, enabling accurate online pose estimation for sequential frames. The Keyframe Management module subsequently performs keyframe retrieval, selection, and pruning based on pose constraints and Gaussian visibility metrics, significantly enhancing the computational efficiency and reconstruction quality of the SLAM system. In the Mapping and Optimization module, FGF-SLAM optimizes Gaussian insertion and parameter updates using refined pose information from preceding stages. The optimization incorporates isotropic regularization to prevent excessive anisotropic deformation in under-observed regions, while utilizing a ‘L1+FGF-SSIM’ strategy to construct an analogous loss function for mapping. The FGF-SSIM component employs adaptive fractional orders that are dynamically adjusted via ATE monitoring, ensuring robustness across varying illumination conditions while preserving structural consistency and contrast sensitivity.

4.2. 3D Gaussian Scene Representation

To establish a dense 3DGS-based VSLAM system capable of high-fidelity geometric reconstruction, we formulate the 3D scene representation as an adaptively controlled parametric model comprising anisotropic 3D Gaussian primitives:
G = G i ( μ i , Σ i , α i ) | i = 1 , , N
where each 3D Gaussian primitive G i is defined by position μ i R 3 in the world coordinate system, 3D covariance matrix Σ i R 3 × 3 representing anisotropic geometry, and opacity α i [ 0 , 1 ] controlling transparency. For numerical stability and optimization efficiency, the covariance matrix is factorized as Σ i = R i S i S i R i where R i S O ( 3 ) represents the rotation matrix, S i = diag ( s x , s y , s z ) represents the scaling matrix, and the factorization enables efficient gradient-based optimization through PyTorch (version 1.12.1) automatic differentiation framework.
When transforming from world coordinates to camera coordinates, the 3D Gaussian G ( μ , Σ ) projects to a 2D Gaussian G ( μ , Σ ) on the image plane through differentiable projective transformation:
Σ = J W Σ W J
where J denotes the Jacobian matrix of the projective transformation, W represents the view transformation matrix, Σ is the original 3D covariance matrix in world coordinates, and Σ is the projected 2D covariance matrix on the image plane. This formulation enables efficient rasterization by skipping blank regions while maintaining differentiability throughout the rendering pipeline.
In our experiments, Gaussian densification triggers every 100 iterations with gradient threshold 0.0002, and opacity resets occur at iterations 300 and 1500. The system maintains up to 100,000 Gaussians through visibility-based pruning using the occ_aware_visibility statistics.

4.3. Camera Pose Tracking

The tracking module estimates camera pose T S E ( 3 ) for each incoming frame while keeping the Gaussian parameters frozen. We formulate the optimization problem as a composite loss function that integrates FGF-SSIM for illumination-robust performance:
L track = λ rgb L rgb + λ depth L depth
where L rgb represents the photometric loss component, L depth represents the geometric loss component when depth information is available, λ rgb and λ depth are weighting coefficients that balance the contributions of each loss term. This separation allows dedicated computational resources for camera motion estimation while avoiding environment model augmentation during tracking phases.
For monocular scenarios, we introduce the FGF-SSIM loss that combines traditional L1-norm with fractional-order structural similarity:
L rgb = 1 | M | p M λ L 1 I ^ ( p ) I ( p ) 1 + λ FGF - SSIM ( 1 FGF - SSIM ( I ^ ( p ) , I ( p ) ) )
where I ^ denotes the rendered image, I represents the ground truth image, M is the valid RGB mask region, p represents pixel coordinates, λ L 1 and λ FGF - SSIM are weighting parameters for the L1 and FGF-SSIM components respectively. The FGF-SSIM metric provides superior edge preservation and multi-scale analysis capabilities critical for SLAM accuracy.
When depth information is available, we incorporate depth residuals to further constrain the camera pose optimization:
L depth = 1 | M d | p M d D ^ ( p ) D ( p ) 1
where D ^ denotes the rendered depth map, D represents the ground truth depth map, M d denotes the valid depth mask region, and p represents pixel coordinates. This combined approach leverages both photometric and geometric constraints for robust pose estimation.
In our experiments, the tracking optimization uses rotation learning rate 1 × 10 3 and translation learning rate 2 × 10 4 , with early stopping at rotation delta < 0.001 rad and translation delta < 0.001 m. The loss combines RGB ( λ rgb = 1.0 ) and depth ( λ depth = 0.8 ) components, and FGF-SSIM weight starts at 0.1 with adaptive adjustment.

4.4. Keyframe Management

Keyframe selection employs strategic covisibility analysis of Gaussian distributions to maximize information diversity while minimizing computational overhead. The visibility of Gaussian G i during rasterization is quantitatively evaluated through:
V ( G i ) = I ray α i N ( x ; μ i , Σ i ) d x > τ v
where I ( · ) denotes the indicator function, α i represents the opacity of Gaussian G i , N ( x ; μ i , Σ i ) is the Gaussian probability density function, the integral represents ray accumulation through the Gaussian primitive, and τ v is a visibility threshold. This visibility assessment serves as preprocessing step prior to keyframe management.
Keyframe selection employs rigorous covisibility analysis between current frame m and latest keyframe n:
IoU ( V m , V n ) = | V m V n | | V m V n | < τ IoU
where V m and V n represent the visibility sets of frames m and n respectively, | · | denotes set cardinality, and τ IoU is the intersection-over-union threshold. Additionally, relative motion constraints ensure sufficient baseline for triangulation:
t m n > max ( τ t , η · median ( D m ) )
where t m n is the relative translation vector between frames m and n, τ t is the absolute translation threshold, η is a scaling factor, and median ( D m ) represents the median depth value in frame m.
In our experiments, keyframe selection uses translation threshold 0.12× median depth, rotation threshold 0.1 rad, and overlap ratio cutoff 0.8. The system maintains 2 protected keyframes and a window size of 10, with covisibility cutoff 0.4 for pruning.

4.5. Mapping and Optimization

To prevent excessive anisotropic deformation of Gaussians in under-observed regions, we enforce isotropic constraints through a dedicated regularization term:
L iso = 1 N i = 1 N s i s ¯ i 1 2 2
where s i = [ s x , s y , s z ] represents the scaling parameters of Gaussian G i , s ¯ i = 1 3 j = 1 3 s i , j is the mean scaling value, 1 = [ 1 , 1 , 1 ] is the unit vector, and N is the total number of Gaussians. This regularization mechanism effectively constrains spatial expansion of Gaussian distributions.
The complete mapping objective integrates FGF-SSIM for superior structural preservation:
L map = λ rgb L rgb + λ depth L depth + λ iso L iso
where λ rgb , λ depth , λ iso , and λ geom are weighting coefficients for the photometric, depth, and isotropic regularization respectively. The FGF-SSIM component employs adaptive fractional orders that are dynamically adjusted based on ATE monitoring.
For efficient optimization, we derive the gradients of the FGF-SSIM loss with respect to the rendered image. The fractional orders α and β in FGF-SSIM are dynamically adjusted based on ATE monitoring to maintain robustness:
L total x = λ 1 sign ( x y ) λ 2 · FGF - SSIM γ L ( α ) L ( α ) x + δ CS ( β ) CS ( β ) x
where λ 1 and λ 2 are weighting parameters, sign ( · ) is the sign function, γ and δ control the relative contributions of the luminance and contrast-structure components, and the partial derivatives are computed using the fractional Gaussian field kernel and chain rule differentiation. The ATE monitoring mechanism adjusts α and β to optimize performance under varying conditions.
Gaussian parameters are optimized through gradient descent with type-specific learning rates:
θ ( t + 1 ) = θ ( t ) η θ L map
where θ represents the Gaussian parameters being optimized, t denotes the iteration step, η is the learning rate, and θ L map is the gradient of the mapping loss with respect to the parameters. The learning rates are carefully tuned per parameter type to ensure stable convergence while preserving geometric fidelity.
In our experiments, initial mapping uses 2000 iterations and subsequent keyframes use 300 iterations. Isotropic regularization weight is 10.0, and the ATE monitor uses a 15-frame history with degradation threshold 1.15. Gaussian updates occur every 100 iterations (offset 50) with densification threshold 0.01 and opacity pruning threshold 0.005.
Through the synergistic integration of these components, FGF-SLAM achieves efficient and high-quality scene reconstruction and rendering performance in challenging illumination environments. This architectural framework provides beneficial insights for enhancing the robustness and practical deployment of VSLAM technology in real-world intelligent transportation systems.

5. Experimental Results

5.1. Experimental Setup

All experiments were conducted on a workstation equipped with an Intel® Core™ i7-13700F processor (16-core, 8 performance-cores and 8 efficient-cores, 5.20 GHz) and an NVIDIA GeForce RTX 4090 GPU (16,384 CUDA cores, 24 GB GDDR6X). Algorithm validation was performed on three established benchmarks: the synthetic Replica dataset [], real-world TUM RGB-D dataset [], and M2UD benchmark []. For quantitative evaluation, we adopted three standardized metrics: Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index Measure (SSIM), and Learned Perceptual Image Patch Similarity (LPIPS) for rendering quality assessment, complemented by Absolute Trajectory Error Root Mean Square (ATE RMSE) for pose estimation accuracy.
To ensure optimal performance across different experimental conditions, we implemented a systematic parameter optimization strategy for the fractal Gaussian kernel and FGF-SSIM framework. The key parameters were optimized through grid search and cross-validation on a held-out validation set. The fractional order parameters α and β , which control the long-range correlation and structure preservation respectively, were tuned over the ranges [0.2, 0.9] and [0.1, 0.8] with step sizes of 0.05. The optimal configuration was determined to be α = 0.8 and β = 0.4 , which balanced the trade-off between illumination robustness and structural detail preservation. For the Gaussian kernel parameters, we employed a kernel size of 3 × 3 with adaptive standard deviation σ computed using the scaling relation σ = size · ( 0.3 + 0.25 α + 0.1 β ) that incorporates the fractional orders. For illumination robustness testing, we standardized the perturbation parameters to brightness change = 30, contrast change = 1.1, and gamma change = 1.6, representing challenging real-world lighting variations. All optimization procedures employed the Adam optimizer with an initial learning rate of 5 × 10 4 , decaying by a factor of 0.8 every 50 epochs, and convergence was determined when the relative improvement in validation loss fell below 0.1% for 5 consecutive epochs.
We incorporate the FGF-SSIM loss function into the MonoGS-SLAM baseline, resulting in an enhanced framework termed FGF-SLAM, to assess the practical efficacy of the fractal Gaussian field representation within a full SLAM framework. The 3D Gaussian Scene Representation module employs a composite loss with progressive FGF-SSIM weighting ( w fgf = min ( 0.7 , 0.1 + 0.6 × iteration / 3000 ) ), visibility-aware pruning (threshold = 3), and periodic opacity reset (every 1000 iterations). The Camera Pose Tracking module employs AdamW optimizer with weight decay 1 × 10 6 . Tracking iterations utilize conservative FGF-SSIM parameters ( α = 0.3 , kernel size = 5) with early stopping based on loss stagnation detection (patience=10 iterations, relative tolerance = 1%). The Keyframe Management module selects frames based on translation (> 0.12 × median depth), rotation (>0.1 rad), and overlap criteria (<0.8), while protecting two keyframes and pruning based on 0.4 covisibility cutoff. The Mapping and Optimization module utilizes aggressive FGF-SSIM parameters ( α = 0.5 , kernel size = 7) with ATE-aware adaptive control. The ATEMonitor employs a 15-frame sliding window and degradation threshold of 1.15, automatically reducing FGF-SSIM weight by 30% when performance decline is detected for two consecutive evaluations.
Through systematic optimization and ablation studies, we will demonstrate that the proposed loss function consistently enhances performance, confirming the advantages of incorporating fractal Gaussian priors in improving both robustness and accuracy in visual SLAM systems. Comparative analysis encompassed NeRF-based SLAM methods (e.g., NICE-SLAM [], Vox-Fusion [], Point-SLAM []) and 3DGS-based SLAM approaches (e.g., GS-SLAM [], SplaTAM [], MonoGS []) on their respective compatible datasets. Specifically, to evaluate performance under monocular RGB input constraints, we further compared against ORB-SLAM2 [] (feature-based SLAM), MonoGS [] (monocular Gaussian Splatting SLAM), and MotionGS [] (explicit motion-guided deformation) on the TUM dataset. All comparisons were based on publicly reported performance metrics for these methods.

5.2. Gaussian vs. Fractional Kernels

Figure 2 presents a comparative analysis of the kernel functions employed in SSIM and FGF-SSIM, providing a visual illustration of the mathematical properties inherent to the fractional Gaussian field kernel. Figure 2a displays a two-dimensional heatmap visualization of the standard Gaussian kernel, which exhibits the characteristic profile of a high central weight with rapid exponential decay from the center. This indicates that during image processing, the kernel primarily incorporates information from immediately adjacent pixels, reflecting its limited receptive field. Figure 2b shows the corresponding heatmap of the fractional Gaussian field kernel. While high weights are also concentrated near the center, the kernel exhibits broader and more extensive weight distribution extending outward. It is worth noting that the redistribution of weight toward longer distances in the fractional kernel comes at the cost of slightly reduced emphasis on immediate neighbors compared to the conventional Gaussian kernel. Even at considerable distances from the center, the fractional kernel maintains non-negligible weight values—unlike the conventional Gaussian kernel, which attenuates rapidly to near-zero values. Figure 2c shows the difference map obtained by subtracting the Gaussian kernel from the FGF kernel, highlighting distinctions in their spatial weight distributions. A blue-toned central region suggests a weight advantage of the Gaussian kernel near the origin. This indicates that while the fractional kernel excels at capturing long-range dependencies, it exhibits a minor compromise in local sensitivity compared to the conventional Gaussian formulation. A neutral white intermediate ring indicates comparable magnitudes in this transitional region. Most notably, a pronounced red outer region reveals significantly higher weights in the FGF kernel at larger distances, providing direct visual evidence of its long-range memory characteristics.
Figure 2. Comparative visualization of the standard Gaussian and fractional Gaussian field kernels.
Figure 3 presents a comparative analysis of the radial weight distribution between the Gaussian kernel and the FGF kernel, highlighting quantitative differences in their spatial weighting characteristics. Figure 3a shows one-dimensional profiles extracted from the center of each kernel, plotting weight value as a function of radial distance. The Gaussian kernel exhibits a steep exponential decay, with values declining rapidly from the center. In the immediate vicinity (0–2 pixels), the conventional Gaussian kernel maintains higher weight values, providing superior emphasis on local pixel relationships. In contrast, the FGF kernel demonstrates a more gradual attenuation, maintaining substantially higher weights at medium to large distances. This behavior confirms the stronger long-range memory characteristics of the FGF kernel. Figure 3b displays the same data with weight values converted to a logarithmic scale. In this representation, the Gaussian kernel closely follows an approximately straight line, consistent with its exponential nature. The FGF kernel, by contrast, exhibits a slowly curving trajectory indicative of power-law decay. This fundamental trade-off between local precision and global context capture represents a deliberate design choice in the fractional Gaussian formulation, where the benefits of extended receptive fields outweigh the minor reduction in local emphasis for VSLAM applications. This slower attenuation rate, a defining feature of heavy-tailed distributions, enables the FGF kernel to integrate information from longer distances, thereby facilitating enhanced structural preservation in image processing.
Figure 3. Comparative radial weight analysis of Gaussian and FGF kernels in linear and logarithmic scales.
The fundamental distinction between these kernels lies in their mathematical formulation and consequent performance characteristics. The standard Gaussian kernel exhibits exponential decay, inherently limiting its receptive field and making it susceptible to local illumination variations. In contrast, the fractional Gaussian kernel combines power-law decay with Gaussian attenuation, enabling capture of both local details and global context. This hybrid structure directly addresses core limitations of conventional SSIM in VSLAM applications. The power-law component provides long-range dependencies necessary for modeling global illumination changes, while the Gaussian term maintains local smoothness. The practical implication is that FGF-SSIM maintains structural consistency across larger spatial extents, making it particularly effective for SLAM applications with significant viewpoint changes and illumination variations.
More precisely, we use a Gaussian kernel in the luminance component, σ α = σ 0 · α . In this way, when α increases, the variance of the Gaussian kernel increases, which means greater smoothing and greater robustness to changes in illumination. When α decreases, the variance of the Gaussian kernel decreases and it becomes more sensitive to changes in illumination. Within the local window, when the image satisfies stationarity (i.e., the local mean and variance remain constant), the variation of the fractional-order statistics is mainly affected by the order α . We prove that when α approaches 1, the derivative of the fractional statistic with respect to α is approximately constant within the local window.
Consider the local mean of the luminance component:
μ x ( α ) = w p ( α ) x ( q ) w p ( α )
When α varies around 1, we can Taylor expand μ x ( α ) at α = 1 :
μ x ( α ) μ x ( 1 ) + ( α 1 ) · d μ x ( α ) d α α = 1
Among them, the derivative:
d μ x ( α ) d α = d w p ( α ) d α · x ( p ) · w p ( α ) w p ( α ) x ( p ) · d w p ( α ) d α w p ( α ) 2
In the local stationary region, x ( p ) is approximately the constant c, then:
d μ x ( α ) d α d w p ( α ) d α · c · w p ( α ) c · w p ( α ) · d w p ( α ) d α w p ( α ) 2 = 0
Therefore, in the stationary region, the derivative of μ x ( α ) with respect to α is 0 when α = 1 . Similarly, variance and covariance also have similar properties. However, in non-stationary regions (such as edges), the derivative is not zero. However, in SLAM, the local areas between consecutive frames usually satisfy a gentle change in lighting, so we can consider that the derivative does not change much, that is, it can be approximately regarded as a constant. Therefore, in backpropagation, we can treat the fractional orders α and β as constants (or through adaptive learning, but with a smaller learning rate set), which will not lead to unstable training.

5.3. Statistical Verification of FGF-SSIM

Figure 4 presents a normality assessment of Structural Similarity (SSIM) index distributions across five image denoising methodologies. The upper row displays quantile-quantile (Q–Q) plots, while the lower row shows corresponding histograms with fitted normal distributions, collectively characterizing the statistical distribution of each method. Although all Shapiro–Wilk test results (L2: p = 0.0002 ; L1: p = 0.0004 ; SSIM: p = 3.65 × 10 10 ; MS-SSIM: p = 0.0002 ; FGF-SSIM: p = 0.0134 ) indicate significant deviations from normality ( α = 0.05 ), the FGF-SSIM method exhibits notably superior distributional characteristics. Its Q–Q plot demonstrates the closest alignment with theoretical normality benchmarks, and the histogram confirms the most symmetric and well-shaped distribution profile among all evaluated methods. Although FGF-SSIM does not achieve complete normality, this distributional advantage provides critical justification for employing distribution-free statistical tests in subsequent analyses. Furthermore, these foundational distributional properties establish the necessary context for interpreting differential effect sizes in methodological comparisons, suggesting that FGF-SSIM’s enhanced distributional stability may contribute to more consistent performance advantages.
Figure 4. Assessment of Distribution Normality for SSIM Scores Across Loss Functions.
Figure 5a presents the results of Wilcoxon signed-rank tests comparing the proposed FGF-SSIM method against four established denoising approaches. The bar chart summarizes the statistical significance of pairwise comparisons, with all p-values for FGF-SSIM versus L2, L1, SSIM, and MS-SSIM falling below 10 10 . This uniformly high level of statistical significance across all comparisons provides robust non-parametric evidence of systematic performance differences. The use of this distribution-free test is directly justified by the non-normal distribution characteristics established in Figure 4, ensuring valid statistical inference despite violations of parametric assumptions. The consistently minimal p-values anticipate the substantial effect sizes quantified later, particularly in comparisons against L1 and MS-SSIM, where the fractional Gaussian field kernel demonstrates superior structural preservation. These statistically significant outcomes are consistent with the methodological strengths of FGF-SSIM, which integrates fractional-order denoising, large-kernel filtering for long-range dependencies, and multi-feature fusion for illumination robustness. The agreement between the distributional characteristics established in Figure 4 and the non-parametric significance testing in Figure 5a provides strong evidence for the method’s consistent performance. This rigorous statistical foundation, which confirms a systematic advantage, is further complemented by an analysis of the practical effect magnitudes in the following section.
Figure 5. Wilcoxon Test Significance and Effect Size Analysis of Loss Functions.
Figure 5b quantifies the practical significance of performance differences through Cohen’s *d* effect size analysis, complementing the non-parametric statistical significance established in Figure 5a. The visualization presents effect sizes for all paired comparisons between FGF-SSIM and conventional methods, explicitly classified into small, medium, and large categories. The results indicate large effect sizes for FGF-SSIM compared to L1 and MS-SSIM, a medium effect compared to L2, and a small effect compared to SSIM. This small effect size indicates that while FGF-SSIM achieves statistically significant gains, its average performance advantage over SSIM in terms of the evaluated metrics is less pronounced than against other baselines. Nonetheless, FGF-SSIM demonstrates superior capabilities in specific aspects critical to SLAM, such as illumination robustness and geometric consistency, as detailed in Section 5.3 and Section 5.4. These differential effect sizes align with the theoretical advancements of FGF-SSIM, including the fractional Gaussian field kernel that couples power-law decay for long-range memory with local smoothness. In summary, this effect size analysis substantiates that FGF-SSIM provides consistent and meaningful improvements, with its most substantial gains realized in scenarios requiring robustness to photometric variations and the modeling of global context.

5.4. Loss Functions’ Comparison Results

Figure 6 compares the illumination robustness of FGF-SSIM with other loss functions on the urban01 scenario of the M2UD dataset []. Under consistent experimental conditions, we provide qualitative comparisons by presenting the resulting output images generated by each loss function. Additionally, we evaluated the proposed FGF-SSIM loss function against other loss functions using three standard image quality metrics: PSNR, SSIM, and MSE. After increasing the original image’s illumination intensity by 30%, the proposed FGF-SSIM loss preserves structural details more effectively than competing loss functions and achieves superior performance across multiple evaluation metrics.The results demonstrate that incorporating fractional-order calculus into SSIM significantly enhances algorithmic performance under real-world illumination variations. FGF-SSIM achieves the optimal performance, marginally surpassing the L2-based results and substantially outperforming conventional SSIM. Specifically, under a 30% luminance adjustment, FGF-SSIM attains peak metrics of 21.94 dB PSNR, 0.9639 SSIM, and 415.62 MSE, exceeding the L1 algorithm by 1.82 dB PSNR, 0.173 SSIM, and 216.21 MSE. This improvement stems from fractional-order local statistics in FGF-SSIM, which replaces first-order Gaussian convolution kernels in traditional SSIM with fractional-order counterparts for decoupled luminance-contrast-structure control. The power-law decaying weights ( m 2 + n 2 ) α / 2 confer long-range dependence to capture non-local illumination changes, while Gaussian attenuation preserves local details and smoothness, enabling robustness to global illumination variations without compromising structural fidelity.
Figure 6. Comparative results of robustness to illumination variations between FGF-SSIM and benchmark metrics using urban01 sequence on M2UD dataset.
As illustrated in Figure 7, We provide a more detailed analysis of FGF-SSIM and other loss functions in localized regions, with a specific focus on their denoising performance and illumination robustness of the M2UD dataset []. In terms of denoising efficacy, although SSIM and L2 achieve superior results, FGF-SSIM remains competitive, securing the third-best performance while preserving the desirable properties of conventional SSIM. More importantly, in assessing photometric robustness, FGF-SSIM outperforms all other methods across three key metrics: PSNR, SSIM, and MSE. This improvement can be attributed to the fractional-order gradient flow and the fractional Gaussian field, which enhance structural sensitivity and illumination invariance. By incorporating fractional calculus, FGF-SSIM more effectively captures fine-grained textures and edges while suppressing high-frequency noise, thereby increasing PSNR and SSIM while reducing MSE. Particularly in the highlighted scenario of direct sunlight exposure within an autonomous driving context, FGF-SSIM exhibits enhanced capability in capturing fine details and maintaining structural integrity. It demonstrates superior adaptability to challenging lighting conditions, such as in regions where sunlight interacts with complex textures like tree branches.
Figure 7. Comparative analysis of FGF-SSIM versus conventional loss functions in varying illumination conditions.

5.5. Illumination Invariance Performance Evaluation

Figure 8 presents the illumulation invariance performance evaluation between the SSIM and FGF-SSIM metrics applied to the urban03 scene from the M2UD dataset [], alongside a comparative decomposition analysis. To underscore the advantages of incorporating fractal calculus into the SSIM framework, we provide a detailed examination of the principal loss components and error distributions of both methods. As illustrated in Figure 8a, both methods are evaluated under a 30% global illumination increase. The results demonstrate that FGF-SSIM exhibits significantly greater robustness to illumination variations. Images processed with FGF-SSIM appear more natural and sharper, with better preservation of structural details such as edges and textures. Figure 8b presents the Main Loss Component, which corresponds to the luminance computation segment of each loss function. The luminance error distribution of FGF-SSIM more effectively captures inter-object illumination variations, indicating its superior ability to maintain the global luminance structure of the scene. In Figure 8d, two localized regions are magnified for direct comparison. FGF-SSIM more accurately renders the brightness of fine structures such as background branches, thereby contributing to improved illumination invariance in the final output. Figure 8c compares the error distributions of both methods relative to the ground truth image. FGF-SSIM achieves lower and more uniformly distributed errors, confirming its higher reconstruction quality. This improvement is attributed to the long-range memory characteristics of FGF-SSIM, which enhance its capacity to preserve global structural information and fine details while effectively smoothing homogeneous regions. Figure 8e provides an additional zoomed-in comparison of two local patches. While SSIM fails to retain structural object details, FGF-SSIM clearly maintains the object’s structure, further corroborating its stability under illumination changes. Illumination variation constitutes a global monotonic transformation of pixel values. Although it significantly alters absolute intensity values, relative structural relationships within the image remain largely intact. FGF-SSIM leverages a fractional Gaussian field kernel endowed with long-range memory properties, enabling a wider receptive field. This allows the model to incorporate broader contextual information when estimating local luminance, emphasizing relational structural cues over absolute pixel values.
Figure 8. Illumination invariance performance evaluation and comparative decomposition analysis of SSIM and FGF-SSIM on the urban03 scene of the M2UD dataset.
Consequently, the Main Loss Component of FGF-SSIM remains substantially more stable under illumination shifts compared to that of conventional SSIM. This inherent stability of the core computational component directly translates to more consistent output quality, culminating in superior performance in illumination invariance evaluation. In summary, the illumination invariance exhibited by FGF-SSIM is not a product of post-processing techniques, but an intrinsic advantage rooted in its mathematical foundation based on fractional Gaussian fields.

5.6. Monocular Reconstruction Comparison Results

As shown in Table 1, we integrated the FGF-SSIM loss function into the SLAM framework and assessed its performance on the TUM monocular dataset [], demonstrating consistent superiority across all evaluation measures. FGF-SLAM achieves a PSNR of 36.04 dB, SSIM of 0.95, and LPIPS of 0.07, exceeding the second-best method by margins of +12.11 dB, +0.15, and −0.18 respectively. The TUM Mono dataset presents considerable challenges including rapid motion-induced blur, underexposed and noisy images from legacy sensors, and textureless regions with repetitive patterns. These conditions impose stringent demands on monocular SLAM systems, where the absence of depth information amplifies sensitivity to photometric variations and structural inconsistencies. While MonoGS-SLAM employs a combination of L1 and conventional SSIM losses, and UDGS-SLAM utilizes an undisclosed loss formulation, both exhibit limitations in handling the complex illumination conditions and structural degradation inherent in monocular visual SLAM.
Table 1. Monocular reconstruction comparison results of SLAM methods.
The exceptional performance of FGF-SLAM in monocular settings stems from fundamental mechanistic advantages specifically designed for photometric robustness. MonoGS-SLAM’s L1+SSIM formulation, while providing basic structural awareness, suffers from the inherent limitations of conventional Gaussian kernels in SSIM—limited receptive fields and sensitivity to local illumination changes. This proves particularly problematic in monocular SLAM where depth ambiguity exacerbates the impact of photometric inconsistencies. UDGS-SLAM, while achieving competitive results, lacks the mathematical foundation to explicitly address long-range dependencies in illumination variations. In contrast, FGF-SSIM introduces a paradigm shift through its fractional Gaussian field formulation, which establishes extended spatial correlations via power-law decay while maintaining local structural fidelity. This is critically important in monocular SLAM, where the absence of geometric constraints from depth sensors places greater emphasis on photometric consistency for both tracking and mapping. The adaptive decomposition of luminance, contrast, and structural components allows FGF-SLAM to dynamically balance photometric invariance with structural preservation—a capability particularly valuable in monocular scenarios where illumination changes can easily mislead traditional photometric-based optimization.

5.7. New View Synthesis Results

Figure 9 presents a comparative analysis of novel view synthesis quality after 3D reconstruction among different 3DGS-based methods on the Replica dataset []. To evaluate the advantages of incorporating fractional Gaussian fields, we selected two state-of-the-art models from CVPR 2024 (SplaTAM [], MonoGS []) for performance comparison. All three methods demonstrate competent scene reconstruction capabilities when provided with RGB-D input data. However, our proposed FGF-SLAM exhibits superior performance in detail preservation and long-range dependency modeling. In the Office1 scenario, for occluded regions in the input images (such as the rear corner of a pillow), SplaTAM employs blank Gaussian primitives for completion, resulting in visible artifacts. While MonoGS-SLAM attempts to reconstruct these areas, it introduces noticeable color discrepancies near boundaries, where whitish artifacts appear inconsistent with the original beige pillow and black sofa. In contrast, FGF-SLAM effectively handles these challenging regions by leveraging its long-range memory capacity during novel view synthesis, producing more natural and coherent scene reconstructions. This advantage becomes more pronounced in the sofa seam reconstruction of Office3. Given frontal views in the input sequence, SplaTAM generates blurred seam structures, MonoGS-SLAM produces moderately defined edges, whereas FGF-SLAM accurately reproduces sharp and geometrically consistent seams. Furthermore, FGF-SLAM demonstrates exceptional capability in fine detail reconstruction. In the Office4 scenario featuring wall graffiti, SplaTAM captures only coarse visual patterns while losing structural details of specific lines. MonoGS-SLAM achieves reasonable reconstruction but exhibits limitations in reproducing intricate patterns, particularly in complex shapes such as facial features and hat contours. FGF-SLAM successfully preserves all fine details with superior line accuracy and geometric fidelity. These results substantiate the effectiveness of fractional-order Gaussian fields in 3D reconstruction, directly attributable to the long-range dependency modeling enabled by fractional-order kernels, which enhances structural consistency across spatial domains.
Figure 9. Qualitative comparison of novel view synthesis between 3DGS-based SLAM methods on Replica dataset.
As shown in Table 2, we further assess the proposed FGF-SLAM on the Replica synthetic dataset [], demonstrating its exceptional capability in novel view synthesis using RGB-D data. The results demonstrate that FGF-SLAM achieves notable improvements across both PSNR and LPIPS metrics, outperforming both NeRF-based SLAM systems (e.g., NICE-SLAM [], Vox-Fusion [], Point-SLAM []) and 3DGS-based approaches (such as GS-SLAM [], SplaTAM [], MonoGS []). Specifically, FGF-SLAM attains state-of-the-art average scores on Replica, reaching 38.70 dB in PSNR and 0.045 in LPIPS, exceed those of the second-best method by margins of +1.20 dB and −0.025, respectively. The Replica dataset consists of high-fidelity synthetic indoor scenes featuring noise-free, blur-free imagery and accurate geometric ground truth, thereby posing challenges such as reconstructing fine geometric details, handling highly reflective surfaces, and maintaining overall visual fidelity. These results indicate that FGF-SLAM, through the incorporation of fractional calculus, enhances the preservation of edge and texture details, effectively reducing pixel-level errors in detailed regions. This leads to superior global pixel fidelity, as reflected in the highest PSNR value. Moreover, by independently modulating the contrast-structure component, FGF-SLAM produces images with improved perceptual clarity, enriched textures, and more coherent structures. Even without exact pixel-wise alignment, the visual appearance and perceptual quality of the generated views are closest to the ground truth, accounting for its significant advantage in LPIPS, with a relative improvement of 35.71% over the second-best method. Although the SSIM score of 0.970 ranks second, it trails the top method by only 0.005. This minor gap can be attributed to an inherent trade-off between low-frequency luminance consistency and high-frequency detail enhancement, suggesting that FGF-SLAM prioritizes perceptually salient features over strict brightness uniformity, thereby achieving better overall perceptual and quantitative performance.
Table 2. New view synthesis results of SLAM methods.

5.8. Ablation Experiments Results

Table 3 presents a systematic ablation study of FGF-SSIM on the office1 scenario from the Replica dataset [], analyzing the role of the proposed FGF-SSIM component within the SLAM system. Due to its low-light conditions, the Office 1 scenario in the Replica dataset is well-suited for assessing the efficacy of various loss terms in recovering scene luminosity and geometric structure. To evaluate their efficacy in low-light conditions, we ablated various loss function formulations (including L1, L2, SSIM, MS-SSIM, and our proposed FGF-SSIM) on the Office 1 scenario, assessing their impact on reconstruction quality. As shown in Table 3, the proposed FGF-SSIM achieves superior performance across all metrics compared to other loss functions. Notably, its ranking within the full VSLAM system (L2 < MS-SSIM < L1 < SSIM < FGF-SSIM) differs from its standalone evaluation (L1 < MS-SSIM < SSIM < L2 < FGF-SSIM). The limitations of L2 loss stem from its inherent rigidity, sensitivity to outliers, and instability in low-signal conditions. When integrated into a VSLAM framework, these shortcomings are magnified, propagating errors through the coupled pose estimation and mapping modules. In contrast, the SSIM loss assesses image fidelity through local luminance, contrast, and structure, prioritizing the preservation of edges, corners, and texture patterns over exact pixel-wise intensity. This makes it inherently more aligned with the requirements of robust SLAM. This discrepancy arises because a loss function’s efficacy in isolation does not fully predict its behavior in a complex, coupled system. Our proposed FGF-SSIM builds upon this principled foundation of SSIM, further enhancing the geometric clarity and structural coherence of the reconstructed map. FGF-SSIM demonstrates optimal synergy with the pose estimation and mapping modules, effectively balancing texture fidelity and geometric consistency, which is critical for overall VSLAM performance.
Table 3. Ablation experiments results of loss function formulations on the Replica dataset’s office1 scene.
Table 4 presents a systematic loss function ablation study of FGF-SLAM conducted on the Replica dataset’s office1 scene []. We analyze three critical components: photometric loss, depth loss, and anisotropic regularization, quantifying their contributions to VSLAM performance. When disabling the photometric loss, the system retains functionality through depth alignment and anisotropic regularization. However, quantitative results indicate severe degradation in pose estimation accuracy (RMSE ATE error increases by 374% relative to the baseline scenario), confirming the essential role of photometric alignment in VSLAM. This component enables cross-frame pixel intensity alignment, serving as a primary visual feedback mechanism for camera localization. Furthermore, under photometric-deprived conditions, the ablation of photometric loss leads to a marked decline in reconstruction quality (PSNR decreases by 18.94%, SSIM decreases by 10.99%, PSNR increases by 393.81%). These results underscore the necessity of incorporating photometric processing via a fractional Gaussian field. Ablating the depth loss, by contrast, preserves reasonable reconstruction quality compared to photometric ablation (PSNR decreases by 10.89%, SSIM decreases by 8.25%, PSNR increases by 292.86%), suggesting RGB inputs alone can sustain mapping integrity. However, the pose estimation accuracy performance degrades substantially (RMSE ATE increases by 293%), indicating that depth constraints help curb error accumulation in pose estimation. This contrast underscores the secondary yet valuable role of depth loss: while not critical for scene representation, it supplies geometric priors that improve tracking accuracy. Notably, our FGF-SLAM framework was originally designed to perform 3D reconstruction without explicit depth input. When anisotropic regularization is disabled, performance degradation is the smallest among all ablations (PSNR decreases by 7.23%, SSIM decreases by 1.83%, PSNR increases by 122.58%, RMSE ATE increases by 52.38%), primarily manifesting as increased surface artifacts in reconstructed meshes. This confirms its auxiliary role in noise suppression rather than in core SLAM functionality. Nevertheless, such regularization aids in recovering finer luminance and structural details in VSLAM.
Table 4. Ablation experiments results of loss function components on the Replica dataset’s office1 scene.

6. Conclusions

This work presents a comprehensive solution to the fundamental challenge in dense visual SLAM where existing loss functions fail to properly balance luminance, contrast, and structural fidelity under varying illumination conditions. We introduce a fractional Gaussian field (FGF) that innovatively combines Caputo fractional derivatives with Gaussian weighting, creating a hybrid kernel coupling power-law decay for long-range memory with Gaussian attenuation for local smoothness. Kernel analysis confirms that the FGF maintains non-negligible weights at larger distances, retaining a weight of 0.01 even at 10-pixel distances while conventional Gaussian kernels attenuate rapidly to near-zero values, establishing superior long-range dependency while preserving local structural coherence. Building on this foundation, we develop FGF-SSIM, a novel loss function that adaptively recalibrates SSIM components through fractional-order statistics, and integrate it into a complete 3D Gaussian Splatting based SLAM framework (FGF-SLAM) where it serves as the primary photometric objective in both tracking and mapping modules. Rigorous validation confirms the effectiveness of our approach. Illumination robustness tests show FGF-SSIM achieves 21.94 dB PSNR under 30% luminance increase, outperforming L1 loss by 1.82 dB. Reconstruction accuracy on standard benchmarks reveals FGF-SLAM attains 36.04 dB PSNR on the TUM dataset, representing a 72% improvement with 1.73 cm ATE RMSE, and achieves 38.70 dB PSNR with 0.045 LPIPS on the Replica dataset. Ablation studies validate our design choices, showing that removing FGF-SSIM causes 18.9% PSNR degradation, confirming its critical role in achieving robust high-quality SLAM performance. While the proposed method demonstrates superior performance, the observed trade-off between long-range dependency modeling and local precision in the fractional Gaussian kernel warrants further investigation for optimal parameter adaptation across different visual conditions. Future work will focus on adaptive parameter optimization and explore broader applications in image processing domains such as medical image enhancement, remote sensing analysis, and high dynamic range imaging, where robust multi-scale feature preservation under varying acquisition conditions is equally crucial as in VSLAM systems.

Author Contributions

Conceptualization, J.Z. and H.Z.; methodology, Z.Z. and M.F.; validation, H.Y.; resources, Y.L.; writing—original draft preparation, H.Z. and Y.L.; writing—review and editing, H.Z. and M.F.; funding acquisition, Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62305393.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, W.; Fu, C.; Zhu, L.; Loo, S.Y.; Zhang, H. Rumination meets VSLAM: You do not need to build all the submaps in realtime. IEEE Trans. Ind. Electron. 2023, 71, 9212–9221. [Google Scholar] [CrossRef]
  2. Cai, D.; Li, S.; Qi, W.; Ding, K.; Lu, J.; Liu, G.; Hu, Z. DFT-VSLAM: A dynamic optical flow tracking VSLAM method. J. Intell. Robot. Syst. Theor. Appl. 2024, 110, 135. [Google Scholar] [CrossRef]
  3. Kerl, C.; Sturm, J.; Cremers, D. Robust odometry estimation for RGB-D cameras. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 3748–3754. [Google Scholar] [CrossRef]
  4. Zhang, J.; Yao, Y.; Quan, L. Learning signed distance field for multi-view surface reconstruction. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 6525–6534. [Google Scholar] [CrossRef]
  5. Yang, K.; Cheng, Y.; Chen, Z.; Wang, J. Slam meets nerf: A survey of implicit slam methods. World Electr. Veh. J. 2024, 15, 85. [Google Scholar] [CrossRef]
  6. Kerbl, B.; Kopanas, G.; Leimkühler, T.; Drettakis, G. 3D Gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 2023, 42, 1–14. [Google Scholar] [CrossRef]
  7. Zhu, H.; Zhang, Z.; Zhao, J.; Duan, H.; Ding, Y.; Xiao, X.; Yuan, J. Scene reconstruction techniques for autonomous driving: A review of 3D Gaussian splatting. Artif. Intell. Rev. 2025, 58, 1–33. [Google Scholar] [CrossRef]
  8. Fei, B.; Xu, J.; Zhang, R.; Zhou, Q.; Yang, W.; He, Y. 3d gaussian splatting as new era: A survey. IEEE Trans. Vis. Comput. Graph. 2025, 31, 4429–4449. [Google Scholar] [CrossRef]
  9. Keetha, N.; Karhade, J.; Jatavallabhula, K.M.; Yang, G.; Scherer, S.; Ramanan, D.; Luiten, J. Splatam: Splat track & map 3d gaussians for dense rgb-d slam. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 21357–21366. [Google Scholar] [CrossRef]
  10. Ha, S.; Yeon, J.; Yu, H. Rgbd gs-icp slam. In Proceedings of the 18th European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 180–197. [Google Scholar] [CrossRef]
  11. Baker, A.H.; Pinard, A.; Hammerling, D.M. On a structural similarity index approach for floating-point data. IEEE Trans. Vis. Comput. Graph. 2023, 30, 6261–6274. [Google Scholar] [CrossRef]
  12. Yan, C.; Qu, D.; Xu, D.; Zhao, B.; Wang, Z.; Wang, D.; Li, X. Gs-slam: Dense visual slam with 3d gaussian splatting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 19595–19604. [Google Scholar] [CrossRef]
  13. Matsuki, H.; Murai, R.; Kelly, P.H.; Davison, A.J. Gaussian splatting slam. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 18039–18048. [Google Scholar] [CrossRef]
  14. Soliman, M.; Ahmed, H.M.; Badra, N.M.; Samir, I.; Radwan, T.; Ahmed, K.K. Fractional wave structures in a higher-order nonlinear Schrödinger equation with cubic–quintic nonlinearity and β-fractional dispersion. Fractal Fract. 2025, 9, 522. [Google Scholar] [CrossRef]
  15. Shao, Z.; Kosari, S.; Yadollahzadeh, M.; Derakhshan, M. Stability analysis of an efficient hybrid numerical approach for solving the two-dimensional space/multi-time fractional Bloch-Torrey model involving the Riesz fractional operator. Numer. Algorithms 2025, 1–27. [Google Scholar] [CrossRef]
  16. Dubey, S.; Chakraverty, S. Hybrid techniques for approximate analytical solution of space-and time-fractional telegraph equations. Pramana-J. Phys. 2022, 97, 11. [Google Scholar] [CrossRef]
  17. Liu, J.G.; Yang, X.J.; Geng, L.L.; Yu, X.J. On fractional symmetry group scheme to the higher-dimensional space and time fractional dissipative Burgers equation. Int. J. Geom. Methods Mod. Phys. 2022, 19, 2250173. [Google Scholar] [CrossRef]
  18. Liu, J.G.; Yang, X.J.; Feng, Y.Y.; Geng, L.L. Symmetry analysis of the generalized space and time fractional Korteweg–de Vries equation. Int. J. Geom. Methods Mod. Phys. 2021, 18, 2150235. [Google Scholar] [CrossRef]
  19. Pu, Y.F.; Zhou, J.L.; Yuan, X. Fractional differential mask: A fractional differential-based approach for multiscale texture enhancement. IEEE Trans. Image Process. 2009, 19, 491–511. [Google Scholar] [CrossRef]
  20. Pu, Y.F.; Siarry, P.; Chatterjee, A.; Wang, Z.N.; Yi, Z.; Liu, Y.G.; Zhou, J.L.; Wang, Y. A fractional-order variational framework for retinex: Fractional-order partial differential equation-based formulation for multi-scale nonlocal contrast enhancement with texture preserving. IEEE Trans. Image Process. 2017, 27, 1214–1229. [Google Scholar] [CrossRef]
  21. Chen, H.; Qiao, H.; Wei, W.; Li, J. Time fractional diffusion equation based on caputo fractional derivative for image denoising. Opt. Laser Technol. 2024, 168, 109855. [Google Scholar] [CrossRef]
  22. Moniri, Z.; Babaei, A.; Moghaddam, B.P. Robust numerical framework for simulating 2D fractional time–space stochastic diffusion equation driven by spatio-temporal noise: L1-FFT hybrid approach. Commun. Nonlinear Sci. 2025, 146, 108791. [Google Scholar] [CrossRef]
  23. Huang, G.; Qin, H.Y.; Chen, Q.; Shi, Z.; Jiang, S.; Huang, C. Research on application of fractional calculus operator in image underlying processing. Fractal Fract. 2024, 8, 37. [Google Scholar] [CrossRef]
  24. Jayasinghe, H.; Brilakis, I. Towards a density preserving objective function for learning on point sets. In Proceedings of the 18th European Conference on Computer Vision, ECVA, Milan, Italy, 29 September–4 October 2024; pp. 128–145. [Google Scholar] [CrossRef]
  25. Hyeon, E.; Ersal, T.; Kim, Y.; Stefanopoulou, A.G. Loss function design for data-driven predictors to enhance the energy efficiency of connected and automated vehicles. IEEE Trans. Intell. Transp. Syst. 2022, 24, 827–837. [Google Scholar] [CrossRef]
  26. Liu, L.; Li, P.; Chu, M.; Gong, R. Nonparallel support vector machine with L2-norm loss and its dcd-type solver. Neural Process. Lett. 2023, 55, 4819–4841. [Google Scholar] [CrossRef]
  27. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  28. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; pp. 1398–1402. [Google Scholar] [CrossRef]
  29. Chen, G.H.; Yang, C.L.; Xie, S.L. Gradient-based structural similarity for image quality assessment. In Proceedings of the 2006 International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 2929–2932. [Google Scholar] [CrossRef]
  30. Li, C.; Bovik, A.C. Content-partitioned structural similarity index for image quality assessment. Signal Process. Image Commun. 2010, 25, 517–526. [Google Scholar] [CrossRef]
  31. Wang, Z.; Li, Q. Information content weighting for perceptual image quality assessment. IEEE Trans. Image Process. 2010, 20, 1185–1198. [Google Scholar] [CrossRef]
  32. Zhao, H.; Gallo, O.; Frosio, I.; Kautz, J. Loss functions for image restoration with neural networks. IEEE Trans. Comput. Imaging 2016, 3, 47–57. [Google Scholar] [CrossRef]
  33. Park, S.; Ma, J.; Oh, S.W.; Kim, W.S. Blur image representation in defocus using generalized Gaussian kernel. J. Inf. Disp. 2025, 26, 321–328. [Google Scholar] [CrossRef]
  34. Wang, Y.; Yang, L. Metric learning-based generalized Gaussian kernel for nonlinear classification. Eng. Appl. Artif. Intell. 2025, 139, 109605. [Google Scholar] [CrossRef]
  35. Straub, J.; Whelan, T.; Ma, L.; Chen, Y.; Wijmans, E.; Green, S.; Engel, J.J.; Mur-Artal, R.; Ren, C.; Verma, S.; et al. The replica dataset: A digital replica of indoor spaces. arXiv 2019, arXiv:1906.05797. [Google Scholar] [CrossRef]
  36. Sturm, J.; Engelhard, N.; Endres, F.; Burgard, W.; Cremers, D. A benchmark for the evaluation of RGB-D SLAM systems. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura-Algarve, Portugal, 7–12 October 2012; pp. 573–580. [Google Scholar] [CrossRef]
  37. Jia, Y.; Wang, S.; Shao, S.; Wang, Y.; Zhang, F.; Wang, T. M2UD: A multi-model, multi-scenario, uneven-terrain dataset for ground robot with localization and mapping evaluation. arXiv 2025, arXiv:2503.12387. [Google Scholar] [CrossRef]
  38. Zhu, Z.; Peng, S.; Larsson, V.; Xu, W.; Bao, H.; Cui, Z.; Oswald, M.R.; Pollefeys, M. Nice-slam: Neural implicit scalable encoding for slam. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12786–12796. [Google Scholar] [CrossRef]
  39. Yang, X.; Li, H.; Zhai, H.; Ming, Y.; Liu, Y.; Zhang, G. Vox-fusion: Dense tracking and mapping with voxel-based neural implicit representation. In Proceedings of the IEEE International Symposium on Mixed and Augmented Reality, Singapore, 17–21 October 2022; pp. 499–507. [Google Scholar] [CrossRef]
  40. Sandström, E.; Li, Y.; Van Gool, L.; Oswald, M.R. Point-slam: Dense neural point cloud-based slam. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 18433–18444. [Google Scholar] [CrossRef]
  41. Mur-Artal, R.; Tardós, J.D. Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  42. Zhu, R.; Liang, Y.; Chang, H.; Deng, J.; Lu, J.; Yang, W.; Zhang, T.; Zhang, Y. Motiongs: Exploring explicit motion guidance for deformable 3d gaussian splatting. Adv. Neural Inf. Process. Syst. 2024, 37, 101790–101817. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.