2.1. Symbolic Kinematic Modeling Based on D-H Parameters
The classical Denavit–Hartenberg (D–H) model, originally proposed by Denavit and Hartenberg [
19], is adopted in this study to serve as the symbolic knowledge module of the proposed framework. It utilizes homogeneous transformation matrices to rigorously describe the rigid geometric topology of the robot. The six-axis ABB IRB 120 industrial manipulator employed in this work is illustrated in
Figure 1, and its nominal kinematic parameters are listed in
Table 1. Unlike traditional methods that attempt to lump all errors into D–H parameters, this study clearly distinguishes between geometric and non-geometric factors. The D–H model is strictly used to identify static geometric deviations, while the complex electromechanical coupling effects (which violate rigid-body assumptions) are treated as unmodeled dynamics to be captured by the subsequent neural network. Identification of these symbolic geometric states is conducted by analyzing the deviation between the measured end-effector position and its theoretical symbolic prediction.
Based on this rigid-body symbolic formulation, the forward kinematic model is established. The homogeneous transformation matrix of the
i-th link, denoted as
, is mathematically expressed as:
where
represents the link length,
denotes the link offset,
specifies the link twist angle, and
denotes the joint angle. In the context of parameter identification, we identify a constant calibrated joint offset
, which absorbs the nominal D–H offset in
Table 1:
therefore, the actual joint angle used in forward kinematics is
In this paper, the parameters
used in Equation (
1) are the
calibrated D–H formulation. The nominal values are first listed in
Table 1, while the calibrated values are reported in
Table 7. By successively multiplying the individual transformation matrices, the global symbolic pose of the system relative to the base frame can be derived as:
where the system considered in this study is a six-axis manipulator (
). The Cartesian position of the end-effector corresponds to the translation vector extracted from the final transformation matrix:
The theoretical distance between this predicted position and the encoder attachment point
(defined in the robot base frame) constitutes the symbolic observation function:
Accordingly, the measurement residual at the
k-th configuration is defined as the discrepancy between the physical measurement and the symbolic prediction:
Here,
denotes the vector of symbolic geometric states to be identified, constructed by stacking the deviations of the D–H parameters for all
joints. Specifically, we define
, where
,
, and
represent the identified corrections to the nominal link length, link offset, and twist angle of the
i-th joint, respectively, and
is the identified joint zero-offset correction. Accordingly, the calibrated parameters used in the forward kinematics are given by
,
,
, and
with
. Therefore,
Table 7 reports the calibrated D–H parameters after compensation, while
Table 1 provides the nominal values for reference. The function
denotes the theoretically predicted cable length computed from the current (calibrated) geometric parameters, and
is the absolute length measured by the draw-wire encoder. The joint configuration vector is defined as
. It should be noted that the measurement residual contains unmodeled non-geometric dynamics, which will be explicitly learned by the Adaptive Wavelet Network in the subsequent state-space formulation.
2.2. Neuro-Symbolic State-Space Formulation with Adaptive Wavelet Networks
In this study, we construct a Neuro-Symbolic State-Space Model to perform dynamic latent state estimation. Unlike standard EKF approaches, which rely on strict rigid-body assumptions, this framework integrates a symbolic kinematic model with an Adaptive Wavelet Network (AWNN) into the observation equation. This fusion allows for the explicit separation of static geometric deviations from configuration-dependent non-geometric residuals (e.g., joint compliance).
The state vector is defined as the
deviations (corrections) to the nominal D–H parameters, including the constant joint zero-offset deviations:
The calibrated joint offset is then reconstructed as
Accordingly, the calibrated geometric parameters used in forward kinematics are
,
,
.
Moreover,
denotes the calibrated constant joint offset (with the nominal offset absorbed) as defined in Equation (
2), such that
. These parameters are used in the forward kinematics via Equations (1) and (2).
In the prediction step, since the base kinematic parameters describe the physical structure of the robot, they are modeled as a stationary process with Gaussian process noise. The symbolic state evolution is formulated as follows:
where
denotes the process noise covariance. The corresponding covariance propagation is given by the following:
The introduction of the process noise covariance partially relaxes the strict stationarity assumption on the symbolic states. It permits minor state fluctuations, enabling the recursive estimator to track slow, random walk geometric drifts over time without destabilising the model. Consequently, a slowly time-varying drift is not fundamentally inconsistent with the present representation, provided is tuned in accordance with the expected drift rate.
The observation model utilizes the absolute cable length measured by the draw-wire encoder. To account for complex dynamic drifts, the observation equation innovatively fuses the symbolic prediction with the neural output:
Here,
is the Euclidean distance derived from the symbolic D-H model:
Distinct from conventional Multi-Layer Perceptrons (MLP), the non-geometric residual term
is modeled by an Adaptive Wavelet Network (AWNN) to leverage its time-frequency localization properties. We employ the Mexican Hat wavelet as the activation function. The forward propagation is mathematically expressed as follows:
where
M is the number of wavelet neurons. The network parameters
include the output weights
, the translation factors
, the dilation factors
, and the input projection weights
. The adaptive nature of
and
allows the network to automatically adjust its receptive field to capture local error singularities. Unlike standard RBFs or Morlet wavelets, the Mexican Hat wavelet possesses a strict zero-mean property and superior time-frequency localization. This morphology effectively isolates sharp, localized non-geometric singularities (e.g., gear backlash or compliance) without globally distorting the learned error field. Furthermore, the zero mean morphology of the Mexican Hat wavelet (
) acts as a powerful implicit regularization on the residual field. Unlike standard activation functions that easily output constant biases, this topological property provides a strong structural inductive bias that heavily suppresses the network’s ability to maintain a constant DC offset over the bounded workspace. Consequently, the network is strongly deterred from illegitimately absorbing the static geometric offsets associated with the rigid body DH parameters, naturally complementing the decoupled global refinement by focusing on high frequency, oscillatory residuals.
To linearize the observation model for the recursive update, we compute the Jacobian of the
predicted measurement with respect to the symbolic states. Define
For notational brevity, let
. The EKF observation Jacobian is
since
depends only on the commanded joint configuration
and does not explicitly depend on
. Therefore, the neural compensation term does not contribute to
, and the EKF linearization is performed only with respect to the symbolic geometric states.
Consequently, the Jacobian reduces to the symbolic model derivative:
Here,
is the unit vector along the cable direction, so each
is the projection of the end-effector sensitivity onto the measured distance. The partial derivative is calculated using differential kinematics. Let the parameter
belong to the
i-th joint, then
where
and
are the cumulative transformation matrices, and the operator
extracts the translational components (i.e., the first three elements of the fourth column) from a
homogeneous matrix. The full Jacobian matrix is assembled as
.
With the linearized model, the recursive inference steps are performed. The innovation covariance
and Kalman gain
are computed as follows:
The measurement innovation
represents the residual after removing both the symbolic prediction and the wavelet compensation:
Finally, the symbolic state estimate is updated via the Kalman gain:
Simultaneously, the neural parameters
are updated using the gradient of the squared innovation loss
. This ensures that the AWNN adaptively learns the residual dynamics that the symbolic model cannot explain:
Notably, this gradient descent updates not only the weights
but also the dilation
and translation
, enabling the network to dynamically refine its time-frequency resolution during the filtering process.
2.3. Decoupled Global Refinement via Levenberg–Marquardt Optimization
After Stage I, the Adaptive Wavelet Network parameters are frozen at their converged values, and the corresponding deterministic compensation field is denoted by . In Stage II, we refine only the symbolic geometric parameter vector via a global Levenberg–Marquardt (LM) batch optimization while keeping fixed.
For the
k-th measurement, we define the (prediction) residual as
where
is the symbolic D–H-based cable-length prediction computed from
, and
is the draw-wire encoder measurement. Accordingly, the residual vector is
The Jacobian matrix
is defined by
Since the frozen compensation term
depends only on the joint configuration and is independent of
, its derivative with respect to
vanishes. Therefore, each Jacobian row in Stage II is identical to the symbolic observation Jacobian
used in Stage I.
Using first-order derivatives, the gradient of the objective function is
and the Gauss–Newton approximation of the Hessian is
To improve numerical robustness, LM introduces a damping factor
and computes the update
by solving the damped normal equation
equivalently,
A candidate parameter vector is then obtained by
With
fixed, the candidate residual and loss are evaluated as
Finally, the damping factor and parameter state are updated using a loss-decrease rule:
where
is a user-defined adjustment factor (typically
). This decoupled formulation ensures that Stage II refines only the symbolic geometric parameters, while the learned non-geometric field
remains a deterministic correction term.
2.4. Design and Analysis of the PSO-Driven Neuro-Symbolic Framework
To provide a comprehensive visualization of the proposed calibration strategy, the complete workflow of the PSO-Driven Neuro-Symbolic State-Space Framework is illustrated in
Figure 2. The framework is structured into three sequential phases, progressing from autonomous meta-optimization to recursive inference, and finally to decoupled global refinement.
The process initiates with the system inputs (grey region), where nominal D–H parameters, measurement dataset , and the search space for hyperparameters are defined. The workflow first enters Stage 0: PSO Meta-Optimization. In this phase, a particle swarm autonomously explores the hyperparameter space to identify optimal values for process noise covariance , measurement noise variance R, and network initialization settings. This step effectively resolves the sensitivity issues inherent in recursive estimation, ensuring a robust starting point.
Subsequently, the system proceeds to Stage I: Recursive Neuro-Symbolic Inference (blue region). This stage functions as an online dual-estimation process. To preserve the physical definition of symbolic kinematic parameters, the Adaptive Wavelet Network (AWNN) compensation
is fused into the observation equation. For each measurement sample, the framework executes a synchronized update mechanism: the Extended Kalman Filter (EKF) recursively updates the symbolic geometric states (
), while the measurement innovation
drives the Stochastic Gradient Descent (SGD) update for the wavelet parameters. Unlike standard networks, this update adjusts not only the weights but also the dilation and translation factors (
), allowing the network to dynamically adapt its time-frequency resolution to capture local error singularities. This stage outputs a refined symbolic prior (
) and a trained wavelet compensator (
). Following the recursive inference, the process transitions to Stage II: Decoupled Global Refinement (orange region). As visually highlighted by the red dashed line in
Figure 2, the AWNN
trained in Stage I is frozen and transferred to the Levenberg–Marquardt (LM) module. The LM algorithm utilizes this frozen dynamic field as a deterministic non-geometric correction term to perform a global batch optimization. This strategy effectively decouples the optimization process, assisting the solver in avoiding local minima and ensuring convergence to the global geometric optimum without interference from dynamic noise. Finally, the framework yields the system output (green region), consisting of the optimal symbolic D–H parameters (
) and the non-geometric compensation model (
). The detailed algorithmic steps are provided in
Table 2.
The computational complexity is analyzed as follows. Let N denote the number of measurements, n the dimension of symbolic states (), and P the number of particles in PSO. In Stage 0, the complexity is proportional to the number of particles and iterations: . In Stage I, since the observation is a scalar (single cable length measurement), the innovation covariance reduces to a scalar, and consequently the Kalman gain computation simplifies to rather than . Combined with the wavelet network forward/backward pass (), Stage I achieves linear complexity: . In Stage II, the LM algorithm involves iterative Jacobian assembly and the solution of a damped normal equation, with complexity . Since n is a small constant (), this simplifies to . Although the PSO stage introduces a constant multiplier, the overall algorithmic complexity remains linear with respect to the dataset size N, ensuring scalability for large-scale industrial calibration tasks.