Next Article in Journal
Evaluation Protocol Sensitivity in Frequency-Constrained Truss Optimization: A Comparative Study of Adaptive and Evolutionary Metaheuristics
Previous Article in Journal
Collagen-Based Microspheres for Biomedical Applications in Drug Delivery and Tissue Engineering
Previous Article in Special Issue
Global Path Planning for Land–Air Amphibious Biomimetic Robot Based on Improved PPO
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Segmentation-Guided Visual Tracking on a Bio-Inspired Quadruped Robot

1
School of Aerospace Science, Harbin Institute of Technology, Shenzhen 518055, China
2
School of Mechanical Engineering and Automation, Harbin Institute of Technology, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Biomimetics 2026, 11(4), 234; https://doi.org/10.3390/biomimetics11040234
Submission received: 8 February 2026 / Revised: 10 March 2026 / Accepted: 29 March 2026 / Published: 2 April 2026

Abstract

Bio-inspired quadrupedal robots exhibit superior adaptability and mobility in unstructured environments, making them suitable for complex task scenarios such as navigation, obstacle avoidance, and tracking in a variety of environments. Visual perception plays a critical role in enabling autonomous behavior, offering a cost-effective alternative to multi-sensor systems. This paper proposes an image segmentation-guided visual tracking framework to enhance both perception and motion control in quadruped robots. On the perception side, a cascaded convolutional neural network is introduced, integrating a global information guidance module to fuse low-level textures and high-level semantic features. This architecture effectively addresses limitations in single-scale feature extraction and improves segmentation accuracy under visually degraded conditions. On the control side, segmentation outputs are embedded into a biologically inspired central pattern generator (CPG), enabling coordinated generation of limb and spinal trajectories. This integration facilitates a closed-loop visual-motor system that adapts dynamically to environmental changes. Experimental evaluations on benchmark image segmentation datasets and robotic locomotion tasks demonstrate that the proposed framework achieves enhanced segmentation precision and motion flexibility, outperforming existing methods. The results highlight the effectiveness of vision-guided control strategies and their potential for deployment in real-time robotic navigation.

1. Introduction

Quadrupedal robots, owing to their bio-inspired leg structures, exhibit superior locomotion capabilities compared to wheeled robots, making them highly promising for complex task scenarios. Enhancing their ability to autonomously complete tasks necessitates advanced environmental perception. Compared to multi-sensor fusion approaches, such as those relying on LiDAR, visual sensors provide a cost-effective means for robots to perceive their surroundings [1]. Furthermore, the rapid advancement of neural networks has significantly improved computer vision’s ability to extract meaningful information from increasingly complex environments [2]. Among computer vision techniques, image segmentation plays a crucial role as an indispensable auxiliary technology. Its primary function is to decompose visual scenes into meaningful regions, thereby facilitating downstream tasks such as detection, tracking, and recognition. Improving segmentation quality directly enhances the robustness and accuracy of robotic perception systems, with extensive applications in obstacle avoidance [3], navigation [4], and target tracking [5].
Traditional image segmentation methods rely heavily on hand-crafted features or intrinsic priors [6], which often limit their adaptability in complex or cluttered scenes. Recent advances in deep learning, particularly Convolutional Neural Networks (CNNs), have significantly boosted segmentation performance by learning multi-level features from data. However, many CNN-based models still struggle to balance fine-grained detail preservation with global contextual understanding, due to limitations in single-scale feature extraction. Effectively addressing this imbalance requires more sophisticated multi-level and multi-scale representation mechanisms to enhance both spatial resolution and semantic abstraction. To this end, recent Transformer-based approaches employ complex aggregation [7] or dynamic fusion [8] units to provide such global guidance. In contrast, our proposed Global Information Guidance Module (GIGM) achieves similar global-to-local interaction through a more lightweight filter-level sharing strategy, facilitating efficient multi-scale information exchange with significantly reduced structural complexity.
Meanwhile, bio-inspired quadrupedal robots have attracted growing attention for their superior mobility and adaptability in unstructured environments such as rugged terrain or confined spaces. Inspired by animal locomotion, recent work has explored various gait optimization and control strategies. Zhornyak et al. [9] applied genetic algorithms to replicate feline gaits, while Gangapurwala et al. [10] introduced guided constrained policy optimization to generate physically feasible locomotion behaviors. However, these methods often focus on specific motion modes and lack generality when adapting to dynamically changing tasks or environments. As a result, there is a growing need for more flexible and adaptive motion control frameworks that can seamlessly respond to perceptual feedback.
To address challenges in visual perception and motion adaptability, this paper proposes an image segmentation-guided visual tracking framework for bio-inspired quadruped robots.
  • On the perception side, we introduce a cascaded neural network equipped with a global information guidance module, which effectively integrates low-level texture details and high-level semantic features across layers, overcoming the limitations of single-scale feature extraction. This design enhances segmentation accuracy, particularly in visually cluttered or blurred environments.
  • On the control side, high-level information is incorporated into a biologically inspired central pattern generator (CPG) model to generate coordinated limb and spinal trajectories, enabling comprehensive motion adaptability in dynamically changing conditions. The segmentation results directly inform visual tracking and influence control decisions, creating a closed-loop visual-motor system.
  • We conducted extensive evaluations on standard image segmentation datasets and robotic tracking tasks to validate our approach. The results demonstrate that our method outperforms existing approaches in segmentation accuracy and motion flexibility, highlighting its potential for real-time robotic navigation in complex environments.

2. Related Work

2.1. Image Segmentation

Image segmentation aims to identify regions of greatest interest to people in images. Traditional image segmentation approaches usually predict saliency scores by utilizing hand-crafted cues or intrinsic priors [11,12]. However, they are limited due to their low efficiency and bad generalization ability. With the rise of deep learning, recent methods mostly leverage convolutional neural networks (CNNs) to make a pixel-to-pixel prediction.
Compared with traditional ones, CNN-based methods have shown superior performance on popular image segmentation benchmarks. Among them, early work [13,14,15] mostly adopted an iterative or stage-wise manner to refine the predictions step by step. Some later methods [16,17,18] focus on designing new multi-scale feature-extracting modules and strategies based on the U-shape architecture. Some [19,20,21] introduced various attention mechanisms to enhance the feature representation ability of the network.
In recent years, generative models have rapidly advanced and significantly influenced visual learning tasks, ranging from image synthesis [22] to reinforcement learning [23]. This trend has likewise motivated progress in image segmentation, where researchers have begun to integrate generative paradigms such as VAE-based approaches [24], GAN-driven frameworks [25], and diffusion model-based techniques [26]. These methods leverage generative priors to refine feature representations and promote more stable and coherent segmentation results.
Compared with existing image segmentation methods, we propose a new cascading interaction mode that combines multi-scale information with a global information guidance model to reduce the loss of detailed information and improve accuracy.

2.2. Low-Level Gait Controller

Low-level gait control involves managing the robot’s leg movements to maintain balance on various terrains. This includes controlling the swing and support phases of the legs, as well as adjusting the gait cycle and timing. Optimizing low-level gait control is crucial for quadrupedal robots, as it determines their stability, mobility, and efficiency across different environments [27,28]. Controlling the gait of quadruped robots involves multiple parameters and interactions, resulting in a complex parameter space. Efficient computation of gait parameters while maintaining real-time performance with limited computational resources is a challenge.
To tackle this challenge, researchers have proposed various methods in recent years. Some approaches [29,30,31] simplify the gait design of robots using model predictive control. For complex tasks, some methods [32,33] use a hierarchical control structure to match leg movements with torso movements. With the introduction of reinforcement learning, Tsounis et al. [34] use the landing point as a network output, allowing the robot to achieve a more flexible gait. Bellegarda et al. [35] combine CPG with deep reinforcement learning to create adaptable and robust movement patterns.
Unlike the above methods, this paper uses CPGs to parameterize the motion of a crawling quadruped robot, enabling omnidirectional motion across multiple gaits with fewer parameters. At the same time, the method also facilitates the integration with the upper framework.

3. Method

Figure 1 shows the flowchart of this paper.
The experimental platform used is a crawling quadruped robot with 3 degrees of freedom in its legs and 5 degrees of freedom in its spine, which can only bend and extend in the horizontal plane.
The tracking framework comprises an image segmentation model and a CPG-based low-level controller. The segmentation model takes an RGB image captured by the robot’s camera as input and outputs a binary image that contains only the target. The target position is calculated by combining the distance information obtained from the camera’s internal reference and the depth camera. The desired deflection angle is then calculated based on the target position. The parameters of the CPG are adjusted based on the desired deflection angle to generate mutually coupled rhythmic signals that control the robot’s joint trajectories and foot end, allowing the robot to track the target.

3.1. Image Segmentation

3.1.1. Cascaded Information Interaction Network

To precisely segment the target and facilitate the visual servoing module in calculating its position, we use the Swin transformer [36] as an encoder because of its unique advantages: the Swin transformer incorporates a local attention mechanism, inherits the advantages of CNNs in processing large images, and uses a window-based approach to exploit the transformer’s capabilities in long-range dependency modeling. To extract scale-specific features based on different backbone networks, we introduce an additional convolutional layer with a kernel size of 1 to standardize the channel dimensions. Consequently, the resulting unified channel features can be denoted as E = { E i , 1 i I } , where I is typically set to 5.
As shown in Figure 2, after applying convolutional pooling for down-sampling and subsequent up-sampling to restore the original resolution, images often suffer from blurring and loss of fine details. The conventional approach involves cascading feature maps at the same resolution along both the bottom-up and top-down paths, which mitigates the loss of local features to some extent. However, a direct feature extraction approach may limit multi-scale information fusion, as hierarchical feature interactions are often underutilized. To overcome this constraint, we propose a Cascaded Information Interaction Network, which enables multi-scale information exchange at the filter level. This technique establishes a structured mechanism for progressive feature refinement, ensuring effective communication across different resolution layers. Additionally, we recognize that deep architectures typically yield enhanced performance due to their ability to model complex patterns. Building on this idea, we expand the interaction layers in our model to strengthen hierarchical feature representation. Given the channel unified feature maps from the encoder E , the features delivered to the decoder D = { D i , 1 j J } could be obtained by cascaded interactors as
D j = F q E k , , E m , 1 j 5 , 1 k m 5
where F denotes the feature fusion in each interaction level, q indicates the number of function actions, which means the number of cascading levels.

3.1.2. Global Information Guidance Module

In segmentation tasks, an efficient multiscale module significantly enhances module performance. Higher-level information can serve to guide and enhance the interaction of lower-level information across different scales. To maintain the compression of both local and relative global information, we introduce a global information guidance module (GIGM). The higher-level information can serve to guide the lower-level information, thereby enhancing the interaction between different scales of information. The module input contains the lower-level information F i , which has been processed by a 1 × 1 convolutional layer. In addition, the higher-level information, F i + 1 , has been subjected to Global Maximum Pooling (GMP) and sigmoid function, as shown by the gray box in Figure 2. The higher-level information is compressed to calibrate the lower-level information, thereby preserving local features. Finally, the output D i is obtained after a 1 × 1 convolutional layer. D i serves as an information guide from the relatively higher level pathway to the lower level pathway. The module is expressed as follows:
G i + 1 = S i g m o i d G M P F i + 1 , 1 i M 1
D i = ( C o n v 1 + 1 ) ( G i + 1 ( C o n v 1 + 1 ) ( F i ) + F i ) , 1 i M .

3.2. Visual Servo Controller

To transform the target binary map generated by the segmentation model into a tracking instruction for the robot, we combine depth measurements to obtain distance information and calculate the desired deflection angle.
The zero-order moments ( M 00 ) and first-order moments ( M 10 ,   M 01 ) are solved for the binary map to obtain the position of the center-of-mass pixel point of the target object, respectively.
M 00 = W H P ( w , h ) , M 10 = W H w · P ( w , h ) , M 01 = W H h · P ( w , h )
P ( w , h ) has only two values, 0 (black) or 1 (white), and therefore M 00 represents the sum of the target regions in the map. M 10 and M 01 represent the accumulation of w and h coordinate values of the target area, respectively. Therefore, the center of mass position of the target in the figure is w c = M 10 M 00 ,   h c = M 01 M 00 .
The depth map is used to obtain distance information corresponding to the pixel point of the center of mass. This is combined with the imaging principle of the camera to solve the position of the target center of mass under the camera coordinate system. The image coordinate system (W, H) is defined in Figure 3 with the upper left corner of the image as the origin. The coordinate origin is the optical center position of the camera. The X-axis and Y-axis are parallel to the W-axis and H-axis of the image coordinate system, and the Z-axis is the optical axis of the camera’s camera coordinate system. The position of the target center of mass in the image under the camera system can be obtained from the imaging principle of the camera by the following equation:
X c = i w 0 f x D ( i , j ) Y c = j h 0 f y D ( i , j ) Z c = D ( i , j )
The depth value of pixel point ( i ,   j ) , designated as D ( i ,   j ) . ( f x ,   f y ,   w 0 ,   h 0 ) represent internal camera parameters. The desired angle of turn in the horizontal direction of the quadrupedal robot is ϕ = arctan ( Xc Zc ) .

3.3. CPG-Based Low-Level Gait Control

The coupling of leg and spinal motion is controlled using CPG. It is designed to mimic the central nervous system of an organism, enabling quadrupedal robots to autonomously generate gaits and adapt to different terrains and environments.
The connection scheme and functional allocation of the oscillators of the CPG are illustrated in Figure 1. The phase expression for the cpg oscillator is given by (6).
θ i ˙ = 2 π f + j ω i j sin θ j θ i ϑ i j
where θ i is the phase of the oscillator and f is the robot walking step frequency. The coupling between the oscillators is realized by the weights ω i j and the phase difference ϑ i j . The value of the phase difference is determined by the gait of the robot.
The robot’s spine is controlled by the oscillators 1–5, and the actual angle sent to the spinal joints is obtained by combining the amplitude and bias based on the phase calculated above, which is calculated as (7)
r ˙ i s p = a r R i s p r i s p x i ˙ = a x X i x i ϕ i = x i + r s p cos θ i
where r i s p and x i are state variables denoting the amplitude and bias of the oscillator, respectively, and R i s p and X i are the desired amplitude and desired bias of the oscillator, and the rate of convergence of the oscillator’s amplitude and bias is expressed in terms of positive gains a r and a x . When the robot steers, the spine is driven to bend in the desired direction by the bias term X i . The bias X i is calculated from the visual servo controller output ψ i . We take 4 maps at equal intervals during a gait cycle and find the bias by counting the distribution of pixel points.
The CPG parameters (e.g., the gains a r and a x ) were selected via manual tuning. We started from stable baseline values and then iteratively adjusted the parameters in simulation to (i) maintain stable oscillations, (ii) ensure smooth convergence of amplitude/bias without overshoot, and (iii) achieve low tracking error while avoiding foot slippage and excessive joint excursions. The final parameter set was chosen based on the best overall trade-off across these criteria.
In this paper, we utilize 6th to 9th oscillators to facilitate the generation of foot trajectories. Subsequently, the joint commands are derived through inverse kinematics. To align the oscillator state with the selected gait, we transform the oscillator phase by the ratio of the gait support phase to the swing phase to obtain θ . Figure 4 illustrates the spine posture and foot trajectory of the robot during a turn. The coordinates of the foot end position are as follows:
p ¯ i , x = d step ± d cos θ i cos ψ i p ¯ i , y = d step ± d cos θ i sin ψ i
p ¯ i , z = h + g c sin θ i if α i 0 , π 2 h + g p sin θ i otherwise
where g c is the maximum ground clearance achieved during the swing phase, while g p denotes the maximum ground penetration attained during the stance phase. The step length and robot height are denoted as d s t e p and h, respectively. The compensation amount d is related to the steering angle. The plus and minus signs, respectively, indicate the outer foot trajectory and the inner foot trajectory.

4. Experimental Results

4.1. Image Segmentation

4.1.1. Experimental Setup

The evaluation datasets utilized in our study include five well-established datasets: ECSSD [37], PASCAL-S [38], DUT-OMRON [39], HKU-IS [40], and DUTS-TE [41]. For model training, we consistently employ the DUTS-TR dataset [41] across all experiments, following established practices in image segmentation research.
Our model was trained for 60 rounds in batches of 30, and we selected the optimizer with a learning rate of 0.005, momentum of 0.9, and weight decay of 5 × 10 5 . The image input size was resized to 384 × 384 for both training and testing. The detailed hyperparameters and preprocessing steps are summarized in Table 1.
To assess the effectiveness of various methods, we utilize three commonly used metrics: the F-measure score ( F β ), the mean absolute error ( M A E ), and the S-measure score ( S α ). ( F β ) is calculated as follows:
F β = ( 1 + β 2 ) × Precision × Recall β 2 × Precision + Recall .
To impose a higher weight for accuracy, we set β 2 to 0.3. At the pixel level, M A E evaluates the average absolute difference between the predicted image P and the labeled image L.
MAE = 1 W × H x = 1 W y = 1 H | P ( x , y ) L ( x , y ) | ,
where the width and height of the image are denoted by W and H, respectively. The S-measure ( S α ) integrates both object-aware ( S o ) and region-aware ( S r ) structural similarity components, and is calculated as follows:
S α = γ S o + ( 1 γ ) S r ,
where γ is 0.5 as is commonly done.
The loss function utilized in this paper combines an intersection-over-union (IoU) loss with a binary cross-entropy loss (BCE): l = l i o u + l b c e . Because of its excellent robustness, the binary cross-entropy (BCE) loss function is widely used in binary classification and is obtained by calculating the pixel-by-pixel loss of the image:
l b c e ( p , l ) = 1 n k = 1 n l k log p k + 1 l k log 1 p k
p and l stand for the predicted image and label, respectively. k is the index of the pixel and n is the number of pixels in x. In contrast to the BCE loss function, which emphasizes differences at the pixel level, the IoU loss considers the overall graph similarity, and its definition is as follows:
l iou ( p , l ) = 1 k = 1 n l k p k k = 1 n l k + p k l k p k .

4.1.2. Comparisons to the State of the Art

We compared the proposed image segmentation method with 22 state-of-the-art approaches, including PAGR [20], DGRL [13], PiCANet [19], MLMS [42], PAGE [21], ICTB [14], CPD [15], BASNet [43], PoolNet [16], CSNet [44], GateNet [17], MINet [45], ITSD [46], VST [47], MSFNet [48], CII [49], PoolNet+ [50], DCN [51], DNA [52], RCSB [53], PriorNet [7], and NASAL [54]. To ensure a fair comparison, we either utilize saliency maps shared by the authors or compute their released models. We then quantitatively compare the obtained results by calculating the F-measure score F β , the S-measure score S α and the mean absolute error (MAE) of our method alongside the other methods. Table 2 presents the results of the other advanced measurement methods mentioned. On the ECSSD dataset, our method achieves the highest F β (0.952) and the lowest MAE (0.028), while maintaining a high S α value of 0.933. These results suggest enhanced capacity for capturing fine details and complex object structures, particularly in cluttered scenes. Similarly, on PASCAL-S, our model maintains leading performance, with minimized MAE and competitive F β and S α values, indicating improved robustness in handling occlusion and challenging backgrounds. Performance on HKU-IS further highlights the model’s generalization capabilities, recording an F β of 0.898, MAE of 0.031, and S α of 0.929, surpassing comparative methods across all metrics. On more challenging datasets such as DUT-OMRON and DUTS-TE, our method maintains its advantages. It shows significant improvement of 1.1 % and 0.8 % compared with the famous PoolNet+ model [50], which confirms its effectiveness in delineating object boundaries under complex scenes. Our model achieves leading performance in salient object detection, owing to its unique architecture that combines multi-scale feature interaction with global information guidance. This design enhances detail preservation while maintaining accurate global context.
In Figure 5, we present example saliency maps generated by our method. The input images used for the visual comparison are taken from standard datasets (not from the simulation camera). These maps demonstrate our method’s ability to produce accurate results with clear boundaries and uniform highlights.

4.1.3. Speed Analysis

In this speed analysis experiment, we evaluate inference latency when processing a 384 × 384 input image on two hardware platforms (one station with a single NVIDIA RTX-4090 GPU and one laptop with a single NVIDIA RTX-2060 GPU). Our method is built by adding lightweight modules on top of the FPN baseline, and we use this FPN implementation as the direct baseline for comparison. Although the proposed method achieves substantially better segmentation and tracking performance in the previous subsection, it introduces only a marginal inference overhead over FPN (about 1.74 ms on the station and 1.11 ms on the laptop, as shown in Table 3). This result indicates that the performance gains are obtained with minimal additional computational cost on both devices. Taken together, the strong perception accuracy and low computational overhead suggest that the proposed model is well suited for practical deployment in real-world robotic navigation tasks.

4.2. Low-Level Controller

In this paper, we use the Mujoco as the simulator. The robot physical parameters (including the 3-DOF legs and the 5-DOF spine) and the MuJoCo simulation configuration follow the setup reported by Horvat et al. [57]. To assess the effectiveness of this method in improving robot locomotion, we compared it with the spineless control and open-loop methods of tracking the straight line and the sine line.
Here, the open-loop control method applies trigonometric signals to the spine joints; the signals are the same as the step frequency. We evaluated the robot’s agility by letting it walk the same number of steps under different methods. Figure 6a shows the experimental results. During the experiment, the robot walked for 10 gait cycles while tracking a straight-line trajectory. The results indicate that the robot’s motion performance is poor without spine coordination (orange line). However, open-loop spine control (green line) greatly improves the motion performance, resulting in a travel distance of more than 2.5 m . The method proposed in this paper (blue line) further increases the travel distance and significantly reduces the tracking error compared with open-loop control.
To compare turning ability across methods, the robot tracked a segment of a sinusoidal trajectory while taking the same number of steps. The experimental results in Figure 6b indicate that the spineless control method (orange line) performs poorly in coordination and in tracking the target trajectory. Similarly, the open-loop method (green line) does not capture the interaction between the spine and the legs, resulting in larger errors at large steering angles. Our method (blue line) exhibits superior motion performance, advancing farther while taking the same number of steps. Furthermore, the tracking error is smaller, particularly during the (0,2m) convergence phase towards the target trajectory. This demonstrates the effectiveness of our method in improving the locomotion ability of the robot.

4.3. Visual Tracking

To validate the performance of the visual tracking framework presented in this paper, we utilize a moving Epuck robot as the tracking target.
In the simulation, the Epuck robot moves at a speed of 0.5 m/s, while the quadrupedal robot continuously captures and tracks it using visual input. The simulation scenario for the tracking experiment is illustrated in Figure 7. The quadrupedal robot gradually reduces the distance to the target robot, eventually catching up and maintaining the track.
As illustrated in Figure 8a, the quadrupedal robot is positioned at the origin to track the Epuck robot at (2.5, 0) and approaches the Epuck robot to within 0.5   m . At the initial stage, the Epuck robot moves in the positive direction of the Y-axis, while the quadrupedal robot moves in the positive direction of the X-axis.
There is a significant difference in the direction of motion between the two. According to the instructions of the vision-tracking framework, the quadrupedal robot narrows the deviation and successfully tracks the Epuck robot’s trajectory. The quadrupedal robot was again tasked with continuously tracking the target, as illustrated in Figure 8b. The tracking framework successfully enabled the quadrupedal robot to capture the target and track its trajectory. The target also altered its direction of motion along the way, and the tracking framework provided the correct instructions for the quadrupedal robot to make timely adjustments.
Throughout the experiment, the target robot altered its direction of motion multiple times. The tracking framework provided the quadrupedal robot with the necessary instructions to make timely adjustments, ensuring continuous and accurate tracking of the Epuck robot’s trajectory. This demonstrated the robustness and effectiveness of the visual tracking framework in dynamic and changing environments.

5. Discussion

One critical consideration for the proposed framework is its applicability to real-world environments, given that the current validation is conducted within the MuJoCo simulation. However, it is important to note that the visual perception module (the cascaded transformer-based network) is trained on large-scale, standard real-world datasets. These datasets encompass a wide spectrum of environmental challenges, such as drastic lighting variations, complex shadows, and diverse surface textures, which are common in unconstrained real-world scenarios. By learning from these diverse real-world samples, the model develops high-level semantic understanding and boundary-awareness that are inherently robust to the environmental factors mentioned. While the MuJoCo simulator provides a controlled environment for testing the closed-loop gait adaptation logic, the perception-action backbone is built upon real-world visual features, significantly bridging the gap between simulation and reality. Future work will further explore the Sim-to-Real transition by deploying the system on physical quadrupedal platforms to account for complex hardware-level dynamics and sensor noise.

6. Conclusions

In this work, we present a visual tracking control framework for quadrupedal robots. A cascade interaction network is introduced to enhance the information interaction ability, which improves the accuracy and efficiency of the image segmentation model. Furthermore, we convert the segmented target image into the desired deflection angle of the quadruped robot. For low-level gait control, this paper introduces a central pattern generator to parameterize quadruped motion and upper-level commands, enabling flexible switching between different motion modes. The effectiveness of the image segmentation proposed in this paper is verified through experiments and comparisons on standard datasets. And we verify the value of the framework in quadrupedal robot tasks.

Author Contributions

Conceptualization, H.X. and W.W.; methodology, H.X., G.M. and W.W.; software, H.X. and G.M.; validation, H.X., G.M. and W.W.; formal analysis, H.X. and G.M.; investigation, H.X. and W.W.; resources, G.M. and W.W.; data curation, H.X. and G.M.; writing—original draft preparation, H.X.; writing—review and editing, H.X., G.M. and W.W.; visualization, H.X. and G.M.; supervision, W.W. and G.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Shenzhen Science and Technology Program, China (No. GXWD2023 1129140908002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, T.; Wu, Z.; Wang, D. Visual perception generalization for vision-and-language navigation via meta-learning. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 5193–5199. [Google Scholar] [CrossRef]
  2. Zhao, X.; Wang, L.; Zhang, Y.; Han, X.; Deveci, M.; Parmar, M. A review of convolutional neural networks in computer vision. Artif. Intell. Rev. 2024, 57, 99. [Google Scholar] [CrossRef]
  3. Zhang, Y.; Wen, L.; Hong, L.; Zhang, L.; Guo, Q.; Li, S.; Bing, Z.; Knoll, A. Safety-Critical Control with Saliency Detection for Mobile Robots in Dynamic Multi-Obstacle Environments. In Proceedings of the 2025 IEEE International Conference on Robotics and Automation (ICRA); IEEE: Piscataway, NJ, USA, 2025; pp. 7756–7762. [Google Scholar]
  4. Liu, Z.; Liu, Y.; Fang, Y.; Guo, X. Autonomous Visual Navigation with Head Stabilization Control for a Salamander-Like Robot. IEEE/ASME Trans. Mechatron. 2025. early access. [Google Scholar]
  5. Roberts, R.; Ta, D.N.; Straub, J.; Ok, K.; Dellaert, F. Saliency detection and model-based tracking: A two part vision system for small robot navigation in forested environment. In Proceedings of the Unmanned Systems Technology XIV; SPIE: Bellingham, WA, USA, 2012; Volume 8387, pp. 306–317. [Google Scholar]
  6. Zhang, D.; Han, J.; Zhang, Y.; Xu, D. Synthesizing supervision for learning deep saliency network without human annotation. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1755–1769. [Google Scholar] [CrossRef] [PubMed]
  7. Zhu, G.; Li, J.; Guo, Y. PriorNet: Two Deep Prior Cues for Salient Object Detection. IEEE Trans. Multimed. 2024, 26, 5523–5535. [Google Scholar] [CrossRef]
  8. Zheng, Q.; Zheng, L.; Deng, J.; Li, Y.; Shang, C.; Shen, Q. Transformer-based hierarchical dynamic decoders for salient object detection. Knowl.-Based Syst. 2023, 282, 111075. [Google Scholar] [CrossRef]
  9. Zhornyak, L.; Emami, M.R. Gait optimization for quadruped rovers. Robotica 2020, 38, 1263–1287. [Google Scholar] [CrossRef]
  10. Gangapurwala, S.; Mitchell, A.; Havoutis, I. Guided constrained policy optimization for dynamic quadrupedal robot locomotion. IEEE Robot. Autom. Lett. 2020, 5, 3642–3649. [Google Scholar] [CrossRef]
  11. Lee, G.; Tai, Y.W.; Kim, J. Deep Saliency with Encoded Low level Distance Map and High Level Features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2016. [Google Scholar]
  12. Xu, J.; Liu, Z.A.; Hou, Y.K.; Zhen, X.T.; Shao, L.; Cheng, M.M. Pixel-Level Non-local Image Smoothing With Objective Evaluation. IEEE Trans. Multimed. 2021, 23, 4065–4078. [Google Scholar] [CrossRef]
  13. Wang, T.; Zhang, L.; Wang, S.; Lu, H.; Yang, G.; Ruan, X.; Borji, A. Detect Globally, Refine Locally: A Novel Approach to Saliency Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2018; pp. 3127–3135. [Google Scholar]
  14. Wang, W.; Shen, J.; Cheng, M.M.; Shao, L. An Iterative and Cooperative Top-Down and Bottom-Up Inference Network for Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  15. Wu, Z.; Su, L.; Huang, Q. Cascaded Partial Decoder for Fast and Accurate Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  16. Liu, J.J.; Hou, Q.; Cheng, M.M.; Feng, J.; Jiang, J. A Simple Pooling-Based Design for Real-Time Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  17. Zhao, X.; Pang, Y.; Zhang, L.; Lu, H.; Zhang, L. Suppress and Balance: A Simple Gated Network for Salient Object Detection. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2020. [Google Scholar]
  18. Chang, Y.; Liu, Z.; Wu, Y.; Fang, Y. Deep-Learning-Based Automated Morphology Analysis with Atomic Force Microscopy. IEEE Trans. Autom. Sci. Eng. 2024, 21, 7662–7673. [Google Scholar] [CrossRef]
  19. Liu, N.; Han, J.; Yang, M.H. PiCANet: Learning Pixel-wise Contextual Attention for Saliency Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2018; pp. 3089–3098. [Google Scholar]
  20. Zhang, X.; Wang, T.; Qi, J.; Lu, H.; Wang, G. Progressive Attention Guided Recurrent Network for Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2018; pp. 714–722. [Google Scholar]
  21. Wang, W.; Zhao, S.; Shen, J.; Hoi, S.C.; Borji, A. Salient Object Detection with Pyramid Attention and Salient Edges. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  22. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2022; pp. 10684–10695. [Google Scholar]
  23. Liu, Z.; Liu, Y.; Fang, Y. Diffusion Model-Based Path Follower for a Salamander-Like Robot. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 14399–14413. [Google Scholar] [CrossRef]
  24. Zhang, J.; Fan, D.P.; Dai, Y.; Anwar, S.; Saleh, F.; Aliakbarian, S.; Barnes, N. Uncertainty inspired RGB-D saliency detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5761–5779. [Google Scholar] [CrossRef] [PubMed]
  25. Wang, C.; Dong, S.; Zhao, X.; Papanastasiou, G.; Zhang, H.; Yang, G. SaliencyGAN: Deep learning semisupervised salient object detection in the fog of IoT. IEEE Trans. Ind. Inform. 2019, 16, 2667–2676. [Google Scholar] [CrossRef]
  26. Sun, K.; Chen, Z.; Lin, X.; Sun, X.; Liu, H.; Ji, R. Conditional diffusion models for camouflaged and salient object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 2833–2848. [Google Scholar] [CrossRef]
  27. Yang, Y.; Zhang, T.; Coumans, E.; Tan, J.; Boots, B. Fast and efficient locomotion via learned gait transitions. In Proceedings of the Conference on Robot Learning; PMLR: Cambridge, MA, USA, 2022; pp. 773–783. [Google Scholar]
  28. Lee, J.; Kim, J.; Ubellacker, W.; Molnar, T.G.; Ames, A.D. Safety-critical Control of Quadrupedal Robots with Rolling Arms for Autonomous Inspection of Complex Environments. arXiv 2023, arXiv:2312.07778. [Google Scholar] [CrossRef]
  29. Liu, K.; Dong, L.; Tan, X.; Zhang, W.; Zhu, L. Optimization-Based Flocking Control and MPC-Based Gait Synchronization Control for Multiple Quadruped Robots. IEEE Robot. Autom. Lett. 2024, 9, 1929–1936. [Google Scholar] [CrossRef]
  30. Ding, Y.; Pandala, A.; Li, C.; Shin, Y.H.; Park, H.W. Representation-Free Model Predictive Control for Dynamic Motions in Quadrupeds. IEEE Trans. Robot. 2021, 37, 1154–1171. [Google Scholar] [CrossRef]
  31. Bjelonic, M.; Grandia, R.; Harley, O.; Galliard, C.; Zimmermann, S.; Hutter, M. Whole-Body MPC and Online Gait Sequence Generation for Wheeled-Legged Robots. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); IEEE: Piscataway, NJ, USA, 2021; pp. 8388–8395. [Google Scholar] [CrossRef]
  32. Wang, J.; Hu, C.; Zhu, Y. CPG-based hierarchical locomotion control for modular quadrupedal robots using deep reinforcement learning. IEEE Robot. Autom. Lett. 2021, 6, 7193–7200. [Google Scholar] [CrossRef]
  33. Sleiman, J.P.; Farshidian, F.; Minniti, M.V.; Hutter, M. A unified mpc framework for whole-body dynamic locomotion and manipulation. IEEE Robot. Autom. Lett. 2021, 6, 4688–4695. [Google Scholar] [CrossRef]
  34. Tsounis, V.; Alge, M.; Lee, J.; Farshidian, F.; Hutter, M. Deepgait: Planning and control of quadrupedal gaits using deep reinforcement learning. IEEE Robot. Autom. Lett. 2020, 5, 3699–3706. [Google Scholar] [CrossRef]
  35. Bellegarda, G.; Ijspeert, A. CPG-RL: Learning central pattern generators for quadruped locomotion. IEEE Robot. Autom. Lett. 2022, 7, 12547–12554. [Google Scholar] [CrossRef]
  36. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE International Conference on Computer Vision; IEEE: Piscataway, NJ, USA, 2021; pp. 10012–10022. [Google Scholar]
  37. Yan, Q.; Xu, L.; Shi, J.; Jia, J. Hierarchical saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2013; pp. 1155–1162. [Google Scholar]
  38. Li, Y.; Hou, X.; Koch, C.; Rehg, J.M.; Yuille, A.L. The secrets of salient object segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2014; pp. 280–287. [Google Scholar]
  39. Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; Yang, M.H. Saliency detection via graph-based manifold ranking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2013; pp. 3166–3173. [Google Scholar]
  40. Li, G.; Yu, Y. Visual saliency based on multiscale deep features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2015; pp. 5455–5463. [Google Scholar]
  41. Wang, L.; Lu, H.; Wang, Y.; Feng, M.; Wang, D.; Yin, B.; Ruan, X. Learning to detect salient objects with image-level supervision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2017; pp. 136–145. [Google Scholar]
  42. Wu, R.; Feng, M.; Guan, W.; Wang, D.; Lu, H.; Ding, E. A Mutual Learning Method for Salient Object Detection with Intertwined Multi-Supervision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  43. Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagersand, M. BASNet: Boundary-Aware Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  44. Gao, S.H.; Tan, Y.Q.; Cheng, M.M.; Lu, C.; Chen, Y.; Yan, S. Highly Efficient Salient Object Detection with 100K Parameters. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2020. [Google Scholar]
  45. Pang, Y.; Zhao, X.; Zhang, L.; Lu, H. Multi-Scale Interactive Network for Salient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2020; pp. 9413–9422. [Google Scholar]
  46. Zhou, H.; Xie, X.; Lai, J.H.; Chen, Z.; Yang, L. Interactive Two-Stream Decoder for Accurate and Fast Saliency Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2020; pp. 9141–9150. [Google Scholar]
  47. Liu, N.; Zhang, N.; Wan, K.; Shao, L.; Han, J. Visual Saliency Transformer. In Proceedings of the IEEE International Conference on Computer Vision, October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 4722–4732. [Google Scholar]
  48. Zhang, M.; Liu, T.; Piao, Y.; Yao, S.; Lu, H. Auto-MSFNet: Search Multi-scale Fusion Network for Salient Object Detection. In Proceedings of the ACM Multimedia Conference; ACM: New York, NY, USA, 2021. [Google Scholar]
  49. Liu, J.J.; Liu, Z.A.; Peng, P.; Cheng, M.M. Rethinking the U-shape structure for salient object detection. IEEE Trans. Image Process. 2021, 30, 9030–9042. [Google Scholar] [CrossRef] [PubMed]
  50. Liu, J.J.; Hou, Q.; Liu, Z.A.; Cheng, M.M. Poolnet+: Exploring the potential of pooling for salient object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 887–904. [Google Scholar] [CrossRef] [PubMed]
  51. Wu, Z.; Su, L.; Huang, Q. Decomposition and Completion Network for Salient Object Detection. IEEE Trans. Image Process. 2021, 30, 6226–6239. [Google Scholar] [CrossRef]
  52. Yao, Z.; Wang, L. Boundary Information Progressive Guidance Network for Salient Object Detection. IEEE Trans. Multimed. 2022, 24, 4236–4249. [Google Scholar] [CrossRef]
  53. Ke, Y.Y.; Tsubono, T. Recursive contour-saliency blending network for accurate salient object detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision; IEEE: Piscataway, NJ, USA, 2022; pp. 2940–2950. [Google Scholar]
  54. Liu, Z.A.; Liu, J.J. Towards efficient salient object detection via U-shape architecture search. Knowl.-Based Syst. 2025, 318, 113515. [Google Scholar] [CrossRef]
  55. Liu, Y.; Cheng, M.M.; Zhang, X.Y.; Nie, G.Y.; Wang, M. DNA: Deeply Supervised Nonlinear Aggregation for Salient Object Detection. IEEE Trans. Cybern. 2022, 52, 6131–6142. [Google Scholar] [CrossRef]
  56. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2017; pp. 2117–2125. [Google Scholar]
  57. Horvat, T.; Melo, K.; Ijspeert, A.J. Spine Controller for a Sprawling Posture Robot. IEEE Robot. Autom. Lett. 2017, 2, 1195–1202. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed visual tracking framework. The on-board camera image is segmented to obtain the target mask; a visual servoing module fuses the mask with depth to estimate target pose and compute a steering command; a CPG-based controller then generates coordinated spine/leg trajectories for tracking. The numbers (1–9) in the Central Pattern Generator (CPG) represent interconnected nodes that facilitate rhythmic signal propagation. The different background colors of the panels (green for Image Segmentation, beige for Visual Servo and blue for CPG) indicate the distinct stages of the processing pipeline.
Figure 1. Overview of the proposed visual tracking framework. The on-board camera image is segmented to obtain the target mask; a visual servoing module fuses the mask with depth to estimate target pose and compute a steering command; a CPG-based controller then generates coordinated spine/leg trajectories for tracking. The numbers (1–9) in the Central Pattern Generator (CPG) represent interconnected nodes that facilitate rhythmic signal propagation. The different background colors of the panels (green for Image Segmentation, beige for Visual Servo and blue for CPG) indicate the distinct stages of the processing pipeline.
Biomimetics 11 00234 g001
Figure 2. Overview of the proposed segmentation network. A Swin-Transformer encoder extracts multi-scale features, which are progressively fused via cascaded cross-resolution interactions. GIGM uses higher-level context to calibrate lower-level features, and FA blocks refine the representations to produce the final saliency map.
Figure 2. Overview of the proposed segmentation network. A Swin-Transformer encoder extracts multi-scale features, which are progressively fused via cascaded cross-resolution interactions. GIGM uses higher-level context to calibrate lower-level features, and FA blocks refine the representations to produce the final saliency map.
Biomimetics 11 00234 g002
Figure 3. Desired deflection angle calculation.
Figure 3. Desired deflection angle calculation.
Biomimetics 11 00234 g003
Figure 4. Robot turning motion.
Figure 4. Robot turning motion.
Biomimetics 11 00234 g004
Figure 5. Visual comparison of saliency maps with state-of-the-art methods. From left to right: Input image, Ground truth, Ours, DNA, CII, MSFNet, VST and ITSD. Our approach consistently produces the best results.
Figure 5. Visual comparison of saliency maps with state-of-the-art methods. From left to right: Input image, Ground truth, Ours, DNA, CII, MSFNet, VST and ITSD. Our approach consistently produces the best results.
Biomimetics 11 00234 g005
Figure 6. Comparison of motion performance.
Figure 6. Comparison of motion performance.
Biomimetics 11 00234 g006
Figure 7. Simulation scenario of quadrupedal robot tracking task.
Figure 7. Simulation scenario of quadrupedal robot tracking task.
Biomimetics 11 00234 g007
Figure 8. Robot tracking performance.
Figure 8. Robot tracking performance.
Biomimetics 11 00234 g008
Table 1. Training hyperparameters and data preprocessing.
Table 1. Training hyperparameters and data preprocessing.
ItemSetting
Training epochs (rounds)60
Batch size30
OptimizerSGD
Initial learning rate0.005
Learning rate scheduleFixed (no decay)
Momentum0.9
Weight decay 5 × 10 5
Input resize384 × 384 (train/test)
NormalizationStandard dataset normalization
Data augmentationNone
Table 2. Comparisons of our method with other state-of-the-art methods on five popular SOD benchmarks.
Table 2. Comparisons of our method with other state-of-the-art methods on five popular SOD benchmarks.
MethodECSSDPASCAL-SDUT-OMRONHKU-ISDUTS-TE
F β  ↑MAE S α F β  ↑MAE S α F β  ↑MAE S α F β  ↑MAE S α F β  ↑MAE S α
PAGR [20]0.9270.0610.8890.8470.0890.8220.7710.0710.7750.9190.0470.8890.8540.0550.839
DGRL [13]0.9220.0410.9030.8440.0720.8360.7740.0620.8060.9100.0360.8950.8280.0490.842
PiCANet [19]0.9350.0470.9170.8640.0750.8540.8200.0640.8300.9200.0440.9040.8630.0500.868
MLMS [42]0.9300.0450.9110.8530.0740.8440.7930.0630.8090.9220.0390.9070.8540.0480.862
PAGE [21]0.9310.0420.9120.8480.0760.8420.7910.0620.8250.9200.0360.9040.8380.0510.855
ICTB [14]0.9380.0410.9180.8550.0710.8500.8110.0600.8370.9250.0370.9090.8550.0430.865
CPD [15]0.9390.0370.9180.8590.0710.8480.7960.0560.8250.9250.0340.9070.8650.0430.869
BASNet [43]0.9420.0370.9160.8570.0760.8380.8110.0570.8360.9300.0330.9080.8600.0470.866
PoolNet [16]0.9440.0390.9210.8650.0750.8500.8300.0550.8360.9340.0320.9170.8860.0400.883
CSNet [44]0.9440.0380.9210.8660.0730.8510.8210.0550.8310.9300.0330.9110.8810.0400.879
GateNet [17]0.9460.0400.9200.8770.0680.8580.8310.0550.8380.9350.0330.9150.8890.0400.885
MINet [45]0.9470.0340.9250.8740.0640.8560.8260.0560.8330.9360.0280.9200.8880.0370.884
ITSD [46]0.9470.0350.9250.8710.0660.8590.8230.0610.8400.9330.0310.9160.8830.0410.885
VST [47]0.9510.0340.9320.8750.0620.8720.8290.0580.8500.9420.0300.9290.8910.0370.896
MSFNet [48]0.9430.0330.9150.8650.0610.8520.8240.0500.8320.9300.0270.9090.8810.0340.877
CII [49]0.9500.0340.9260.8820.0620.8650.8310.0540.8390.9390.0290.9200.8900.0360.888
PoolNet+ [50]0.9490.0400.9250.8790.0680.8640.8310.0560.8420.9410.0340.9210.8940.0390.890
DCN [51]0.9520.0310.9280.8720.0620.8610.8230.0510.8450.9400.0270.9220.8940.0350.891
DNA [55]0.9400.0430.9150.8550.0790.8370.8030.0630.8180.9270.0360.9050.8730.0460.860
RCSB [53]0.9450.0330.9220.8790.0590.8600.8490.0490.8350.9390.0270.9180.8970.0350.881
PriorNet [7]0.9530.0310.9310.8810.0590.8690.8390.0510.8490.9400.0290.9200.9010.0330.897
NASAL [54]0.9250.0520.9040.8360.0920.8250.8000.0690.8180.9130.0440.8980.8330.0600.841
Ours 0.9520.0280.9330.8880.0540.8790.8420.0490.8580.9430.0250.9290.8980.0310.900
Note: The “Ours” row is highlighted in bold to emphasize the performance of the proposed method in this study. The arrows (↑) next to the relevant metrics indicate that higher values are better, whereas the arrows (↓) indicate that lower values are better.
Table 3. Comparison of inference latency between the proposed network and FPN on different devices.
Table 3. Comparison of inference latency between the proposed network and FPN on different devices.
MethodStationLaptop
FPN [56]13.48 ms44.15 ms
Proposed15.22 ms45.26 ms
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiao, H.; Ma, G.; Wu, W. Image Segmentation-Guided Visual Tracking on a Bio-Inspired Quadruped Robot. Biomimetics 2026, 11, 234. https://doi.org/10.3390/biomimetics11040234

AMA Style

Xiao H, Ma G, Wu W. Image Segmentation-Guided Visual Tracking on a Bio-Inspired Quadruped Robot. Biomimetics. 2026; 11(4):234. https://doi.org/10.3390/biomimetics11040234

Chicago/Turabian Style

Xiao, Hewen, Guangfu Ma, and Weiren Wu. 2026. "Image Segmentation-Guided Visual Tracking on a Bio-Inspired Quadruped Robot" Biomimetics 11, no. 4: 234. https://doi.org/10.3390/biomimetics11040234

APA Style

Xiao, H., Ma, G., & Wu, W. (2026). Image Segmentation-Guided Visual Tracking on a Bio-Inspired Quadruped Robot. Biomimetics, 11(4), 234. https://doi.org/10.3390/biomimetics11040234

Article Metrics

Back to TopTop