Next Article in Journal
Influence and Quantitative Analysis of Different Parameters on Adaptive Cycle Fan Performance
Previous Article in Journal
UAV Sensor Data Fusion for Localization Using Adaptive Multiscale Feature Matching Mechanisms Under GPS-Deprived Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LisseMars: A Lightweight Semantic Segmentation Model for Mars Helicopter

1
CVIR Lab, Changchun University of Science and Technology, Changchun 130022, China
2
Shanghai Aerospace Control Technology Institute, Shanghai 201109, China
3
School of Artificial Intelligence, Jilin University, Changchun 130012, China
4
School of Astronautics, Harbin Institute of Technology, Harbin 150001, China
5
Beijing Institute of Control Engineering, Beijing 100190, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Aerospace 2025, 12(12), 1049; https://doi.org/10.3390/aerospace12121049
Submission received: 9 October 2025 / Revised: 14 November 2025 / Accepted: 21 November 2025 / Published: 25 November 2025
(This article belongs to the Section Astronautics & Space Science)

Abstract

With the continuous deepening of Mars exploration missions, the Mars helicopter has become a key platform for acquiring high-resolution near-ground imagery. However, accurate semantic segmentation of the Martian surface remains challenging due to complex terrain morphology, sandstorm interference, and the limited onboard computational resources that restrict real-time processing. Existing models either introduce high computational overhead unsuitable for deployment on Mars aerial platforms or fail to jointly capture fine-grained local texture and global contextual structure information. To address these limitations, we propose LisseMars, a lightweight semantic segmentation network designed for efficient onboard perception. The model integrates a Window Movable Attention (WMA) module for enhanced global context extraction and a multi-convolutional feedforward module (CFFN) to strengthen local detail representation. A Dynamic Polygon Convolution (DPC) module is further introduced to improve segmentation performance on geometrically heterogeneous objects, while a Group Fusion Module (GFM) enables effective multi-scale semantic integration. Extensive experiments are conducted on both real Tianwen-1 Mars helicopter imagery and synthetic datasets. The results show that our method achieved a mean IoU of 78.56% with only 0.12 MB of model parameters, validating the effectiveness of the proposed framework. The real-time performance of proposed method on edge device deployment further demonstrate potential application for real Mars airborne missions.

1. Introduction

In Mars exploration, researchers are seeking new technologies to enable smooth navigation and the discovery of scientific objects across the complex Martian terrain [1,2,3]. The success of the Mars helicopter “Ingenuity” [4,5] has reached a milestone as the first powered flight on another planet. It provides unprecedented perspectives and data for the Mars rover. The exploration of the Martian surface is crucial and indispensable to the Mars exploration mission. By processing image data, different scientific objects on the surface of Mars can be quickly identified [6,7] and therefore will effectively optimise the trajectory planning and exploration of the Mars rover [8,9,10].
With the continuous development of deep learning, neural networks have achieved remarkable success in the field of computer vision, particularly in semantic segmentation [11], object detection [12,13,14,15], 3D reconstruction [16,17,18] and object classification [19,20,21] tasks. For extraterrestrial exploration missions, semantic segmentation technology is crucial as it enables accurate identification and distinction of various surface objects, such as rocks, sand, and gravel. This capability not only aids in enhancing the navigation accuracy of Mars rovers but also has significant implications for effectively discovering scientific objects. However, Mars has diverse and complex unstructured terrain with harsh sandstorm environments, making it challenging to clearly distinguish the objects from the surrounding ground. As a result, it becomes essential for semantic segmentation networks to capture global contextual information so as to accurately explore the scene. Some studies [22,23] introduce pyramid structures embedded in CNNs to capture global contextual information at different scales. However, such methods typically extract local features from CNNs and then aggregate to global information, rather than directly encoding the global context. Consequently, in case of complex Martian scenes, CNN-based models often struggle to capture clear global information.
A type of solution is to design the transformer model [24], which can sufficiently extract global information. However, transformer-based methods still have some limitations, especially in the Martian environment. The original transformer architecture relies primarily on global contextual information and pays relatively limited attention to local details. The surface of Mars is scattered with various objects, which requires more precise processing of object locations and associated complex details, making the limited local focus of transformers even more problematic. To address this challenge, some researchers [25,26] have introduced convolutional position encoding and relative position encoding to improve the model’s ability to capture spatial and sequential dependencies. However, these methods may not fully express the spatial structure of the objects, which could adversely affect the model’s performance in complex, dynamic environments.
The deployment of sophisticated models on Mars rovers and helicopters is significantly constrained by their limited computational resources, necessitating a careful trade-off between model complexity and accuracy. Existing CNN-based or heavy Transformer-based models often fail to jointly preserve fine local boundaries and global contextual terrain cues, and many are unsuitable for deployment on Mars aerial platforms due to limited onboard computational resources. To this end, in this paper we proposed LisseMars, a lightweight Transformer-based segmentation network tailored for the unstructured, low-texture, and scale-varying Martian environments. LisseMars is built upon a U-shaped encoder–decoder architecture with skip connections. For the encoder, we proposed a Window Movable Attention (WMA) mechanism, which combines the local window partitioning with cross-window interaction to adaptively balance global context modeling and local detail preservation. This design will significantly enhance the robustness of semantic segmentation against the low-contrast and complex geological textures of Mars environment. Furthermore, to better handle rocks with geometrically heterogeneous structures, we design a Dynamic Polygon Convolution (DPC) that adjusts its sampling geometry according to local shape variations, allowing the network to more precisely characterize uneven, non-uniform object boundaries. For the decoder, we employ an enhanced Group Fusion Module (GFM), derived from UperNet [27], to hierarchically integrate multi-scale features and maintain semantic consistency across regions of varying size. As illustrated in Figure 1, the proposed LisseMars achieves a favorable balance between segmentation accuracy and model efficiency, demonstrating clear advantages over existing state-of-the-art methods and strong potential for real deployment in Mars exploration missions. The main contributions are as follows:
(1)
We proposed a lightweight semantic segmentation model (LisseMars) for the Mars helicopter, which can effectively segment objects on Mars.
(2)
We designed a novel Window Movable Attention Module (WMA), which enhances feature extraction and effectively promote the fusion of global and local information.
(3)
We developed a Dynamic Polygon Convolution (DPC) that effectively focuses on irregular and uneven local features, enabling accurate segmentation of polygonal structures.
Figure 1. m I o U values of current SOTA models on SynMars-Air. Our proposed LisseMars model outperforms all the baselines.
Figure 1. m I o U values of current SOTA models on SynMars-Air. Our proposed LisseMars model outperforms all the baselines.
Aerospace 12 01049 g001

2. Related Works

2.1. Deep Learning for Mars

Perceiving unstructured environments on Mars has always been a key factor in autonomous navigation for current Mars missions [28,29]. For the Mars helicopter, accurately assessing the environment and identifying scientific objects are prerequisites for future successfully completing large-scale Mars exploration missions. This can provide crucial supports for the rovers’ navigation and path planning, thereby improving overall exploration efficiency [30,31]. Edge-based detection [32] and region-based detection [33] have been widely applied in the past. However, due to the complex and dynamic environment of Mars, relying solely on traditional methods is no longer sufficient to meet the requirements of future Mars exploration missions. With the rapid advancement of artificial intelligence, deep learning is gaining increasing attention [17].
Compared with classical semantic segmentation methods, CNN-based methods have significant advantages in feature extraction. Ref. [13] developed a lightweight deep learning framework named mini-SDD to detect various types of landforms on Mars. Ref. [34] using multi-scale dilated convolution in the extended path, integrating with the U-Net network for accurate segmentation of objects on Mars. Ref. [35] proposed Mobile-DeepRFB, a lightweight segmentation framework for Martian terrain classification.
Although existing studies have made notable progress in Martian terrain segmentation, several limitations remain insufficiently addressed. Traditional CNN-based approaches struggled to capture multi-scale contextual relationships and long-range dependencies, which are crucial for delineating geometrically heterogeneous rock boundaries and low-texture surface. To overcome these issues, recent works have attempted to integrate Transformers with convolutional architectures. SegMarsViT [36] combined ViT module with CNN model to jointly extract local and global features; MarsFormer [37] enhanced remote contextual modeling while preserving local details. RockFormer [38] designed a layered encoder–decoder structure for precise rock segmentation. HASS [39] incorporated multi-scale attention modules to improve the segmentation consistency. Light4Mars [40] further designed a lightweight window attention mechanism for Mars rover to characterize complex, unstructured surface. However, despite these advancements, current methods still face significant challenges, including insufficient fine-grained geometric modeling, limited robustness in extreme multi-scale scenarios, high computational and memory overhead due to heavy global-attention modules, and difficulty achieving lightweight onboard deployment on Mars aerial or rover platforms with strict power and hardware constraints.

2.2. Mars Terrain Datasets

A high-quality dataset is the foundation of deep learning methods. Currently, there exist various types of unstructured terrain datasets. For Mars, there are three types of datasets: the real data captured by Mars rovers or helicopter, the physical simulation data, and synthetic simulation data. Ref. [41] established the first large-scale Mars landscape dataset, AI4Mars, using real images from Curiosity mission. Ref. [42] created the first real dataset for Martian rock segmentation, MarsData, and later refined it to MarsDataV2 [38]. Ref. [43] developed the first Mars panoramic dataset, called MarsScapes.
Physical simulation and synthetic simulation datasets also play a crucial role in supplementing real Martian data. These datasets provide controlled conditions and allow the generation of diverse scenarios, which are difficult or impossible to capture on Mars. Ref. [44] verified the feasibility of Mars helicopter navigation by simulating Martian terrain in the desert. Ref. [37] proposed a publicly available synthetic Martian surface dataset, SynMars, for rock segmentation and extended to SynMars-TW dataset [40] based on images of Zhurong Rover.
The aforementioned datasets have propelled advances in the progress of Mars exploration by Mars rovers. However, the viewpoint between the Rover and the helicopter is quite different, which leads to entirely distinct observations of the surrounding environment and objects. In the case of occlusion and partial visibility, the segmentation networks designed for Mars rovers may not be suitable for Mars helicopter. Differences in sensor types and data processing methods can lead to an incomplete understanding of the object’s position and size, affecting object segmentation. Currently, there are still few real Mars images captured from a helicopter. To this end, Ref. [45] proposed the first aerial perspective dataset of Mars, SynMars-Air. Sample images from these datasets are shown Figure 2.

3. Method

In this section, we first introduce the overall framework of our lightweight semantic segmentation model, LisseMars. Then, we provide a detailed description of the proposed Window Moving Attention Module (WMA), the Improved Feedforward Network (CFFN), and the Dynamic Polygon Convolution (DPC), as well as the Group Fusion Module (GFM).

3.1. Overrall Framework

LisseMars model balanced modularity and flexibility, aiming to provide a scalable system architecture. As shown in Figure 3, it followed the encoder-decoder structure of U-Net. The input image X H × W × 3 is first fed into the encoder, which includes four improved self-attention modules and a dynamic polygon convolution module (DPCM). Each improved attention module consists of a WMA and CFFN structure, which is repeated multiple times to obtain four hierarchical feature maps: X i R H i × W i × C i .
Subsequently, the obtained feature X i is processed through the Group Fusion Module (GFM), which plays a crucial role in fusing and enhancing multi-scale feature maps. The high-dimensional feature layer X 4 undergoes pooling operations at multiple scales in the High Dimensional Pooling Fusion Module and then merges with features from other layers in the GFM to obtain different output features: F i R H i × W i × C i . Finally, these features F i are sampled and fused, and the predicted semantic segmentation map is output.

3.2. Window Movable Attention (WMA)

The working principle of WMA is illustrated in Figure 4. In order to reduce computational costs while maintaining high accuracy, we adopted the window partitioning strategy to extract feature X i from feature X and project it onto the Q tensor, the K tensor, and the V tensor. During the process again, W h and W w represent the window pooling sizes in the vertical and horizontal directions, while C is the number of channels. The specific steps are as follows:
Q = L N X i W Q W h W w i = 0 W h 1 j = 0 W w 1 Q C ( i , j ) K = L N X i W K W h W w i = 0 W h 1 j = 0 W w 1 K C ( i , j ) V = L N ( X W V )
In this context, L N denotes Layer Normalization, while i and j represent the indices in the horizontal and vertical directions of the feature map, respectively. As shown in Figure 5c, we applied deep convolutions for feature aggregation to the Q, K, and V tensors to aggregate more global context. This enables the model to focus on the relationships between adjacent elements in the input sequence, thereby enhancing its ability to capture spatial features. The main process is as follows:
Q = D P C N ( r i , Q ) + D P C N 1 ( r i , Q ) + Q K = D P C N ( r i , K ) + D P C N 1 ( r i , K ) + K V = D P C N ( r i , V ) + D P C N 1 ( r i , V ) + V
DPCN represents the application of deep convolution to input X, utilizing contextual information from r i . This operation captures complex spatial relationships and enhances the local features of each component effectively. This dual-branch mechanism preserves and amplifies significant information within Q, K, and V while dynamically capturing the contextual details of surrounding elements. The complete workflow of WMA is depicted in Figure 6, and self-attention is defined as follows:
A t t e n t i o n ( Q , K , V ) = S o f t m a x ( Q K T d h ) V
Figure 5. Different implementation forms of local attention. (a) Swin extracts local features by dividing the image into windows and achieves feature fusion by shifting windows. (b) Slide Transformer [46] extracts local information through window partitioning and convolution. (c) Our model extracts local information through window partitioning and fuses multi-scale information extracted through convolution.
Figure 5. Different implementation forms of local attention. (a) Swin extracts local features by dividing the image into windows and achieves feature fusion by shifting windows. (b) Slide Transformer [46] extracts local information through window partitioning and convolution. (c) Our model extracts local information through window partitioning and fuses multi-scale information extracted through convolution.
Aerospace 12 01049 g005
Figure 6. Structure of Window Movable Attention.
Figure 6. Structure of Window Movable Attention.
Aerospace 12 01049 g006

3.3. Convolutional Feedforward Neural Network

The extraction and expression of local features provide detailed contextual information, which is crucial for processing subsequent global features. In the WMA, effectively fusing global and local information from feature maps can significantly improve segmentation performance. Ref. [47] has demonstrated that convolutions of various sizes can capture multi-scale features. However, the Martian surface is complex, with numerous small objects, making it essential to preserve important local and multi-scale information for effective segmentation. Inspired by this challenge, we designed the Convolutional Feedforward Neural Network (CFFN) module, which better guides feature fusion while preserving the locality of the fused features.
As shown in the Figure 7, we use a 1 × 1 convolution instead of a linear layer, which effectively performs a linear transformation of features while preserving spatial integrity. Using a 1 × 1 convolution, we can increase the network’s non-linear capacity without increasing computational complexity, thus improving feature fusion. Deep convolutions with varying kernel sizes are used to extract multi-scale features. It effectively captures detailed information at multiple scales. Combining varied kernel sizes and receptive fields enhances feature learning, thus improving the identification and segmentation of small objects on the surface of Mars. The specific steps are as follows:
x 1 = L C o n v 1 × 1 ( x i n ) x 2 = G E L U ( x 1 ) x 3 = G P C N ( x 2 , r i , k i ) + x 2 x 4 = G E L U ( x 3 ) x o u t = L C o n v 1 × 1 ( x 4 )
where x i n is the output of the WMA module, L C o n v represents point-wise convolution using 1 × 1 convolution, GELU represents a nonlinear activation function, and GPCN stands for grouped convolution, where r i , k i represents group convolution with different convolution kernels and stride sizes.

3.4. Dynamic Polygon Convolution Module

To enhance the adaptability of convolutional kernels to complex and geometrically heterogeneous shapes, we designed the DPCM module with a Dynamic Polygon Convolution (DPC) cooperation, as shown in Figure 8.
Inspired by deformable convolutions [48] and serpentine convolutions [49], in our DPC, only the central coordinate of a standard convolution kernel is fixed; the remaining kernels are configured into irregular polygonal shapes by the Convex Hull algorithm [50], as shown in Figure 9. It is utilized to define the polygonal boundaries and to verify the interior membership of spatial points. If a point is outside the irregular polygon, it will be moved back to the base coordinate. Such an irregular receptive field can significantly improve the feature extraction capability for diverse complex objects on Mars.
Under such, the process of DPCM can be represented as follows:
x 1 = C o n v e x H U l l ( K , X ) x 2 = D P C ( x 1 ) x 3 = G P C N ( X ) x 4 = C o n v 1 × 1 ( X ) x o u t = C o n v 1 × 1 ( C a t ( x 2 + x 3 + x 4 + X ) )
where K denotes the center coordinates of convolution K i = ( x i , y i ) , and X is the input feature. By fusing features under different channels through 1 × 1 convolution, multiple scales through group convolution, and irregular receptive fields through DPC, DPCM will accurately identify complex and irregular geometric targets in the Martian terrain.

3.5. Group Fusion Module

Feature fusion combined low-level detailed information with high-level semantic features, which can effectively improve the model’s performance on complex tasks. For accurate segmentation of Martian rocks, it preserves fine-grained structures while simultaneously improving the overall contextual understanding. However, most existing feature fusion methods focus on separately extracting internal information from each individual scale, which may introduce noisy or irrelevant features. In fact, it has been demonstrated in [51] that not all feature maps obtained through feature fusion contribute positively to the final prediction of the decoder. To mitigate the instability caused by feature fusion and achieve a lightweight design while preserving accuracy, we proposed the Group Fusion Module (GFM), as shown in Figure 10. By leveraging multi-scale information obtained through high-dimensional pooling and group fusion, the GFM module can efficiently increase the capability of feature representation, which is particularly important for object segmentation on complex Martian scenes.
The denoted feature X i ( i 1 , 2 , 3 , 4 ) is obtained through the encoder. We first performed pooling fusion on the high-dimensional feature X 4 :
P S i ( x 4 i n ) = A d a p t i v e A v g P o o l 2 d ( x , S i ) ( i = 1 , 2 , 3 ) Y = i = 1 n P ^ S i x 4 = G P C N ( Y , r 1 , k 1 )
where S i represents different adaptive average pooling scales, and the obtained feature x 4 is grouped and fused with other features X i , ( i = 1 , 2 , 3 ) . The F i (i = 1, 2, 3, 4), here, represents the output from different channels.

3.6. Architecture Details and Variants

To balance performance and accuracy, we designed three lightweight models: LisseMars-T, LisseMars-S, and LisseMars-B. The corresponding network structure parameters are listed in Table 1. Given an input image with a size of H × W × 3 , the output feature map size obtained through Patch Embedding is ( H / 2 i ) × ( W / 2 i ) ( i 1 , 2 , 3 , 4 ). Each model comprises five stages, including four stages of WMA and an intermediate DPCM layer. The output feature maps at each stage have sizes [32, 64, 128, 256], denoted as F 1 , F 2 , F 3 , F 4 .

4. Experiments

In this section, we will describe the experimental setup and conduct an extensive analysis to validate the performance of the LisseMars on three datasets: SynMars-Air, MarsScapes, and the TianWen dataset.

4.1. Dataset

The SynMars-Air dataset is the first synthetic aerial viewpoint dataset of Mars helicopter released by [45]. It consists of 8 categories: Gravel, SmallRock, Big rock, Bedrock, Ridge, Sand, Soil, and Sky, as shown in Figure 11. Totally, 11,700 images with semantic annotation are divided into training with 9400 images, testing, and validation; each has 1150 images.
MarsScapes [43] is a region-level object annotation panoramic dataset created from images collected by the Curiosity rover, as shown in Figure 12. The dataset contains 9 categories: Big Rock, Soil, Gravel, Bedrock, Ridge, Sand, Sky, Rover, and Unknown. Totally 13,002 images for training, 4192 for testing, and 3608 for validation.
The TianWen dataset is composed of Martian images collected by NaTeCam of the Zhurong rover in the Tianwen-1 mission (Courtesy of NAOC/GRAS: https://www.nssdc.ac.cn/nssdc_zh/html/task/tianwen1.html (accessed on 10 January 2025)), as shown in Figure 13. We labeled all the rocks in 1194 images with a resolution of 512 × 512, and split them into a training set with 896 images and a testing set with 298 images.

4.2. Implement Details

Our model is implemented based on the mmsegmentation framework for semantic segmentation. We trained the model using an NVIDIA GeForce RTX 3090 GPU and deployed it on the JETSON XAVIER NX platform. The batch sizes for training LisseMars-T, LisseMars-S, and LisseMars-B were set to 15, 10, and 5, respectively. The AdamW optimizer is used with a learning rate of 2 × 10 4 .
Metrics like Params (Parameters) and mIoU (Mean Intersection over Union) are used to evaluate the performance of the proposed method. Params evaluates the network’s parameter count, while mIoU measures segmentation accuracy. The mIOU calculation formula is as follows:
I o U = T P T P + F P + F N m I o U = 1 N i = 1 N IoU i
where TP, FN, and FP represent the number of samples correctly classified as positive, the number of positive samples misclassified as negative, and the number of negative samples misclassified as positive, respectively.

4.3. Comparative Experiment

4.3.1. Semantic Segmentation on SynMars-Air

To better validate the superiority of our framework, we compared LisseMars with the most advanced semantic segmentation algorithms currently available. For fairness, other algorithms adopt the mmsegmentation framework and are evaluated based on the Upernet as the decoder. In Table 2, we separated the results of SOTA backbones into three parts sorted by parameters. Our LisseMars-B model achieved the highest mIoU of 87.37%, with relatively fewer parameters compared to other algorithms. Furthermore, it demonstrated exceptional performance in the segmentation of Gravel (0 to 10 pixels) and SmallRock (10 to 100 pixels), reaching 36.48% and 77.65%, respectively, which are 19.87% and 10.27% higher than the second-best Light4Mars model. Gravel and Smallrock are the main objects distributed on the surface of Mars. Accurate identification of them would help the Mars rover proactively plan obstacle-avoidance paths, effectively preventing them from damaging the rover’s wheels while also providing valuable support for scientific exploration.
We visualized the segmentation results of these methods on the SynMars-Air test set, as shown in Figure 14. Other SOATs failed to identify gravel (red pixels) accurately, and our method has better recognition results for both large and small objects. The red boxes in Figure 14 highlight the significant differences between methods. Specifically, our model accurately segments the objects, as shown in the middle two rows of Figure 14, while other methods either fail or predict incorrect segmentation results. when handling low-contrast objects, as seen in the first and last rows of Figure 14, our model successfully segments the objects and captures their boundaries with high precision, whereas the other methods fail to do so.

4.3.2. Semantic Segmentation on MarsScapes

To verify the robustness of our proposed LisseMars, we conducted comparative experiments on the real Mars dataset, MarsScapes. As listed in Table 3, although test accuracy decreased, our Base model (LisseMars-B) still achieved the most competitive mIoU of 66.81%, outperforming other state-of-the-art methods by 0.26% and 0.5% over Swin and Segformer, respectively. Taking the prediction results in Figure 15 as an example, LisseMars demonstrates superior segmentation performance compared to other models. As shown in the first and third rows of Figure 15, LisseMars accurately segments the boundary contours of multiple objects. The second and fourth rows demonstrate that the object categories predicted by LisseMars are closer to the true categories. The segmentation results indicate that our model demonstrates outstanding performance in real Martian scenes. On very large rocks (in BigRock category), LisseMars-T does not rank the top performance. This is because, for the segmentation of large-scale rocks, the core strategy of learning-based methods focuses on employing deeper network architectures to achieve a larger receptive field, thus enabling the effective extraction of large-scale features. The proposed modules in LisseMars, however, are primarily optimized for small-scale object segmentation and for edge device deployment. Driven by such constraints, the segmentation model necessitated a lightweight design, at the expense of segmentation performance on large rocks.

4.3.3. Semantic Segmentation on TianWen

We further used the real Mars images of the TianWen-1 mission to validate the genrealizataion of our model. As shown in Table 4, we compared the Tiny model (LisseMars-T) with other Tiny SOATs; our model outperformed MobileNetV3 and Segformer by 0.26% and 0.72% in mIoU. In addition, LisseMars attained the most competitive mIoU 57.93% in Rock segmentation, achieved 0.6% and 1.64% mIoU higher than those of MobileNetV3 and Convnext, respectively. Compared with SOTA, the visualization of the segmentation results of our method is shown in Figure 16. The red box in Figure 16 shows a significant difference in the segmentation results. Our model accurately segmented the object, as shown in the first and last rows of Figure 16, while other methods exhibit a false negative. For the excessive clustering of multiple targets shown in the second and third rows of Figure 16, our model accurately captures the precise object boundaries, while other models suffer from over-segmentation and mistakenly identify multiple objects as a single object. The results demonstrate that our model exhibits excellent generalization ability in real Martian scenes.

4.4. Ablation Studies

4.4.1. The Impact of Different Proposed Modules

We conducted further research and improvements to validate the efficiency of the proposed modules. As shown in Table 5, we progressively added the WMA, CFFN, GFM, and DPCM modules to the BaseFormer model and analyzed their impact on various categories. The results in Table 5 demonstrate that the model’s performance improves with the incremental addition of each module. Specifically, after adding the WMA module, the model exhibited notable improvements in the Gravel and Smallrock, with increases of 4% and 2.37%, respectively. The next module, CFFN, brings further improvements, particularly in BedRock (+0.44%) and Sand (+2.36%). Adding the GFM module provides a more noticeable enhancement, especially in categories such as Gravel (+14.24%) and SmallRock (+7.8%). Finally, the addition of the DPCM module further enhanced the model’s performance, leading to significant improvements in object segmentation, with the final mIoU reaching 87.37%. These results demonstrate the effectiveness of our proposed modules. As shown in the visualization results in Figure 17, as the number of stages in the model increases, the heatmap shows that the model’s attention area to the objects becomes clearer and distinct, ultimately successfully segmenting the objects.

4.4.2. The Impact of WMA

To validate the efficiency of WMA, we changed the transformer module of the encoder while keeping the GFM decoder unchanged. As shown in Table 6, our model ranked the top with mIoU of 87.37%, which is 2.67% higher than the second Light4Mars. In addition, WMA performs better than other models in terms of Gravel, SmallRock, Ridge, etc. This indicates that WMA can effectively improve the overall performance of image segmentation by fully integrating global and local information to optimize the encoder module without changing the GFM decoder.

4.4.3. The Impact of CFFN

To verify the role of CFFN, we replaced the multilayer while keeping the other structures of the network unchanged. As shown in Table 7, CFFN achieves significantly better segmentation performance for various objects compared to CMLP and UFFN. While CMLP and UFFN effectively capture local information through their layer-by-layer fully connected structures, our CFFN incorporates multi-head convolution to enhance global information extraction, enabling more effective fusion of both global and local features.

4.4.4. The Impact of GFM

As shown in Table 8, we explore the performance of different decoders in our model. Among the evaluated decoders, GFM consistently outperforms other methods in predicting various objects. In particular, GFM achieved the highest segmentation accuracy of 36.48% and 77.65%, respectively, in detecting Gravel and small rocks, which proves the effectiveness of GFM in small object segmentation.

4.4.5. The Impact of DPC

To prove the advantages of our proposed DPC, we replaced it with the state-of-the-art convolution, as shown in Table 9. Our DPC achieved an mIoU that is 0.15% higher than that of DCN and 0.18% higher than Gconv. In addition, our DPC achieved the highest prediction accuracy of 92.71% and 99.61% respectively in Segmenting BedRock and Soil, and ranked the top in the accuracy of predicting Grave and SmallRock. These results prove the effectiveness of DPC. Moreover, to further validate the impact of DPCM’s position in different stages on the performance of the LisseMars model, we conducted additional experiments by adjusting the placement of DPCM. As shown in Table 10, when DPCM is positioned in stage 3, it achieves optimal segmentation accuracy by better integrating low-level features and low-level features.

4.5. Runtime Performance

To demonstrate the performance of the proposed model on an edge device, we deployed a resource-efficient and high-precision LisseMars-T model on Nvidia Jeston NX platform. It is a credit card-sized and low-power (10 W/15 W) system-on-a-chip, making it capable of being deployed on a lightweight helicopter or as a key component for deploying algorithms in deep space exploration applications.
The model is exported in TensorRT format (https://developer.nvidia.cn/tensorrt (accessed on 10 April 2025)) with 512 × 512 resolution of the input image. As shown in Table 11, the average inference time per image of LisseMars on NX is 0.048 s on the SynMras-Air, 0.04 s on MarsScapes, respectively. On real TianWen datasets, our model achieved a processing time of 0.044 s per image, showcasing its potential for deployment in future Mars helicopter missions.

5. Conclusions

In this paper, we propose a novel lightweight model semantic segmentation LisseMars, specifically designed for the Mars helicopter. The model integrates a new window movable attention (WMA) module to enhance the fusion of global and local information. Additionally, we developed a multi-head convolutional feedforward network (CFFN) to further improve feature extraction and fusion, along with a dynamic polygon convolutional (DPC) to achieve precise segmentation of polygonal structures. The proposed method was evaluated on the challenging SynMars-Air, MarsScapes, and TianWen datasets. Experimental results demonstrated that LisseMars effectively segmented object contours across various complex Martian scenes. It has achieved or approached state-of-the-art results across most terrain categories, and only fell short on large rocks due to the constraints of its lightweight architecture. Furthermore, the model was successfully deployed on the edge device, validating its real-time performance and potential integration for future Mars helicopter mission. Future research will be dedicated to refining model lightweighting and multi-scale feature extraction, ultimately enabling accurate semantic segmentation for the diverse range of object on the Martian surface.

Author Contributions

Conceptualization, B.L., M.Y. and X.X.; methodology, B.L., F.W. and Y.Q.; formal analysis, B.Z. and Q.L.; investigation, B.L., Y.Q. and F.W.; resources, B.Z., H.C. and X.H.; data curation, Y.Q., F.W. and Q.L.; writing—original draft preparation, Y.Q., F.W. and B.L.; writing—review and editing, X.X. and M.Y.; visualization, B.L. and Y.Q.; supervision, H.C., M.Y. and X.X.; project administration, B.Z. and X.H.; funding acquisition, M.Y. and X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work is sponsored by the National Natural Science Foundation of China (NSFC) through grants Nos. 52472448, 62394354, 62394355, and the Natural Science Foundation of Jilin Province through grants No. YDZJ202301ZYTS396. The authors also thank the China National Space Administration for providing the TianWen-1 data processed and produced by the Ground Research and Application System (GRAS) of China’s Lunar and Planetary Exploration Program to make this study possible.

Data Availability Statement

SynMars-Air dataset: https://github.com/CVIR-Lab/SynMars/tree/SynMars-Air (accessed on 10 January 2025); MarsScapes dataset: https://github.com/InRobots/MarsScapes (accessed on 10 January 2025); Tianwen-1 dataset: https://www.nssdc.ac.cn/nssdc_zh/html/task/tianwen1.html (accessed on 10 January 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xiao, X.; Yao, M.; Cui, H.; Fu, Y. Safe Mars landing strategy: Towards lidar-based high altitude hazard detection. Adv. Space Res. 2019, 63, 2535–2550. [Google Scholar] [CrossRef]
  2. Leake, C.; Grip, H.; Steyert, V.; Hasseler, T.D.; Cacan, M.; Jain, A. HeliCAT-DARTS: A High Fidelity, Closed-Loop Rotorcraft Simulator for Planetary Exploration. Aerospace 2024, 11, 727. [Google Scholar] [CrossRef]
  3. Giacomini, E.; Westerberg, L.G. Rotorcraft Airfoil Performance in Martian Environment. Aerospace 2024, 11, 628. [Google Scholar] [CrossRef]
  4. von Ehrenfried, M.D. Ingenuity. In Perseverance and the Mars 2020 Mission: Follow the Science to Jezero Crater; Springer: Cham, Switzerland, 2022; pp. 111–125. [Google Scholar]
  5. Zhao, P.; Gao, X.; Zhao, B.; Liu, H.; Wu, J.; Deng, Z. Machine learning assisted prediction of airfoil lift-to-drag characteristics for mars helicopter. Aerospace 2023, 10, 614. [Google Scholar] [CrossRef]
  6. Xiao, X.; Cui, H.; Yao, M.; Fu, Y.; Qi, W. Auto rock detection via sparse-based background modeling for mars rover. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–6. [Google Scholar]
  7. Li, G.; Geng, Y.; Xiao, X. Multi-scale rock detection on Mars. Sci. China Inf. Sci. 2018, 61, 1. [Google Scholar] [CrossRef]
  8. Wang, A.; Wang, L.; Zhang, Y.; Hua, B.; Li, T.; Liu, Y.; Lin, D. Landing site positioning and descent trajectory reconstruction of Tianwen-1 on Mars. Astrodynamics 2022, 6, 69–79. [Google Scholar] [CrossRef]
  9. Xu, C.; Huang, X.; Guo, M.; Li, M.; Hu, J.; Wang, X. End-to-end Mars entry, descent, and landing modeling and simulations for Tianwen-1 guidance, navigation, and control system. Astrodynamics 2022, 6, 53–67. [Google Scholar] [CrossRef]
  10. Liu, S.; Kong, J.; Cao, J.; Huang, H.; Man, H.; Yan, J.; Li, X. Precise orbit determination for Tianwen-1 during mapping phase. Astrodynamics 2024, 8, 471–481. [Google Scholar] [CrossRef]
  11. Li, Z.; Wu, B.; Chen, Z.; Ma, Y. Transformer-Based Method for Semantic Segmentation and Reconstruction of the Martian Surface. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, 48, 1643–1649. [Google Scholar] [CrossRef]
  12. Wang, Y.; Wu, B. Active machine learning approach for crater detection from planetary imagery and digital elevation models. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5777–5789. [Google Scholar] [CrossRef]
  13. Jiang, S.; Lian, Z.; Yung, K.L.; Ip, W.H.; Gao, M. Automated detection of multitype landforms on mars using a light-weight deep learning-based detector. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 5015–5029. [Google Scholar] [CrossRef]
  14. Yang, S.; Cai, Z. High-resolution feature pyramid network for automatic Crater detection on Mars. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4601012. [Google Scholar] [CrossRef]
  15. Cheng, S.; Yao, M.; Xiao, X. Dc-mot: Motion deblurring and compensation for multi-object tracking in uav videos. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 789–795. [Google Scholar]
  16. Chen, Z.; Wu, B.; Liu, W.C. Mars3DNet: CNN-based high-resolution 3D reconstruction of the Martian surface from single images. Remote Sens. 2021, 13, 839. [Google Scholar] [CrossRef]
  17. Tian, P.; Yao, M.; Xiao, X.; Zheng, B.; Cao, T.; Xi, Y.; Liu, H.; Cui, H. 3D Semantic Terrain Reconstruction of Monocular Close-up Images of Martian Terrains. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar]
  18. Li, Z.; Wu, B.; Liu, W.C.; Chen, Z. Integrated photogrammetric and photoclinometric processing of multiple HRSC images for pixelwise 3-D mapping on Mars. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4601113. [Google Scholar] [CrossRef]
  19. Meoni, G.; Märtens, M.; Derksen, D.; See, K.; Lightheart, T.; Sécher, A.; Martin, A.; Rijlaarsdam, D.; Fanizza, V.; Izzo, D. The OPS-SAT case: A data-centric competition for onboard satellite image classification. Astrodynamics 2024, 8, 507–528. [Google Scholar] [CrossRef]
  20. Hu, R.; Huang, X.; Xu, C. Integrated visual navigation based on angles-only measurements for asteroid final landing phase. Astrodynamics 2023, 7, 69–82. [Google Scholar] [CrossRef]
  21. Gui, Y.; Qi, Y.; Xiao, X.; Lin, B.; Cui, H.; Huang, X. SPTN: Transformer-based spacecraft pose estimation network for space objects tracking. Astrodynamics 2025, 9, 713–725. [Google Scholar] [CrossRef]
  22. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  23. Ma, X.; Guo, J.; Sansom, A.; McGuire, M.; Kalaani, A.; Chen, Q.; Tang, S.; Yang, Q.; Fu, S. Spatial pyramid attention for deep convolutional neural networks. IEEE Trans. Multimed. 2021, 23, 3048–3058. [Google Scholar] [CrossRef]
  24. Tolstikhin, I.O.; Houlsby, N.; Kolesnikov, A.; Beyer, L.; Zhai, X.; Unterthiner, T.; Yung, J.; Steiner, A.; Keysers, D.; Uszkoreit, J.; et al. Mlp-mixer: An all-mlp architecture for vision. Adv. Neural Inf. Process. Syst. 2021, 34, 24261–24272. [Google Scholar]
  25. Wu, K.; Peng, H.; Chen, M.; Fu, J.; Chao, H. Rethinking and improving relative position encoding for vision transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10033–10041. [Google Scholar]
  26. Su, J.; Ahmed, M.; Lu, Y.; Pan, S.; Bo, W.; Liu, Y. Roformer: Enhanced transformer with rotary position embedding. Neurocomputing 2024, 568, 127063. [Google Scholar] [CrossRef]
  27. Deng, X.; Wang, P.; Lian, X.; Newsam, S. NightLab: A dual-level architecture with hardness detection for segmentation at night. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16938–16948. [Google Scholar]
  28. Guastella, D.C.; Muscato, G. Learning-based methods of perception and navigation for ground vehicles in unstructured environments: A review. Sensors 2020, 21, 73. [Google Scholar] [CrossRef] [PubMed]
  29. Liu, S.; Zhao, H.; Yuan, Z.; Xiao, L.; Shen, C.; Wan, X.; Tang, X.; Zhang, L. A Machine Learning Approach for the Autonomous Identification of Hardness in Extraterrestrial Rocks from Digital Images. Aerospace 2024, 12, 26. [Google Scholar] [CrossRef]
  30. Zhang, Z.; Cheng, Y.; Bu, L.; Ye, J. Rapid SLAM Method for Star Surface Rover in Unstructured Space Environments. Aerospace 2024, 11, 768. [Google Scholar] [CrossRef]
  31. Giacomini, E.; Westerberg, L.G. Numerical Study on Particle Accumulation and Its Impact on Rotorcraft Airfoil Performance on Mars. Aerospace 2025, 12, 368. [Google Scholar] [CrossRef]
  32. Solarna, D.; Gotelli, A.; Le Moigne, J.; Moser, G.; Serpico, S.B. Crater detection and registration of planetary images through marked point processes, multiscale decomposition, and region-based analysis. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6039–6058. [Google Scholar] [CrossRef]
  33. Xiao, X.; Cui, H.; Yao, M.; Tian, Y. Autonomous rock detection on mars through region contrast. Adv. Space Res. 2017, 60, 626–635. [Google Scholar] [CrossRef]
  34. Li, H.; Qiu, L.; Li, Z.; Meng, B.; Huang, J.; Zhang, Z. Automatic rocks segmentation based on deep learning for planetary rover images. J. Aerosp. Inf. Syst. 2021, 18, 755–761. [Google Scholar] [CrossRef]
  35. Feng, L.; Wang, S.; Wang, D.; Xiong, P.; Xie, J.; Hu, Y.; Zhang, M.; Wu, E.Q.; Song, A. Mobile-DeepRFB: A Lightweight Terrain Classifier for Automatic Mars Rover Navigation. IEEE Trans. Autom. Sci. Eng. 2023, 22, 17442–17451. [Google Scholar] [CrossRef]
  36. Dai, Y.; Zheng, T.; Xue, C.; Zhou, L. SegMarsViT: Lightweight mars terrain segmentation network for autonomous driving in planetary exploration. Remote Sens. 2022, 14, 6297. [Google Scholar] [CrossRef]
  37. Xiong, Y.; Xiao, X.; Yao, M.; Liu, H.; Yang, H.; Fu, Y. Marsformer: Martian rock semantic segmentation with transformer. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4600612. [Google Scholar] [CrossRef]
  38. Liu, H.; Yao, M.; Xiao, X.; Xiong, Y. RockFormer: A U-shaped transformer network for Martian rock segmentation. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4600116. [Google Scholar] [CrossRef]
  39. Liu, H.; Yao, M.; Xiao, X.; Cui, H. A hybrid attention semantic segmentation network for unstructured terrain on Mars. Acta Astronaut. 2023, 204, 492–499. [Google Scholar] [CrossRef]
  40. Xiong, Y.; Xiao, X.; Yao, M.; Cui, H.; Fu, Y. Light4Mars: A lightweight transformer model for semantic segmentation on unstructured environment like Mars. ISPRS J. Photogramm. Remote Sens. 2024, 214, 167–178. [Google Scholar] [CrossRef]
  41. Swan, R.M.; Atha, D.; Leopold, H.A.; Gildner, M.; Oij, S.; Chiu, C.; Ono, M. Ai4mars: A dataset for terrain-aware autonomous driving on mars. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1982–1991. [Google Scholar]
  42. Xiao, X.; Yao, M.; Liu, H.; Wang, J.; Zhang, L.; Fu, Y. A kernel-based multi-featured rock modeling and detection framework for a mars rover. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 3335–3344. [Google Scholar] [CrossRef]
  43. Liu, H.; Yao, M.; Xiao, X.; Zheng, B.; Cui, H. MarsScapes and UDAFormer: A Panorama Dataset and a Transformer-Based Unsupervised Domain Adaptation Framework for Martian Terrain Segmentation. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4600117. [Google Scholar] [CrossRef]
  44. Allak, E.; Brommer, C.; Dallenbach, D.; Weiss, S. AMADEE-18: Vision-Based Unmanned Aerial Vehicle Navigation for Analog Mars Mission (AVI-NAV). Astrobiology 2020, 20, 1321–1337. [Google Scholar] [CrossRef]
  45. Qi, Y.; Xiao, X.; Yao, M.; Xiong, Y.; Zhang, L.; Cui, H. AirFormer: Learning-Based Object Detection For Mars Helicopter. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 100–111. [Google Scholar] [CrossRef]
  46. Pan, X.; Ye, T.; Xia, Z.; Song, S.; Huang, G. Slide-transformer: Hierarchical vision transformer with local self-attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 2082–2091. [Google Scholar]
  47. Chen, W.; Shi, K. Multi-scale attention convolutional neural network for time series classification. Neural Netw. 2021, 136, 126–140. [Google Scholar] [CrossRef]
  48. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable {DETR}: Deformable Transformers for End-to-End Object Detection. In Proceedings of the International Conference on Learning Representations, Online, 3–7 May 2021. [Google Scholar]
  49. Qi, Y.; He, Y.; Qi, X.; Zhang, Y.; Yang, G. Dynamic snake convolution based on topological geometric constraints for tubular structure segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 6070–6079. [Google Scholar]
  50. Barber, C.B.; Dobkin, D.P.; Huhdanpaa, H. The quickhull algorithm for convex hulls. ACM Trans. Math. Softw. (TOMS) 1996, 22, 469–483. [Google Scholar] [CrossRef]
  51. Ge, Y.; Yang, Z.; Huang, Z.; Ye, F. A multi-level feature fusion method based on pooling and similarity for HRRS image retrieval. Remote Sens. Lett. 2021, 12, 1090–1099. [Google Scholar] [CrossRef]
  52. Yu, C.; Gao, C.; Wang, J.; Yu, G.; Shen, C.; Sang, N. Bisenet v2: Bilateral network with guided aggregation for real-time semantic segmentation. Int. J. Comput. Vis. 2021, 129, 3051–3068. [Google Scholar] [CrossRef]
  53. Koonce, B.; Koonce, B. MobileNetV3. Convolutional Neural Networks with Swift for Tensorflow: Image Recognition and Dataset Categorization; Apress: Berkeley, CA, USA, 2021; pp. 125–144. [Google Scholar]
  54. Wang, W.; Chen, W.; Qiu, Q.; Chen, L.; Wu, B.; Lin, B.; He, X.; Liu, W. Crossformer++: A versatile vision transformer hinging on cross-scale attention. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 3123–3136. [Google Scholar] [CrossRef] [PubMed]
  55. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  56. Dong, X.; Bao, J.; Chen, D.; Zhang, W.; Yu, N.; Yuan, L.; Chen, D.; Guo, B. Cswin transformer: A general vision transformer backbone with cross-shaped windows. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12124–12134. [Google Scholar]
  57. Zhang, J.; Li, X.; Li, J.; Liu, L.; Xue, Z.; Zhang, B.; Jiang, Z.; Huang, T.; Wang, Y.; Wang, C. Rethinking mobile block for efficient attention-based models. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 1389–1400. [Google Scholar]
  58. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pvt v2: Improved baselines with pyramid vision transformer. Comput. Vis. Media 2022, 8, 415–424. [Google Scholar] [CrossRef]
  59. Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; Yan, S. Metaformer is actually what you need for vision. In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10819–10829. [Google Scholar]
  60. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  61. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  62. Guo, M.H.; Lu, C.Z.; Hou, Q.; Liu, Z.; Cheng, M.M.; Hu, S.M. Segnext: Rethinking convolutional attention design for semantic segmentation. Adv. Neural Inf. Process. Syst. 2022, 35, 1140–1156. [Google Scholar]
  63. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  64. Lin, W.; Wu, Z.; Chen, J.; Huang, J.; Jin, L. Scale-aware modulation meet transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 6015–6026. [Google Scholar]
  65. Li, K.; Wang, Y.; Zhang, J.; Gao, P.; Song, G.; Liu, Y.; Li, H.; Qiao, Y. Uniformer: Unifying convolution and self-attention for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 12581–12600. [Google Scholar] [CrossRef]
  66. Xu, J.; Xiong, Z.; Bhattacharyya, S.P. PIDNet: A real-time semantic segmentation network inspired by PID controllers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19529–19539. [Google Scholar]
  67. Lv, W.; Wei, L.; Zheng, D.; Liu, Y.; Wang, Y. Marsnet: Automated rock segmentation with transformers for tianwen-1 mission. IEEE Geosci. Remote Sens. Lett. 2023, 20, 3506605. [Google Scholar] [CrossRef]
  68. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 June 2018; pp. 3–19. [Google Scholar]
  69. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  70. Vaswani, A. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 11, 6000–6010. [Google Scholar]
  71. Kirillov, A.; Girshick, R.; He, K.; Dollár, P. Panoptic feature pyramid networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6399–6408. [Google Scholar]
  72. Cheng, B.; Collins, M.D.; Zhu, Y.; Liu, T.; Huang, T.S.; Adam, H.; Chen, L.C. Panoptic-deeplab: A simple, strong, and fast baseline for bottom-up panoptic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12475–12485. [Google Scholar]
  73. Howard, A.G. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  74. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  75. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
Figure 2. From left to right, the examples show SynMars-Air, MarsScapes, and images captured by the Ingenuity helicopter. (ad) display a comparison of BigRock, Sand, Ridge, and BedRock.
Figure 2. From left to right, the examples show SynMars-Air, MarsScapes, and images captured by the Ingenuity helicopter. (ad) display a comparison of BigRock, Sand, Ridge, and BedRock.
Aerospace 12 01049 g002
Figure 3. The overall architecture of our LisseMars. X i represent the feature output through each encoder modules, and F i represent the feature output through each decoder modules.
Figure 3. The overall architecture of our LisseMars. X i represent the feature output through each encoder modules, and F i represent the feature output through each decoder modules.
Aerospace 12 01049 g003
Figure 4. The working principle of WMA. Compared to models like PVT, Deformable DETR, and Swin, our WMA demonstrates a significant advantage in terms of receptive field.
Figure 4. The working principle of WMA. Compared to models like PVT, Deformable DETR, and Swin, our WMA demonstrates a significant advantage in terms of receptive field.
Aerospace 12 01049 g004
Figure 7. The structure of CFFN. (a) FFN in ViT. (b) Mix-FFN in PVTv2. (c) Ours.
Figure 7. The structure of CFFN. (a) FFN in ViT. (b) Mix-FFN in PVTv2. (c) Ours.
Aerospace 12 01049 g007
Figure 8. (a) Basic shape of Martian rocks. (b) Structure of DPCM. (c) DPC. The input X is processed through various paths, including DPC, and then merged to produce the final output features.
Figure 8. (a) Basic shape of Martian rocks. (b) Structure of DPCM. (c) DPC. The input X is processed through various paths, including DPC, and then merged to produce the final output features.
Aerospace 12 01049 g008
Figure 9. (Left): Initial coordinate display of DPC and a demonstration of the convolution kernel’s movement. (Right): The receptive field of DPC. When the formed polygon has concave edges, the convolution kernel returns to the initial coordinates.
Figure 9. (Left): Initial coordinate display of DPC and a demonstration of the convolution kernel’s movement. (Right): The receptive field of DPC. When the formed polygon has concave edges, the convolution kernel returns to the initial coordinates.
Aerospace 12 01049 g009
Figure 10. The structure of GFM.
Figure 10. The structure of GFM.
Aerospace 12 01049 g010
Figure 11. Samples of different categories in the SynMars-Air dataset (Source: [45]).
Figure 11. Samples of different categories in the SynMars-Air dataset (Source: [45]).
Aerospace 12 01049 g011
Figure 12. A sample of MarsScapes panoramic image (Source: [43]).
Figure 12. A sample of MarsScapes panoramic image (Source: [43]).
Aerospace 12 01049 g012
Figure 13. Samples from the TianWen dataset. (Source: NAOC/GRAS).
Figure 13. Samples from the TianWen dataset. (Source: NAOC/GRAS).
Aerospace 12 01049 g013
Figure 14. Visual comparison on the SynMars-Air test set. The red boxes mask some challenging objects that are difficult to cut.
Figure 14. Visual comparison on the SynMars-Air test set. The red boxes mask some challenging objects that are difficult to cut.
Aerospace 12 01049 g014
Figure 15. Visual comparison on the MarsScapes test set.
Figure 15. Visual comparison on the MarsScapes test set.
Aerospace 12 01049 g015
Figure 16. Visual comparison on the TianWen test set. The red boxes mask some challenging objects that are difficult to cut.
Figure 16. Visual comparison on the TianWen test set. The red boxes mask some challenging objects that are difficult to cut.
Aerospace 12 01049 g016
Figure 17. The visualization result of feature heatmaps with different stages.
Figure 17. The visualization result of feature heatmaps with different stages.
Aerospace 12 01049 g017
Table 1. Model variants of LisseMars. N represents the number of blocks, h e a d represents the number of heads, and p and K represent the size of patches and the number of convolution kernels, respectively.
Table 1. Model variants of LisseMars. N represents the number of blocks, h e a d represents the number of heads, and p and K represent the size of patches and the number of convolution kernels, respectively.
ModuleLisseMars-TLisseMars-SLisseMars-B
Stage 1WMAp = 4p = 4p = 4
C 1  = 8, h e a d  = 1, N 1  = 1 C 1  = 32, h e a d  = 2, N 1  = 1 C 1  = 32, h e a d  = 2, N 1  = 2
Stage 2WMAp = 2p = 2p = 2
C 2  = 16, h e a d  = 2, N 2  = 1 C 2  = 64, h e a d  = 2, N 2  = 1 C 2  = 64, h e a d  = 2, N 2  = 2
Stage 3DPCMK = 9K = 9K = 9
N 3  = 1 N 3  = 1 N 3  = 1
C 3  = 32, C 3  = 64 C 3  = 128
Stage 4WMAp = 2p = 2p = 2
C 4  = 32, h e a d  = 4, N 4  = 1 C 4  = 128, h e a d  = 4, N 4  = 1 C 4  = 128, h e a d  = 4, N 4  = 2
Stage 5WMAp = 2p = 2p = 2
C 5  = 64, h e a d  = 8, N 5  = 1 C 5  = 256, h e a d  = 8, N 5  = 1 C 5  = 256, h e a d  = 8, N 5  = 2
Table 2. Semantic Segmentation with 160 K on SynMars-Air (the first and second best results are highlighted in red and underlined, respectively).
Table 2. Semantic Segmentation with 160 K on SynMars-Air (the first and second best results are highlighted in red and underlined, respectively).
BackboneParamsFLOPs160 K
(M)(G)GravelSmallRockBigRockBedRockSandSoilRidgeSkymIoU (%)
Bisenetv2 [52]3.3512.31029.1481.4888.4395.3698.4798.1199.7873.85
MobileNetV3 [53]3.288.600.7736.2280.988.9794.6598.7198.3399.7674.79
Crossformer++-T [54]23.34.90.0144.0580.588.3496.098.8798.1799.5675.79
Convnext-T [55]284.52.8154.2285.9491.4895.8798.8391.9799.0577.52
Cswin-T [56]38.961.50.0031.6462.3425.2194.7675.792.9747.83
EMO-1M [57]5.62.40.00.0465.1778.1491.2897.3394.6699.265.72
Pvtv2-B0 [58]29.145.82.6247.6277.1484.3495.6198.897.5399.5975.41
Poolformer-s12 [59]12.4754.310.0327.1155.9679.2783.2697.5188.5598.0366.21
DeepLabV3+-r18 [60]15.6530.751.1556.2289.8893.1382.8298.1290.5599.4776.44
Segformer-b0 [61]3.76.412.938.9684.8788.4486.9898.0495.9899.5674.46
Segnext-T [62]4.36.6024.8379.6388.4196.0498.4498.5299.7573.2
Light4Mars-T [40]0.10.534.7341.3876.885.693.4298.6194.2398.4474.15
LisseMars-T0.120.907.2548.4684.5389.9196.2698.9396.3398.9777.58
Crossformer++-S [54]52.09.51.3651.6980.1887.5796.198.994.2798.1876.03
Convnext-S [55]508.73.9160.9686.6291.7795.7999.1699.3899.9379.69
Swin-T [63]60.02363.0355.0976.3383.4194.8498.9484.4195.1473.90
Swin-S [63]81.03292.2957.2182.6788.1795.7999.193.6197.9977.10
Cswin-S [56]51.373.7017.3349.170.3977.0797.1497.3498.2862.64
EMO-2M [57]6.93.50.00.0267.1579.3192.2197.3394.2398.9966.16
SMT-B [64]51.776.21.956.991.0594.3597.9499.2599.099.8282.19
Uniformer-B [65]70.090.347.4969.1390.192.7197.4999.398.499.9382.51
Pvtv2-B3 [58]49.062.45.5366.3888.6491.6397.8899.3599.599.9381.11
PIDNet-s [66]7.647.6031,5287.4891.2397.298.8698.5399.7875.58
Poolformer-s36 [59]34.642.00.0127.157.1978.8383.4697.5288.6597.9666.34
DeepLabV3+-r50 [60]47.062.70.956.3790.1893.4497.4199.1398.3499.8479.45
Segformer-b2 [61]25.415.10.9953.4289.9294.0497.6999.1998.9499.7979.25
Segnext-B [62]27.835.70.0947.586.6590.0196.1599.0199.1499.8977.31
SMT-B [64]61.83288.2770.0389.8792.1198.499.4199.3799.8882.17
Uniformer-B [65]804717.4969.1390.192.7197.4999.3898.499.581.78
Light4Mars-S [40]1.505.4515.0963.9790.9893.9197.699.3499.7799.9682.51
LisseMars-S1.3556.4917.6367.7791.5394.0497.6399.4199.399.8583.39
Crossformer++-B [54]92.016.61.9449.9780.3187.6195.3698.8796.7999.8376.33
Convnext-B [55]8945.05.3364.6486.9192.0796.4899.299.4299.9480.50
Swin-B [63]121.04791.956.991.0594.3597.9499.2599.099.8280.03
Cswin-B [56]96.7115.53025.0256.9277.1679.097.3778.5994.863.61
EMO-5M [57]10.35.80.00.0171.1480.8892.7797.5393.8398.8266.87
SMT-L [64]97.1132.29.0570.1790.3292.4998.3599.4299.5699.9582.41
Light4Mars-B [40]2.579.4316.6167.3891.5394.4597.8799.499.2899.983.30
Uniformer-L [65]119.0130.610.3671.5191.6493.6598.3199.4499.5699.9583.05
Pvtv2-B5 [58]85.791.18.7669.3589.5992.4597.8799.499.5899.9582.12
PIDNet-L [66]36.9275.8033.6289.3592.6397.7498.9498.7599.7876.35
Poolformer-m48 [59]77.147.20.9450.3286.7691.8897.3399.199.0999.8378.16
DeepLabV3+-r101 [60]66.783.42.1960.1189.9493.1696.8799.1898.5899.8279.98
Segformer-b5 [61]82.022.53.9163.490.6894.0797.7699.2999.3299.8881.04
Segnext-L [62]48.970.00.1547.3687.6190.7296.3699.0299.1299.977.53
SMT-L [64]1025469.0570.1790.3292.4998.3599.4299.5699.9582.41
Uniformer-L [65]10049010.3671.5191.6493.6598.3199.4499.5699.9583.05
LisseMars-B9.1621.736.4877.6592.7195.098.0499.6199.5699.9587.37
Table 3. Semantic Segmentation with 160 K on MarsScapes. (the first and second best results are highlighted in red and underlined, respectively).
Table 3. Semantic Segmentation with 160 K on MarsScapes. (the first and second best results are highlighted in red and underlined, respectively).
BackboneParamsFLOPs160 K
(M)(G)SoilBedRockGravelSandBigRockRidgeSkyRoverUnknowmIoU (%)
Bisenetv2 [52]3.356.1666.7864.7937.8059.4644.0245.9470.6381.9846.0157.49
MobileNetV3 [53]3.284.3769.4273.4849.6465.4347.4748.8477.3862.3045.9159.99
Crossformer++-T [54]23.34.961.960.1733.1148.4431.5542.3464.8764.1443.7250.03
Convnext-T [55]284.563.8149.5838.6449.938.6454.8879.1360.6443.0953.15
EMO-2M [57]6.93.557.2950.5523.8543.0133.8336.3971.9836.6725.5642.12
Light4Mars-T [40]0.10.5369.3570.2550.8260.2945.6153.8781.8855.4245.8959.26
Pvtv2-B0 [58]7.2511.7964.9159.1737.6351.7736.6649.5475.1886.0848.5456.61
Segnext-T [62]40.052.362.5157.7443.7249.5540.4128.3263.9212.79039.89
Uniformer-S [65]52.024765.8557.2538.4552.6741.7546.6177.6663.4139.6853.7
PIDNet-s [66]7.647.660.7955.8432.3253.9839.934.3949.6187.4546.6751.22
LisseMars-T0.120.9068.9567.5145.8158.9440.7462.2483.8489.953.4363.48
MarsNet [67]33.21120.2976.0256.3944.7756.4435.4535.8379.7271.5945.6955.77
Crossformer++-S [54]52.09.570.1261.6749.3360.4339.438.6375.0542.2651.7154.30
Convnext-S [55]508.764.4251.1340.051.2239.6348.473.8592.5945.3256.28
EMO-3M [57]6.93.558.8855.533.7750.4734.2846.3975.4461.6831.0349.71
Light4Mars-S [40]1.505.4569.4171.5746.1560.7247.6760.8482.1395.0046.2664.30
Segnext-B [62]27.835.753.8339.0122.5346.0635.8134.5857.8769.2944.8344.87
Uniformer-B [65]80.047165.2455.5138.7552.2442.3751.3280.2469.9140.6655.14
Poolformer [59]15.6515.3867.2166.5043.6160.7845.9651.5878.0094.7948.2061.85
DeepLabV3+ [60]12.4727.1669.5373.9943.4861.8648.5858.4082.5497.8046.8164.78
LisseMars-S1.3556.4970.2171.9250.0858.7446.9862.7280.7692.3253.3165.23
Crossformer++-B [54]92.016.669.5466.5344.8959.5543.6244.064.6183.1849.8258.42
Convnext-B [55]8945.062.7951.0241.6150.2238.256.4578.0495.1949.258.06
EMO-5M [57]10.283.0568.0171.0640.0358.6742.7648.9780.3295.7945.3361.22
Light4Mars-B [40]2.579.4370.4673.4451.6262.1448.6560.4084.1092.0852.3966.14
PIDNet-L [66]36.9275.866.7065.1236.4557.0439.3151.8776.6789.6546.0158.76
Segnext-L [62]48.070.063.4860.4236.8353.2840.2141.8472.1160.97047.68
Uniformer-L [65]10049067.1162.5341.4951.0142.9146.4977.6975.0540.356.06
Cswin [56]52.01115.4164.1863.0333.4252.0735.3449.3875.3086.9146.4256.23
Swin [63]58.95121.8669.2875.5046.4761.1347.1166.6583.5396.1753.1366.55
SMT [64]52.24117.068.9475.9644.7664.1951.0258.0480.7296.6647.3265.29
Segformer [61]3.723.7069.4875.2646.6262.8548.9661.9880.5298.0053.1366.31
LisseMars-B9.1621.770.4870.8150.3161.2848.1467.7381.5196.4154.766.81
Table 4. Semantic Segmentation with 160 K on TianWen. (the first and second best results are highlighted in red and underlined, respectively).
Table 4. Semantic Segmentation with 160 K on TianWen. (the first and second best results are highlighted in red and underlined, respectively).
BackboneParamsFLOPs160 KmIoU (%)
(M)(G)BackgroundRock
Bisenetv2 [52]3.3512.3199.1953.5876.39
MobileNetV3 [53]3.288.6099.2757.3478.30
Convnext-T [55]28.04.599.2256.2977.76
Light4Mars-T [40]0.10.5399.1856.1577.84
Segformer-b0 [61]3.76.4199.1753.4176.74
LisseMars-T0.120.90499.1957.9378.56
Table 5. Effect of different proposed modules (The improvement is highlighted in red).
Table 5. Effect of different proposed modules (The improvement is highlighted in red).
BackboneFusionGravelSmallRockBigRockBedRockSandSoilRidgeSkymIoU (%)
BaseFormer 16.6167.3891.5394.4597.8799.499.2899.983.30
+WMA20.6469.7592.0394.2695.6999.4598.3799.983.76 (↑0.46)
+CFFN19.469.2592.1894.798.0599.4499.3699.9284.04 (↑0.28)
+GFM33.6477.0291.7394.1398.2899.699.799.9686.76 (↑2.72)
+DPCM36.4877.6592.7195.098.0499.6199.5699.9187.37 (↑0.61)
Table 6. Comparison of different decoders based on LisseMars-B on SynMars-Air (the first and second best results are highlighted in red and underlined, respectively).
Table 6. Comparison of different decoders based on LisseMars-B on SynMars-Air (the first and second best results are highlighted in red and underlined, respectively).
AttentionEncoderDecoderGravelSmallRockBigRockBedRockSandSoilRidgeSkymIoU (%)
CBAM [68] 026.2573.1479.5689.8588.3290.6394.7567.81
SE attention [69] 022.2461.5570.4784.2183.3786.7892.4262.63
Shift Window Attention [61] 2.1157.0692.1494.8497.8499.699.3399.7580.3
Effcient Attention [61] 5.965.9691.894.4798.099.3099.599.9981.8
Linear Spatial Reduction Attention [58] 10.2670.1790.4593.1397.9099.4998.8099.9982.52
Mix Attention [24] 14.065.293.8194.6895.1398.2199.5299.7682.53
Scale-Aware Aggregation [27]GFM9.370.0490.3892.5898.0499.4199.5499.9582.5
Multi-head Relation Attention [65] 12.3172.292.0295.498.2799.5899.5699.983.65
Squeeze Window Attention [40] 23.0470.9492.3795.097.999.4799.4499.8884.7
Window Movable Attention (Ours) 36.4877.6592.7195.098.0499.6199.5699.9187.37
Table 7. Comparison of different multi-layer perceptrons based on LisseMars-B on SynMars-Air (the first and second best results are highlighted in red and underlined, respectively).
Table 7. Comparison of different multi-layer perceptrons based on LisseMars-B on SynMars-Air (the first and second best results are highlighted in red and underlined, respectively).
ModelMulti-LayerGravelSmallRockBigRockBedRockSandSoilRidgeSkymIoU (%)
LisseMarsCMLP [65]22.0170.490.8793.4492.9999.4697.1299.8583.27
MLP [70]21.8870.4690.4493.0291.5899.4696.2899.8282.87
UFFN [40]21.7570.2890.5593.1792.5299.4696.8499.8383.05
CFFN (Ours)36.4877.6592.7195.098.0499.6199.5699.9187.37
Table 8. Comparison of different decoders based on LisseMars-B on SynMars-Air (the first and second best results are highlighted in red and underlined, respectively).
Table 8. Comparison of different decoders based on LisseMars-B on SynMars-Air (the first and second best results are highlighted in red and underlined, respectively).
EncoderDecoderGravelSmallRockBigRockBedRockSandSoilRidgeSkymIoU (%)
WMAEncHead [61]018.1186.7489.197.0998.698.3399.7573.47
SegformerHead [61]23.0771.1790.4593.3395.3299.4998.4599.9983.89
FPN [71]26.072.7791.9594.6898.0399.5199.5299.9185.35
UPerHead [27]31.5376.292.0294.398.2799.5899.4699.986.41
FCNHead [72]23.0470.9492.3795.097.999.4799.4499.8884.77
ALA [40]20.6469.7592.0394.2695.6999.4598.3799.983.76
GFM (Ours)36.4877.6592.7195.098.0499.6199.5699.9187.37
Table 9. The effects of different convolutions in DPC (the first and second best results are highlighted in red and underlined, respectively).
Table 9. The effects of different convolutions in DPC (the first and second best results are highlighted in red and underlined, respectively).
BaseMoudleConvolutionGravelSmallRockBigRockBedRockSandSoilRidgeSkymIoU (%)
DPCMConv35.6277.2192.494.7798.0999.6199.6699.9587.16
DWConv [73]35.3977.292.6995.098.2799.6199.499.8587.18
GConv [74]35.8677.2192.3394.7898.1999.6199.6399.9387.19
DCN [75]35.9277.2292.695.098.199.6199.4499.8687.22
DPC (Ours)36.4877.6592.7195.098.0499.6199.5699.9187.37
Table 10. The position of DPCM (the first result is highlighted in red).
Table 10. The position of DPCM (the first result is highlighted in red).
PositionStage 1Stage 2Stage3Stage 4Stage 5mIoU (%)
DPCM 86.07
86.83
86.92
86.5
87.37
Table 11. Run-time performance of LisseMars on Nvidia Jetson Xavier NX (NX).
Table 11. Run-time performance of LisseMars on Nvidia Jetson Xavier NX (NX).
MetricsSynMars-AirMarsScapesTianWen
Times (s)0.0480.040.044
mIoU (%)77.5861.4878.56
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, B.; Wang, F.; Li, Q.; Zheng, B.; Yao, M.; Xiao, X.; Qi, Y.; Cui, H.; Huang, X. LisseMars: A Lightweight Semantic Segmentation Model for Mars Helicopter. Aerospace 2025, 12, 1049. https://doi.org/10.3390/aerospace12121049

AMA Style

Lin B, Wang F, Li Q, Zheng B, Yao M, Xiao X, Qi Y, Cui H, Huang X. LisseMars: A Lightweight Semantic Segmentation Model for Mars Helicopter. Aerospace. 2025; 12(12):1049. https://doi.org/10.3390/aerospace12121049

Chicago/Turabian Style

Lin, Boyu, Fei Wang, Qingeng Li, Bo Zheng, Meibao Yao, Xueming Xiao, Yifan Qi, Hutao Cui, and Xiangyu Huang. 2025. "LisseMars: A Lightweight Semantic Segmentation Model for Mars Helicopter" Aerospace 12, no. 12: 1049. https://doi.org/10.3390/aerospace12121049

APA Style

Lin, B., Wang, F., Li, Q., Zheng, B., Yao, M., Xiao, X., Qi, Y., Cui, H., & Huang, X. (2025). LisseMars: A Lightweight Semantic Segmentation Model for Mars Helicopter. Aerospace, 12(12), 1049. https://doi.org/10.3390/aerospace12121049

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop