Next Article in Journal
A Remote Maintenance Support Method for Complex Equipment Based on Layered-MVC-B/S Integrated AR Framework
Previous Article in Journal
MPGH-FS: A Hybrid Feature Selection Framework for Robust Multi-Temporal OBIA Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EfficientSegNet: Lightweight Semantic Segmentation with Multi-Scale Feature Fusion and Boundary Enhancement

1
School of Instrument and Electronics, North University of China, Taiyuan 030051, China
2
School of Electrical and Control Engineering, North University of China, Taiyuan 030051, China
3
North Automatic Control Technology Institute, Taiyuan 030006, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(19), 5934; https://doi.org/10.3390/s25195934
Submission received: 11 August 2025 / Revised: 10 September 2025 / Accepted: 14 September 2025 / Published: 23 September 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

Semantic segmentation is a crucial task in computer vision with broad applications in autonomous driving, intelligent surveillance, drone vision, and other fields. The current high-precision segmentation models generally suffer from large parameter sizes, high computational complexity, and substantial memory consumption, which limits their efficient deployment in embedded systems and resource-constrained environments. In addition, traditional methods exhibit significant limitations in handling multi-scale targets and object boundaries, particularly during deep feature extraction, where the loss of shallow spatial information often results in blurred boundaries and reduced segmentation accuracy. To address these challenges, we propose EfficientSegNet, a lightweight and efficient semantic segmentation network. This network features an innovative architecture that integrates the Cascade-Attention Dense Field (CADF) module and the Dynamic Weighting Feature Fusion (DWF) module, effectively reducing computational resource requirements while balancing global semantic information and local detail recovery. Experimental results demonstrate that EfficientSegNet achieves an excellent balance between segmentation accuracy and computational efficiency on multiple public datasets, providing robust support for real-time segmentation tasks and applications on resource-constrained devices.

1. Introduction

With the continuous evolution of artificial intelligence and the widespread application of deep learning technologies, semantic segmentation has become one of the most essential tasks in computer vision. It plays a vital role in numerous real-world applications, such as autonomous driving [1], medical image analysis [2], intelligent surveillance [3], and drone-based scene understanding [4]. Unlike coarse-grained classification or bounding-box detection, semantic segmentation assigns a category label to each individual pixel in an image, thereby enabling high-precision scene interpretation. This fine-grained recognition capability forms the foundation for higher-level visual reasoning tasks, including object detection [5], spatial localization [6], target tracking [7], and comprehensive environmental perception [8].
Despite decades of research, designing robust and efficient semantic segmentation algorithms remains challenging. Earlier algorithmic solutions—although efficient in specific scenarios—largely depend on manually crafted low-level image features, which inherently lack adaptability to high intra-class variation or spatial context dynamics. Such limitations become particularly evident when segmenting cluttered environments with densely distributed targets or ambiguous boundaries [9,10,11].
With the rapid development of deep learning, Convolutional Neural Networks (CNNs) have gradually become dominant in semantic segmentation research. By learning from large amounts of image data, CNNs can automatically extract high-level features, achieving more accurate and refined semantic segmentation. However, these methods inherently rely on local convolution operations, which limit their ability to model long-range dependencies and often result in insufficient capture of global contextual information. To address the challenges posed by the coexistence of multi-scale targets and complex spatial contexts, a series of multi-scale feature fusion modules have been proposed, such as Atrous Spatial Pyramid Pooling (ASPP) and its extended version DenseASPP [12]. These methods significantly expand the receptive field and enhance the model’s ability to represent multi-scale features by introducing dilated convolutions or densely connected structures with different dilation rates, thereby improving the modeling of semantic consistency and spatial structural details in complex scenes. However, dilated structures are always sparsely sampled, which not only dilutes the feature response related to the target but also introduces significant background noise, resulting in the loss of local detail information, and determining a reasonable combination of dilation rates is still a problem that needs further exploration. Meanwhile, this architecture typically relies on feature concatenation or cascading operations, which can sharply increase channel dimensions and introduce substantial redundant information. This not only raises computational costs but also weakens the representation of effective features.
In recent years, Transformer-based semantic segmentation models have achieved state-of-the-art performance on multiple benchmark datasets. Although these models demonstrate excellent capabilities in modeling long-range dependencies and integrating global context, their high computational and memory overhead makes them difficult to deploy on resource-constrained mobile or embedded platforms [13]. In contrast, convolutional neural network (CNN)-based models remain the mainstream choice for embedded systems and mobile devices due to their efficiency and hardware compatibility.
To address these issues, attention mechanisms have been introduced into CNN-based semantic segmentation networks. By adaptively assigning weights to features, these mechanisms highlight important spatial regions and channel information while preserving global contextual cues, thereby alleviating feature redundancy, improving model efficiency, and enhancing semantic prediction consistency. Zhang et al. [14] effectively integrated global and local information through a global local attention module and multi-scale feature fusion, significantly improving the accuracy and real-time performance of drone image segmentation. Lian et al. [15] proposed a lightweight semantic segmentation network, which achieved efficient segmentation through ghost dependent separable revolution with efficient channel attention block and lightweight feature pyramid network. Mo et al. [16] enhanced the adaptive weighting of key channel and spatial position semantic information by introducing a unit attention module, thereby improving the segmentation accuracy of non-structured road drivable areas. Li et al. [17] used multi-layer cross attention module (MCAM) and efficient channel attention module (ECAM) to enhance the interaction between position and channel features and combined semantic aggregation pyramid pooling (SAPPM) to achieve multi-scale feature fusion.
Although the attention mechanism reduces feature redundancy by emphasizing key features, it only reweights existing features and cannot fundamentally recover the spatial structure lost through multiple downsamplings. To mitigate this problem, existing methods typically incorporate shallow features to supplement the representation. However, even state-of-the-art approaches struggle to strike a balance between semantic abstraction and spatial precision. Specifically, deep feature representations (rich in semantics) often lose detailed spatial structures because of repeated downsampling. Conversely, shallow features preserve local spatial textures but lack the semantic depth necessary for accurate region-level classification. The simple merging of these features fails to fully utilize their complementary advantages, often resulting in fuzzy object boundaries, semantic ambiguity, or redundancy in feature maps.
To address these challenges, we propose EfficientSegNet, a novel lightweight semantic segmentation framework that emphasizes the efficient integration of multi-scale semantic abstraction and boundary-aware spatial refinement. The core architecture is designed to bridge the semantic gap between deep and shallow features while ensuring minimal computational overhead, thereby enabling real-time segmentation on mobile or embedded platforms. Through the integration of hierarchical receptive fields, attention-guided fusion, and dynamic weighting mechanisms, EfficientSegNet aims to overcome the limitations of existing approaches in handling redundant information and preserving object boundaries in complex scenes.
The main contributions of this work are as follows:
(1)
Cascade Attention Dense Field (CADF) Module: The CADF module captures deep semantic information through progressively enlarged multi-scale receptive fields. Information sharing across scales is facilitated to maintain high-resolution representations and produce finer segmentation outputs.
(2)
Dynamic Weighting Fusion (DWF) Module: The DWF module supervises the fusion process of deep and shallow features by dynamically weighting feature contributions. This mechanism compensates for resolution-induced information loss and enhances boundary clarity.
(3)
Lightweight Architecture: EfficientSegNet employs a resource-efficient design that significantly reduces computational demands, enabling real-time inference on mobile or resource-constrained devices without compromising segmentation accuracy.

2. Related Work

The field of semantic segmentation has made significant progress, evolving from traditional image processing techniques to modern deep learning-based solutions, and more recently to Transformer-based architectures. Traditional methods, such as thresholding, edge detection, region growing, and graph-based approaches, operate on pixel-level intensity or texture features [18,19,20,21]. While computationally inexpensive, these techniques typically lack robustness in complex scenes, and their performance is highly sensitive to noise and scene variation.
With the rise of deep learning, convolutional neural networks (CNNs) have enabled end-to-end learning of semantic representations. Fully Convolutional Networks (FCN) were among the first to replace fully connected layers with convolutional layers, enabling dense pixel-wise predictions on images of arbitrary size [22]. This innovation laid the foundation for more advanced architectures such as U-Net, which introduced an encoder–decoder structure with skip connections to restore spatial detail [23]. Subsequent developments, including PSPNet [24] and HRNet [25], emphasized multi-scale contextual understanding and high-resolution feature preservation. Meanwhile, the DeepLab series integrated dilated convolutions and spatial pyramid pooling modules to enhance receptive field coverage and contextual aggregation [26,27,28,29], providing an important foundation for subsequent research on multi-scale feature fusion. Building on this, ConTriNet [30] proposed the Residual Atrous Spatial Pyramid Module (RASPM), which combines multi-branch atrous convolution with residual connections to aggregate multi-scale contextual information while preserving spatial details, thereby significantly improving feature representation capabilities. LiteFusionNet [31] designed Efficient Atrous Spatial Pyramid Pooling (EASPP), which reduces computational complexity by lowering the dilation rate and introduces additional upsampling operations to enhance the fusion of semantic and spatial information. SWDE Net [32] proposed Cross-Contextual Atrous Spatial Pyramid (CC-ASP), which adopts a cascaded structure to achieve stronger multi-scale contextual modeling. However, if dilated convolutions are stacked excessively or the dilation rate is improperly set, they not only struggle to cover key feature regions effectively but also introduce significant noise, leading to spatial structure blur and semantic information loss. In addition, multi-branch or cascaded structures often introduce substantial redundant computation, limiting inference efficiency.
To overcome the limitations of heavyweight models, researchers have focused on designing lightweight networks that reduce computational cost while preserving segmentation accuracy [33,34,35]. Depthwise separable convolutions, as utilized in the MobileNet family [36,37,38], significantly reduce the number of parameters and floating-point operations. ShuffleNet introduced channel shuffling and grouped convolutions to further enhance efficiency [39]. Meanwhile, some studies have introduced improved feature fusion modules or attention mechanisms to remove redundant features and highlight key semantics, thereby enhancing global semantic consistency while maintaining a lightweight design. SFRSeg [40] proposes a semantic-guided feature recombination strategy based on a lightweight backbone, achieving a better balance between accuracy and efficiency through multi-level feature fusion and attention enhancement. BiSeNet v3 [41] adopts a dual-branch design, effectively combining deep branches for extracting global contextual information with shallow branches for preserving local details via a feature fusion module. MRBANet [42] introduces a lightweight multi-resolution branch attention neural network that adaptively adjusts the fusion weights of each branch to highlight prominent contextual features. LMFDCNet [43] employs three sub-networks to extract multi-level features and utilizes a three-edge fusion module to enhance low-level details and high-level semantic features, thereby enabling more accurate object representation in complex environments. Although these methods achieve a good balance between accuracy and efficiency, they still face issues related to insufficient feature fusion or fixed fusion weights. The downsampling process is irreversible, and even with upsampling interpolation, the lost high-frequency texture and boundary information cannot be recovered. As a result, deep and shallow features become semantically and spatially inconsistent. Existing feature fusion methods mainly rely on concatenation or weighting strategies, lacking effective semantic alignment and spatial correction mechanisms, which limits the improvement of model performance in detail perception and boundary localization.
Recently, Transformer-based architectures have garnered attention for their capacity to capture global dependencies through self-attention mechanisms. SegFormer, for instance, introduces a hierarchical Transformer encoder tailored for segmentation tasks [44]. Swin Transformer enhances computational efficiency using a shifted window strategy and patch merging hierarchy [45], while Mask2Former integrates mask-based decoding with attention modules to improve segmentation quality [46]. Although these models achieve competitive performance, their practical deployment is hampered by high resource demands and dependency on large-scale annotated datasets [47,48,49].
Despite these diverse approaches, several key challenges remain unresolved: (1) Multiple downsampling in CNN backbone networks irreversibly discards high-frequency texture and boundary information, resulting in severe degradation of the spatial representation ability of deep features and making it difficult to accurately characterize small targets and fine-grained structures. (2) The fusion process of deep and shallow features lacks an adaptive adjustment mechanism, which makes it difficult to alleviate semantic and spatial misalignment. (3) Although multi-scale perception structures enhance feature representation, they often introduce a large amount of redundant computation and parameter overhead, limiting the practical deployment of models on resource-constrained platforms such as embedded or mobile devices.
To this end, our EfficientSegNet architecture is designed to unify deep semantic abstraction and shallow spatial cues through attention-guided modules and dynamic fusion strategies. By addressing both computational constraints and segmentation quality, the proposed model offers a practical and scalable solution for real-world semantic segmentation.

3. Methodology

3.1. Network Architecture

The overall architecture of the proposed EfficientSegNet is illustrated in Figure 1. The network enhances semantic representation and segmentation accuracy while maintaining computational efficiency. It consists of lightweight backbone network, CADF Module, and DWF Module.
In the feature extraction phase, EfficientSegNet employs MobileNetV3-large as the backbone, leveraging deformable convolutions to enhance channel attention and integrating a lightweight inverted residual block to improve feature reuse, thereby increasing the model’s representational capacity without significantly raising computational cost. For an input image of size H × W, the backbone extracts feature maps at three scales: Branch 1 (H/4 × W/4) captures low-level features containing edges and textures; Branch 2 (H/8 × W/8) contains richer structural information; and Branch 3 (H/16 × W/16) aggregates global semantic information. After extracting high-level features, the CADF module is introduced to further enhance feature representation, generating Branch 4.
During the feature fusion process, Branch 2 is first processed with pointwise convolution to reduce channel dimensions and then fused with Branch 1 using the DWF module to avoid scale mismatches while preserving edge details, resulting in Branch 5 (S5 = DWF(Conv1×1(S2), S1)). Similarly, Branch 3 undergoes pointwise convolution and is fused with Branch 2 via the DWF module to strengthen the combination of high-level semantic and mid-level structural features, yielding Branch 6 (S6 = DWF(Conv1×1(S3), S2)).
Subsequently, Branch 4 is processed with pointwise convolution to reduce channels and upsampled 2× to increase resolution before being concatenated with Branch 6 to form Branch 7 (S7 = Concat(Conv1×1(S4)_↑2, S6)), enriching high-level features with finer details. Further, Branch 7 is compressed with pointwise convolution and upsampled 2× before being concatenated with Branch 5 to form Branch 8 (S8 = Concat(Conv1×1(S7)_↑2, S5)), ensuring a comprehensive fusion of semantic and detailed information for clearer boundaries.
Finally, Branch 8 is processed with a 3 × 3 convolution to enhance feature expression and maintain spatial consistency, followed by a 4 upsampling operation that restores the segmentation result to the original input size, yielding the final semantic segmentation output (Sout = (Conv3×3(S8)_↑4)). This approach, through effective feature extraction, enhancement, and multi-scale fusion, improves segmentation accuracy without compromising computational efficiency, thereby providing an efficient solution for real-time semantic segmentation tasks.

3.2. Cascade Attention Dense Field

The Atrous Spatial Pyramid Pooling module captures multi-scale contextual features and extends the receptive field; however, its fixed atrous convolution configuration is unable to adaptively balance global context with local details and often introduces excessive redundant information during feature fusion, which impairs the expression of critical features. Convolutions with large dilation rates capture extensive context but may overlook fine details, whereas those with small dilation rates preserve high spatial resolution for detail capture yet lack sufficient global perception. Thus, balancing global information extraction with detail preservation remains a critical challenge.
To address this issue, we propose the CADF module that fully utilizes the complementarity between features of different scales to achieve effective multi-scale information fusion.
As shown in Figure 2, the CADF module first extracts multi-scale features from the input feature map Sin using four groups of dilated depthwise separable convolution with dilation rates r∈{3, 6, 12, 18}, thereby generating corresponding feature maps Fr:
F r = f r e c o n f i g ( S i n , r ) , r 3 , 6 , 12 , 18
After obtaining the multi-scale feature maps, these features are concatenated along the channel dimension. To address the redundancy and insufficient multi-scale fusion in the concatenated feature Fcat, we perform global average pooling along both the horizontal and vertical directions to extract long-range dependencies and positional information, thereby enabling the model to better understand contextual relationships in all directions.
f h ( c , x ) = 1 W i = 1 W F c a t ( c , x , i ) , x = 1 , , H
f w ( c , y ) = 1 H i = 1 H F c a t ( c , j , y ) , y = 1 , , W
This process produces vectors fhR5C×H and fwR5C×W representing horizontal and vertical information, respectively, which are concatenated along the spatial dimension to form a joint descriptive feature.
F j o i n t = C o n c a t ( [ f h , f w ] , dim = 1 ) R 5 C × ( H + W )
A subsequent pointwise convolution followed by a nonlinear activation function transforms this feature into an intermediate representation Fmid,
F m i d = δ ( C o n v 1 × 1 ( F j o i n t ) ) + b c o n v ) R 5 C × ( H + W )
where bconv and δ denote the bias and activation function, respectively, and C’ is the number of channels after transformation; this step integrates and compresses information from different directions to provide a compact representation for generating fine attention.
Fmid is then processed through two independent pointwise convolution layers to produce horizontal and vertical attention maps.
A h = σ ( C o n v 1 × 1 ( F m i d ) + b h ) R 5 C × H × 1
A w = σ ( C o n v 1 × 1 ( F m i d ) + b w ) R 5 C × 1 × W
with bh and bw as their respective biases and σ representing the Sigmoid activation function to normalize attention values to the [0, 1] range.
Finally, these attention weights are applied to the horizontal and vertical components of Fcat via element-wise multiplication, achieving dynamic recalibration.
S o u t = F c a t A h A w
This module not only preserves the advantages of multi-scale features but also effectively suppresses redundant information, thereby significantly enhancing the model’s ability to capture key regions and fine details in complex scenarios, ultimately providing more precise feature support for subsequent segmentation tasks and improving overall segmentation performance.

3.3. Dynamic Weighting Feature Fusion

In complex segmentation tasks, deep and shallow features exhibit inherent complementarity. Deep features encode rich semantic information, providing strong support for object recognition; however, their low resolution often leads to the loss of fine details. In contrast, shallow features preserve precise spatial information but struggle to capture sufficient global semantics. Simply adding or concatenating these features may fail to fully leverage their respective strengths and may even introduce redundant information.
To overcome this limitation, we propose the DWF module for fusing low-level and high-level semantic information. Its design employs dynamic weighting and fine-grained embedding to preserve the smoothness and naturalness of high-resolution images while fully integrating deep semantic features, thereby enhancing the representation of target features.
As depicted in Figure 3, the DWF module first processes the high-level feature Shigh with a convolution to obtain enhanced local features Fhigh. These features are then upsampled via bilinear interpolation to yield continuous, smooth high-resolution features Fup, which compensate for the spatial resolution loss in deep features and better transmit global semantic information at the detail level. Subsequently, Fup is fused with the shallow feature Slow, which retains complete spatial information, through element-wise multiplication to form the preliminary fused feature Ffuse (where denotes element-wise multiplication).
F f u s e = S l o w ( B i l i n e a r U p ( C o n v 3 × 3 ( S h i g h ) ) )
This direct pixel-wise fusion effectively combines the detailed information of shallow features with the semantic information of the upsampled deep features, laying the foundation for subsequent dynamic weighting.
However, simple fusion can introduce redundancy and fail to reconcile differences between the two feature types. To address this, the DWF module further processes Ffuse with a convolution followed by ReLU activation to obtain an intermediate feature Fnl.
F n l = R E L U ( C o n v 3 × 3 ( F f u s e ) )
Next, two dynamic weight masks, Mlow and Mhigh, are generated via pointwise convolution combined with a softmax function.
[ M l o w , M h i g h ] = s o f t m a x ( C o n v 1 × 1 ( F n l ) )
These masks are applied to the shallow and upsampled deep features, respectively, to achieve refined dynamic fusion.
S o u t = C o n v 1 × 1 ( S l o w M l o w + F u p M h i g h )
This dynamic weighting mechanism establishes a precise pixel-wise correspondence between global semantic information and local details, suppressing redundancy while enhancing shared semantics, thereby ensuring that key features are prominently expressed in the final output.
Finally, a pointwise convolution adjusts the fused features to match the input dimensions, ensuring the continuity of the network structure and the integrity of information transmission.

4. Experiments

4.1. Experimental Equipment

The embedded autonomous driving data acquisition system is shown in Figure 4. This system is built on the Yunle Intelligent Commuter Vehicle, with dimensions of 2179 mm (L) × 1300 mm (W) × 1529 mm (H) and a battery range of 60 km. It supports secondary development and autonomous driving. The video acquisition device used is the Orbbec (LeTV) Astra+, which is mounted at the front of the vehicle. For data processing, the system is equipped with a Jetson Nano mobile development board running the Ubuntu operating system, featuring an Arm Cortex-A78AE processor, an NVIDIA Ampere GPU, and 4 GB of memory.

4.2. Experimental Environment

The experiments in this study were implemented using Python 3.7, and all models were trained on an Ubuntu system. The hardware setup includes Intel Core i7-12850HX CPU and NVIDIA RTX A3000 GPU. The software environment consists of PyTorch 1.13.1, CUDA 11.6, and cuDNN 8.5.0.96. The key parameter settings include an image size of 512 × 512 pixels, the Adam adaptive learning rate optimization algorithm with an initial learning rate of 5 × 10−4, and a total of 100 training epochs.

4.3. Datasets

This study employs three publicly available datasets for comparative experiments: PASCAL VOC 2012 Augmented Dataset [50], PASCAL Context [51], and CityScapes [52].
PASCAL VOC 2012 Augmented Dataset consists of 1449 validation images, 10,582 training images, and 1456 test images, covering 20 semantic categories, including humans, common animals, vehicles, and furniture.
PASCAL Context provides detailed pixel-level annotations for 459 classes, with 4998 training images and 5105 validation images. This dataset includes both indoor and outdoor scenes, but in this study, only 15 commonly used indoor semantic categories are considered.
CityScapes, one of the most authoritative and professional image semantic segmentation benchmarks in autonomous driving, captures street scenes from 50 different cities across various seasons. It contains 2975 training images, 500 validation images, and 1525 test images, categorized into 31 semantic classes. However, this study only utilizes 19 commonly used categories.

4.4. Evaluation Metrics

To quantitatively assess the performance of the multi-class semantic segmentation models, this study adopts standard evaluation metrics, defined as follows:
m I o U = 1 k k = 1 k T P k T P k + F P k + F N k
M P A = 1 k k = 1 k T P k + T N k T P k + F P k + F N k + T N k
The performance evaluation of image segmentation models is based on a confusion matrix comprising four key components: True Positives (TP) denote the number of target class pixels correctly identified, reflecting the model’s ability to capture core features; True Negatives (TN) represent the number of non-target class pixels accurately excluded, indicating the model’s effectiveness in suppressing background noise; False Positives (FP) indicate the number of non-target pixels mistakenly classified as target pixels, which measures the false alarm rate; and False Negatives (FN) refer to the number of target pixels incorrectly omitted, reflecting the risk of missed detections.
Based on these definitions, this study employs Mean Intersection over Union (mIoU) and Mean Pixel Accuracy (MPA) to construct a comprehensive evaluation system. mIoU is computed as the mean ratio of the intersection to the union of predicted and ground truth regions across k classes, focusing on quantifying localization accuracy in challenging boundary cases—such as blurred edges and small object segmentation. In contrast, MPA averages the global pixel classification accuracy, thereby assessing the model’s overall robustness across different classes. Together, these metrics complement each other: mIoU highlights variations in segmentation quality among classes, while MPA verifies the model’s stability under imbalanced class distributions, ensuring a thorough and reliable performance evaluation.

5. Results

5.1. Ablation Experiment

The effectiveness of the multi-scale convolution within the CADF module and its impact on segmentation performance were investigated through multiple ablation experiments. This experiment did not consider depthwise separable convolution and Attention but primarily focused on the contribution of different combinations of dilation rates.
As shown in Table 1, the dilation rate combination of (3, 6, 12, 18) achieved the best performance across all metrics. This configuration preserves global contextual information while retaining local detail, thereby enhancing the model’s precision in delineating object boundaries. Specifically, a dilation rate of 3 aids in capturing small objects and edge textures, bolstering the model’s ability to represent fine structures; dilation rates of 6 and 12 provide medium-scale contextual information that completes the feature representation of target regions; and a dilation rate of 18 effectively covers a wider range of global information, improving semantic consistency in complex scenes. In contrast, the combination of (6, 12, 18) models a limited range of global information but lacks a branch for small dilation rates, weakening its capacity to delineate fine details and resulting in poor edge segmentation. Similarly, while (6, 12, 18, 24) enhances long-range dependency modeling by incorporating a larger dilation rate, the excessive receptive field leads to a loss of local details, thereby reducing segmentation accuracy for small objects and boundary areas. Moreover, although the (3, 6, 12, 18, 24) combination contains richer scale information, the additional branch with a dilation rate of 24 introduces redundant information, ultimately failing to improve segmentation performance and even diminishing the model’s efficiency.
To further investigate the joint influence of convolution types and attention mechanisms in CADF modules, we conducted ablation experiments with four different structural combinations: dilated convolution only (D-Conv), dilated depthwise separable convolution only (DDW Conv), dilated convolution combined with attention (D-Conv + attention), and dilated depthwise separable convolution combined with attention (DDW Conv + attention). From the overall results shown in Table 2, The segmentation accuracy of DDW Conv is slightly higher than that of D-Conv, and the parameter count decreases by 0.81 M. After further introducing the CA attention mechanism, the performance of both types of convolutional structures has been improved. DDW Conv + attention yields the best results, achieving a synergistic optimization of global context awareness and local feature refinement, while maintaining model efficiency. It is the optimal structural design choice in CADF module.
The results presented in Table 3 validate the effectiveness of the DWF module in the feature fusion process. In the model marked as ‘without transition layer’, the decoder directly concatenates and fuses shallow features (H/4 × W/4) from MobileNetV3 with deep features (H/16 × W/16) from ASPP, without employing any mechanism to reconcile the distributional differences between deep and shallow features, leading to semantic misalignment. With the introduction of a transition layer (H/8 × W/8), model performance is notably improved, indicating that transition-based supervision can mitigate multi-scale semantic mismatches. However, direct concatenation-based fusion tends to introduce substantial redundant information, which is not conducive to the lightweight deployment and real-time performance of the model. The DWF module addresses this issue by assigning adaptive weights to feature channels through a dynamic weighting mechanism, thereby reducing redundancy and enhancing the representation of critical information. This facilitates a synergistic optimization of segmentation accuracy and computational efficiency.

5.2. Module Ablation Experiment

The effectiveness of the proposed modules was rigorously validated through ablation experiments on the core components of EfficientSegNet, as presented in Table 4. We selected the classic multi-branch parallel structure DeepLabV3 as the baseline model, which achieved a mIoU ranging from 68.09% to 76.95%, serving as the reference for subsequent improvements.
Initially, MobileNetV3 was integrated into the backbone to reduce the model’s parameters and computational complexity. The experimental results indicated that although the parameter count decreased from 54.71 M to 4.84 M, the segmentation performance only experienced a slight drop. This demonstrates that the modified backbone effectively maintains segmentation capability while ensuring computational efficiency, thereby laying the groundwork for a lightweight design.
On the basis of lightweight backbone network, we conducted quantitative analysis on CADF module and DWF module, respectively. The experimental results show that when only the CADF module is introduced, although the computational burden caused by dense connections is alleviated by reducing the number of convolution channels and DDW Conv, the decoder still uses simple concatenation of large channels, which significantly increases FLOPs and parameters, and brings about a significant improvement in mIoU. When only the DWF module is introduced, compared to Baseline+V3, the parameter count is reduced by about 38%, and a moderate improvement in mIoU is achieved. This is because the DWF module first maps high-level channels to low-level channels during feature fusion, and then performs adaptive weighting, effectively avoiding the dilation caused by channel concatenation and significantly reducing the computational complexity of subsequent convolutions.
When the two are combined to build the complete EfficientSegNet, CADF replaces the original ASPP module, which had a high computational cost, significantly reducing the computational load through depthwise separable convolution and compressing the number of convolution channels from 256 to 112. Meanwhile, DWF avoids the computational overhead caused by concatenating high channels in the decoder, ensuring that subsequent convolutions are always performed with low channel counts. The complete model thus integrates the complementary lightweight advantages of DDW Conv and low channel counts, resulting in a further 0.49% reduction in overall computational complexity compared to Baseline+V3+DWF. Compared with the baseline, it not only outperforms all individually improved variants in accuracy but also achieves a substantial improvement of 4.8–5.9% in mIoU, while reducing the number of parameters by 94.6% and FLOPs by 96.9%. These results fully validate that EfficientSegNet can achieve efficient and accurate semantic segmentation while maintaining a lightweight structure.
In order to more intuitively verify the impact of each optimization module on model performance, Figure 5 presents the attention maps under different configurations. Since the differences in attention are mainly concentrated in small target areas, we enlarged and highlighted the corresponding regions. The results show that after introducing the CADF module, the model exhibits stronger responses to small-scale targets, indicating that the module can effectively integrate multi-scale features and improve the recognition of small objects. With the introduction of the DWF module, the attention shifts more toward the target itself rather than the background or edge regions, demonstrating that its dynamic weighting mechanism can suppress redundant information and highlight key regions during feature selection. When the two modules are combined, the model shows significant improvements in suppressing background noise and focusing on the target, thereby verifying from a visual perspective the complementary roles of the two modules in enhancing overall performance.

5.3. Contrast Experiment

To further validate the effectiveness and generalizability of our model in complex semantic segmentation tasks, experiments were conducted on various datasets and compared with other state-of-the-art models, including U-Net [23], PSPnet [24], DeepLab V3+ [29], SFRSeg [37], BiseNet v3 [38] and Mask2Former [41].
As shown in Table 5, on the PASCAL VOC, CityScapes, and PASCAL Context datasets, the mIoU of EfficientSegNet is 1.47%, 7.21%, and 1.67% lower than that of PSPNet, respectively. However, its parameter count is smaller and its prediction speed is faster. On the PASCAL VOC and CityScapes datasets, its segmentation accuracy is comparable to that of BiseNet v3, with differences of 0.39% and 0.81%, respectively. On the PASCAL Context dataset, however, its segmentation accuracy is 5.18% higher, while also achieving higher inference speed. Compared with the lightweight architecture SFRSeg, EfficientSegNet has slightly higher segmentation accuracy, but with an increase of 1.32 M in parameter count, its real-time performance is slightly inferior. Compared with other models, EfficientSegNet performs well in both segmentation accuracy and efficiency. Overall, EfficientSegNet achieves a good balance between segmentation accuracy and efficiency, demonstrating strong overall advantages.
To more intuitively compare the segmentation performance of various models, a series of experiments were conducted on multiple datasets, focusing on aspects such as single-class multi-scale segmentation capability, multi-class multi-scale segmentation capability, edge clarity, and depth detail preservation.
As shown in Figure 6, in the visualization experiments on the PASCAL VOC dataset, due to the inherent limitations of the encoder–decoder architecture, U-Net and PSPNet performed poorly in handling small and distant targets as well as fine-grained details such as bird beaks, legs, and aircraft tails. DeepLabv3+ has improved semantic consistency in mesoscale scenes through the ASPP module, but errors still occur in large-scale bus regions and aircraft tail fins. Mask2Former emphasizes global context modeling but still falls short in capturing local details, especially in complex structures such as bird legs and aircraft tail fins. Moreover, its global modeling strategy incurs relatively high computational overhead. The shallow semantic representation capacity of BiSeNetV3 is limited, which is mainly reflected in segmentation errors along bus edges. SFRSeg supports fast inference but lacks effective modeling of long-range dependencies, which often leads to semantic inconsistencies within the target, such as disconnections observed in the tail region of an aircraft. In contrast, EfficientSegNet extends the receptive field through CADF modules, effectively preserving global semantic consistency and detailed local structures. For example, it can accurately capture small-scale objects such as buses, trains, and airplane tail fins. Meanwhile, the DWF module adaptively assigns higher weights to boundary-sensitive features, better characterizes the details of complex contours such as bird feet and beaks.
The PASCAL Context dataset is mainly used to evaluate the generalization ability of models in unstructured indoor environments with enclosed spaces, the result is shown in Figure 7. U-Net has difficulty accurately separating corner lights and table legs. PSPNet shows instability when dealing with overlapping regions at wall corners. DeepLabv3+ and SFRSeg exhibit significant semantic discontinuities in structures such as beds and doors. Mask2Former provides better perception of overall structure but still fails to recognize corner lights, table legs, and continuous flooring. BiSeNetV3 lacks wall features in overlapping areas of Multi-class multi-scale scenes. In contrast, EfficientSegNet exhibits strong generalization ability and robustness in complex indoor scenes. The CADF module employs multi-scale dilated convolutions to capture long-range dependencies, enabling the model to maintain semantic consistency between partially occluded and distant targets. This is particularly evident in single-class, multi-scale scenes, where EfficientSegNet preserves contour integrity of lamps even near corners and occluded areas. In addition, the DWF module compensates for missing boundary features in deep networks, facilitating the distinction of highly similar objects such as beds, sofas, tables, and chairs. It also effectively restores connection areas between desktops and table legs, as well as between doors and walls. Moreover, EfficientSegNet demonstrates stronger depth perception and successfully handles complex scenes with cluttered foregrounds, including ground and chairs.
Figure 8 shows the visualization results of the comparison models on the CityScapes dataset. U-Net is prone to losing small targets such as traffic poles and bicycles in complex scenes. PSPNet improves its global perception ability with the help of the pyramid pooling structure, but it does not provide sufficient features for edge segmentation in areas with dense small targets, such as traffic signs. DeepLabv3+ is insufficient for segmenting the smallest targets, such as bicycle wheels. Mask2Former maintains relatively stable large-scale area segmentation in urban street scenes, but its local feature modeling is weak and cannot accurately segment detailed structures such as traffic signs. The shallow feature representation of BiSeNetV3 is weak, leading to confusion in the segmentation of complex background regions. SFRSeg exhibits structural fragmentation and semantic drift in overlapping areas involving buildings, traffic signals, and other background elements. In contrast, EfficientSegNet shows significant improvement in multi-scale urban scenarios. The CADF module provides clear advantages for multi-scale object recognition, effectively restoring the detailed forms of traffic signs of different sizes while improving the perception of subtle structures such as bicycle wheels and fences. In complex urban scenes with multiple categories, such as pedestrians, vehicles, and buildings, the model maintains stable and detailed segmentation performance, demonstrating strong multi-category adaptability and regional consistency. Furthermore, the DWF module enhances boundary discrimination and optimizes the feature representation of challenging areas such as bridge intersections, traffic signals, and building edges.
In addition, on-site semantic segmentation tests were conducted through the embedded autonomous driving data acquisition system to further validate the model’s generalization ability in real-world application scenarios. The test route started from the South Ring Road of North University of China, driving east to the East Ring Road, then extending north to Central Avenue, then turning west to Xingzhi East Road, and finally returning to the starting point, with a total length of about 850 m.
As shown in Figure 9, in the testing of campus road scenes, other models still exhibit issues such as errors, blurred contours, and semantic interruptions. EfficientSegNet, however, produces consistent segmentation results for vehicles of different scales, regardless of whether the objects appear in the center or at the edge of the image. This improvement is largely attributed to the CADF module, which captures cross-scale context and ensures stable semantic consistency between near and distant regions. In scenes containing multiple object categories, such as pedestrians, non-motorized vehicles, and motor vehicles, EfficientSegNet maintains a high degree of category differentiation, even for small-scale targets such as bicycles. The DWF module adaptively emphasizes fine-grained boundary cues and suppresses noise features introduced by shadows and light rays.
Although the collaborative mechanism of CADF and DWF enables EfficientSegNet to achieve robust segmentation performance in scenes with complex interference, there are still issues with incomplete feature extraction and difficulty in focusing on the target object when encountering elongated structures such as fences, streetlights, and table legs, as well as highly overlapping complex backgrounds. In addition, although the model has a certain ability to suppress noise, it is still prone to category confusion in areas with severe lighting changes (such as the intersection of sidewalks and roads), which affects the continuity and accuracy of the segmentation results.

6. Conclusions

In this paper, we propose EfficientSegNet, a segmentation model that achieves both high precision and a lightweight design. We theoretically and experimentally demonstrate the importance of integrating deep and shallow features for accurate segmentation. Compared to traditional models, EfficientSegNet effectively balances global semantic understanding with local detail recovery.
This study employs a lightweight backbone network to reduce computational overhead while enhancing embedded deployment performance. Additionally, we introduce the CADF module, which dynamically captures multi-scale information using adaptive atrous convolutions and an improved attention mechanism. This design significantly improves the recognition of road signs, obstacles, and edge details in complex environments. Furthermore, the DWF module efficiently integrates deep and shallow features through bilinear interpolation-based upsampling and pixel-wise weighting. This approach mitigates information loss due to downsampling, enabling precise fine-structure characterization while maintaining global semantic consistency.
Experimental results demonstrate that EfficientSegNet not only achieves superior mIoU performance on standard datasets but also significantly outperforms existing models in terms of robustness when processing dynamic scenarios and complex road environments. Ablation studies further confirm the critical role of the CADF and DWF modules in enhancing segmentation performance, particularly in improving edge detection and small-object segmentation, while maintaining computational efficiency. We argue that this balanced approach combining lightweight architecture and high segmentation accuracy has significant theoretical and practical value.
In conclusion, EfficientSegNet provides a novel solution for effectively integrating deep semantic information with fine-grained spatial details under limited computational resources. Future research will focus on exploring more advanced feature representation mechanisms to enhance the model’s ability to capture slender structural targets while further improving its robustness in environments with severe lighting variations and highly dynamic scenes.

Author Contributions

Conceptualization, L.Z. and P.Z.; methodology, L.Z.; software, L.Z.; validation, L.Z., P.Z. and M.L.; formal analysis, P.L.; investigation, P.Z.; resources, P.L.; data curation, L.Z.; writing—original draft preparation, L.Z.; writing—review and editing, P.L., P.Z. and M.L.; visualization, L.Z.; supervision, M.L.; project administration, P.Z.; funding acquisition, P.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Defence Fund (2023-JJ-0353), and the National Natural Foundation of China Grant No. 62373247.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The public sources of the data mentioned in this study are described in the paper.

Acknowledgments

We thank the anonymous reviewers for their valuable suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Elhassan, M.A.M.; Zhou, C.; Khan, A.; Benabid, A.; Adam, A.B.M.; Mehmood, A.; Wambugu, N. Real-time semantic segmentation for autonomous driving: A review of CNNs, Transformers, and beyond. J. King Saud Univ. Comput. Inf. Sci. 2024, 36, 102226. [Google Scholar] [CrossRef]
  2. Lyu, J.; Meng, J.; Zhang, Y.; Ling, S.H. MUF-Net: A Novel Self-Attention Based Dual-Task Learning Approach for Automatic Left Ventricle Segmentation in Echocardiography. Sensors 2025, 25, 2704. [Google Scholar] [CrossRef] [PubMed]
  3. Zeng, F.; Wang, R.; Jiang, Y.; Liu, Z.; Ding, Y.; Dong, W.; Xu, C.; Zhang, D.; Wang, J. Growth monitoring of rapeseed seedlings in multiple growth stages based on low-altitude remote sensing and semantic segmentation. Comput. Electron. Agric. 2025, 232, 110135. [Google Scholar] [CrossRef]
  4. Zhang, P.; Zhang, S.; Wang, J.; Sun, X. Identifying rice lodging based on semantic segmentation architecture optimization with UAV remote sensing imaging. Comput. Electron. Agric. 2024, 227, 109570. [Google Scholar] [CrossRef]
  5. Acharya, D.; Tennakoon, R.; Muthu, S.; Khoshelham, K.; Hoseinnezhad, R.; Bab-Hadiashar, A. Single-image localisation using 3D models: Combining hierarchical edge maps and semantic segmentation for domain adaptation. Autom. Constr. 2022, 136, 104152. [Google Scholar] [CrossRef]
  6. Feng, H.; Liu, W.; Xu, H.; He, J. A lightweight dual-branch semantic segmentation network for enhanced obstacle detection in ship navigation. Eng. Appl. Artif. Intell. 2024, 136, 108982. [Google Scholar] [CrossRef]
  7. Kos, A.; Majek, K.; Belter, D. SegTrackDetect: A window-based framework for tiny object detection via semantic segmentation and tracking. SoftwareX 2025, 30, 102110. [Google Scholar] [CrossRef]
  8. Zhang, L.; Wei, Z.; Xiao, Z.; Ji, A.; Wu, B. Dual hierarchical attention-enhanced transfer learning for semantic segmentation of point clouds in building scene understanding. Autom. Constr. 2024, 168, 105799. [Google Scholar] [CrossRef]
  9. Zhang, H.; Li, L.; Xie, X.; He, Y.; Ren, J.; Xie, G. Entropy guidance hierarchical rich-scale feature network for remote sensing image semantic segmentation of high resolution. Appl. Intell. 2025, 55, 528. [Google Scholar] [CrossRef]
  10. Wang, X.; Zhang, Y.; Ca, J.; Qin, Q.; Feng, Y.; Yan, J. Semantic segmentation network for mangrove tree species based on UAV remote sensing images. Sci. Rep. 2024, 14, 29860. [Google Scholar] [CrossRef]
  11. Li, Z.R.; Yan, L.; Leng, Z.; Liu, X.; Zhaodong, F.; Zhou, Q. Enhancing precision in automated segmentation of ambiguous boundaries based on multi-model deep learning and boundary confidence computation. Int. J. Radiat. Oncol. 2024, 120, 640. [Google Scholar] [CrossRef]
  12. Yang, M.; Yu, K.; Zhang, C.; Li, Z.; Yang, K. DenseASPP for Semantic Segmentation in Street Scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 3684–3692. [Google Scholar]
  13. Zhao, S.; Wu, X.; Tian, K.; Yuan, Y. A Transformer-Based Hierarchical Hybrid Encoder Network for Semantic Segmentation. Neural Process. Lett. 2025, 57, 66. [Google Scholar] [CrossRef]
  14. Zhang, Z.; Li, G. UAV Imagery Real-Time Semantic Segmentation with Global–Local Information Attention. Sensors 2025, 25, 1786. [Google Scholar] [CrossRef]
  15. Lian, X.; Kang, M.; Tan, L.; Sun, X.; Wang, Y. LSSMask: A lightweight semantic segmentation network for dynamic object. Signal Image Video Process. 2025, 19, 216. [Google Scholar] [CrossRef]
  16. Mo, X.; Feng, Y.; Liu, Y. Deep semantic segmentation for drivable area detection on unstructured roads. Comput. Vis. Image Underst. 2025, 259, 104420. [Google Scholar] [CrossRef]
  17. Li, T.; Yang, X.; Zhang, Z.; Cui, Z.; Maoxia, Z. Mix-layers semantic extraction and multi-scale aggregation transformer for semantic segmentation. Complex Intell. Syst. 2024, 11, 36. [Google Scholar] [CrossRef]
  18. Matić, T.; Aleksi, I.; Hocenski, Z.; Kraus, D. Real-time biscuit tile image segmentation method based on edge detection. ISA Trans. 2018, 76, 246–254. [Google Scholar] [CrossRef]
  19. Li, J.; Tang, W.; Wang, J.; Zhang, X. Multilevel thresholding selection based on variational mode decomposition for image segmentation. Signal Process. 2018, 147, 80–91. [Google Scholar] [CrossRef]
  20. Cheng, Z.; Wang, J. Improved region growing method for image segmentation of three-phase materials. Powder Technol. 2020, 368, 80–89. [Google Scholar] [CrossRef]
  21. Azam, S.; Rony, M.A.H.; Raiaan, M.A.K.; Fatema, K.; Karim, A.; Jonkman, M.; Beissbarth, J.; Leach, A.; De Boer, F. Reimagining otitis media diagnosis: A fusion of nested U-Net segmentation with graph theory-inspired feature set. Array 2024, 23, 100362. [Google Scholar] [CrossRef]
  22. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  23. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  24. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar]
  25. Wang, J.; Sun, K.; Chen, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep High-Resolution Representation Learning for Visual Recognition. arXiv 2020, arXiv:1904.04514. [Google Scholar] [CrossRef] [PubMed]
  26. Chen, L.; Papandreou, G.; Kokkinos, L.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected CRFs. arXiv 2016, arXiv:1412.7062. [Google Scholar] [CrossRef]
  27. Chen, L.; Papandreou, G.; Kokkinos, L.; Murphy, K.; Yuille, A.L. DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. arXiv 2017, arXiv:1606.00915. [Google Scholar] [CrossRef]
  28. Chen, L.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2018, arXiv:1706.05587. [Google Scholar]
  29. Chen, L.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. arXiv 2018, arXiv:1802.02611. [Google Scholar] [CrossRef]
  30. Tang, H.; Li, Z.; Zhang, D.; He, S.; Tang, J. Divide-and-Conquer: Confluent Triple-Flow Network for RGB-T Salient Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 47, 1958–1974. [Google Scholar] [CrossRef]
  31. Khan, A.; Khan, S.; Elhassan, M.A.M.; Deng, H. Advancing Road Safety: A Lightweight Feature Fusion Model for Robust Road Crack Segmentation. Adv. Eng. Inform. 2025, 65, 103262. [Google Scholar] [CrossRef]
  32. Wei, X.; Li, J.; Zhang, X.; Gu, H.; Van de Weghe, N.; Huang, H. An Innovative Framework Combining a CNN-Transformer Multiscale Fusion Network and Spatial Analysis for Cycleway Extraction Using Street View Images. Sustain. Cities Soc. 2025, 126, 106384. [Google Scholar] [CrossRef]
  33. Song, M.; Wang, Z.; Chen, Y.; Li, Y.; Jin, Y.; Jia, B. A lightweight multi scale fusion network for IGBT ultrasonic tomography image segmentation. Sci. Rep. 2025, 15, 888. [Google Scholar] [CrossRef]
  34. Ma, T.; Wang, K.; Hu, F. LMU-Net: Lightweight U-shaped network for medical image segmentation. Med. Biol. Eng. Comput. 2024, 62, 61–70. [Google Scholar] [CrossRef]
  35. Lu, S.; Chen, Y.; Chen, Y.; Li, P.; Sun, J.; Zheng, C.; Zou, Y.; Liang, B.; Li, M.; Jin, Q.; et al. General lightweight framework for vision foundation model supporting multi-task and multi-center medical image analysis. Nat. Commun. 2025, 16, 2097. [Google Scholar] [CrossRef]
  36. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  37. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  38. Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. arXiv 2019, arXiv:1905.02244. [Google Scholar] [CrossRef]
  39. Ma, N.; Zhang, X.; Zheng, H.; Sun, J. ShuffleNet V2: Practical guidelines for efficient CNN architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Volume 11218, pp. 122–138. [Google Scholar]
  40. Singha, T.; Pham, D.; Krishna, A. A real-time semantic segmentation model using iteratively shared features in multiple sub-encoders. Pattern Recognit. 2023, 140, 109557. [Google Scholar] [CrossRef]
  41. Tsai, T.; Tseng, Y. BiSeNet V3: Bilateral segmentation network with coordinate attention for real-time semantic segmentation. Neurocomputing 2023, 532, 33–42. [Google Scholar] [CrossRef]
  42. Dai, Y.; Wang, Y.; Meng, L.; Ren, J. A Multiple Resolution Branch Attention Neural Network for Scene Understanding of Intelligent Autonomous Platform. Expert Syst. Appl. 2025, 267, 126253. [Google Scholar] [CrossRef]
  43. Dai, Y.; Meng, L.; Sun, F.; Wang, S. Lightweight Multi-Scale Feature Dense Cascade Neural Network for Scene Understanding of Intelligent Autonomous Platform. Expert Syst. Appl. 2025, 259, 125354. [Google Scholar] [CrossRef]
  44. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with Transformers. arXiv 2021, arXiv:2105.15203. [Google Scholar] [CrossRef]
  45. Lu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  46. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-attention mask transformer for universal image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 1290–1299. [Google Scholar]
  47. Zhang, L.; Huang, W.; Fan, B. SARFormer: Segmenting Anything Guided Transformer for semantic segmentation. Neurocomputing 2025, 635, 129915. [Google Scholar] [CrossRef]
  48. Liu, W.; Zhang, A. Plant disease detection algorithm based on efficient Swin Transformer. Comput. Mater. Continua 2025, 82, 3045–3068. [Google Scholar]
  49. Zhu, P.; Liu, J. Joint U-Nets with hierarchical graph structure and sparse Transformer for hyperspectral image classification. Expert Syst. Appl. 2025, 275, 127046. [Google Scholar] [CrossRef]
  50. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The PASCAL Visual Object Classes (VOC) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  51. Mottaghi, R.; Chen, X.; Liu, X.; Cho, N.; Lee, S.W.; Fidler, S.; Urtasun, R.; Yuille, A.L. The role of context for object detection and semantic segmentation in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 891–898. [Google Scholar]
  52. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
Figure 1. Schematic of EfficientSegNet network structure.
Figure 1. Schematic of EfficientSegNet network structure.
Sensors 25 05934 g001
Figure 2. Schematic of CADF module structure.
Figure 2. Schematic of CADF module structure.
Sensors 25 05934 g002
Figure 3. Schematic of DWF module structure.
Figure 3. Schematic of DWF module structure.
Sensors 25 05934 g003
Figure 4. Embedded acquisition system for autonomous driving.
Figure 4. Embedded acquisition system for autonomous driving.
Sensors 25 05934 g004
Figure 5. Module ablation experiment attention maps.
Figure 5. Module ablation experiment attention maps.
Sensors 25 05934 g005aSensors 25 05934 g005b
Figure 6. PASCAL VOC visualization results. The yellow box highlights the segmented region of interest.
Figure 6. PASCAL VOC visualization results. The yellow box highlights the segmented region of interest.
Sensors 25 05934 g006aSensors 25 05934 g006b
Figure 7. PASCAL Context visualization results. The yellow box highlights the segmented region of interest.
Figure 7. PASCAL Context visualization results. The yellow box highlights the segmented region of interest.
Sensors 25 05934 g007aSensors 25 05934 g007b
Figure 8. CityScapes visualization results. The yellow box highlights the segmented region of interest.
Figure 8. CityScapes visualization results. The yellow box highlights the segmented region of interest.
Sensors 25 05934 g008
Figure 9. Campus road scenes visualization results. The yellow box highlights the segmented region of interest.
Figure 9. Campus road scenes visualization results. The yellow box highlights the segmented region of interest.
Sensors 25 05934 g009
Table 1. CADF module expansion rate ablation experiment data.
Table 1. CADF module expansion rate ablation experiment data.
(6, 12, 18)(3, 6, 12, 18)(6, 12, 18, 24)(3, 6, 12, 18, 24)
Parameter (M)5.055.755.756.48
FLOPs (G)76.1788.8588.85101.67
mIoU
(%)
PASCAL VOC76.8178.678.4378.79
PASCAL Context67.1469.9369.0670.26
CityScapes66.8468.668.3368.85
MPA
(%)
PASCAL VOC87.588.388.2889.05
PASCAL Context78.5780.3480.4580.65
CityScapes77.7278.7178.579.06
Table 2. Ablation results of different CADF structural configurations.
Table 2. Ablation results of different CADF structural configurations.
D-ConvDDW ConvD-Conv + AttentionDDW Conv + Attention
Parameter (M)5.754.945.875.07
mIoU (%)PASCAL VOC78.678.8280.5180.73
PASCAL Context69.9369.8571.7771.69
CityScapes68.668.6870.3070.38
MPA
(%)
PASCAL VOC88.389.0789.6890.45
PASCAL Context80.3480.7781.3281.75
CityScapes78.7179.1280.3880.79
Table 3. Ablation results of DWF module structural design.
Table 3. Ablation results of DWF module structural design.
Without Transition LayerConcatenation Transition LayerDWF
Parameter (M)4.846.23.07
mIoU (%)PASCAL VOC76.2277.2477.74
PASCAL Context67.6369.269.5
CityScapes66.5668.3267.97
MPA
(%)
PASCAL VOC89.3490.2689.99
PASCAL Context78.5278.579.33
CityScapes77.1476.9677.71
Table 4. Module ablation experiment data.
Table 4. Module ablation experiment data.
BaselineBaseline+V3Baseline+V3+CADFBaseline+V3+DWFBaseline+V3+CADF+DWF
Parameter (M)54.714.845.073.072.92
FLOPs (G)16748.4887.35.725.23
mIoU
(%)
PASCAL VOC76.9576.2280.7377.7481.13
PASCAL Context70.7167.6371.6969.574.9
CityScapes68.0966.5670.3867.9771.39
mPA
(%)
PASCAL VOC79.9589.3490.4589.9990.46
PASCAL Context80.178.5281.7579.3390.2
CityScapes76.9677.1480.7977.7182.04
Table 5. Contrast experiment data.
Table 5. Contrast experiment data.
ModelmIoU (%)Parameter
(M)
Average FPS
PASCAL VOCCityScapesPASCAL Context GPUJetson Nano
U-Net67.5360.656.5943.9323.335.55
PSPnet82.678.676.5746.7131.684.86
DeepLab V3+76.9568.0970.7154.7111.32.58
Mask2Former74.162.159.743.9633.911.2
BiseNet v380.7472.269.7239.3756.4910.51
SFRSeg79.370.661.41.619441.1
EfficientSegNet81.1371.3974.92.9276.5631.88
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Li, M.; Zhang, P.; Liu, P. EfficientSegNet: Lightweight Semantic Segmentation with Multi-Scale Feature Fusion and Boundary Enhancement. Sensors 2025, 25, 5934. https://doi.org/10.3390/s25195934

AMA Style

Zhang L, Li M, Zhang P, Liu P. EfficientSegNet: Lightweight Semantic Segmentation with Multi-Scale Feature Fusion and Boundary Enhancement. Sensors. 2025; 25(19):5934. https://doi.org/10.3390/s25195934

Chicago/Turabian Style

Zhang, Le, Mengwei Li, Peng Zhang, and Peng Liu. 2025. "EfficientSegNet: Lightweight Semantic Segmentation with Multi-Scale Feature Fusion and Boundary Enhancement" Sensors 25, no. 19: 5934. https://doi.org/10.3390/s25195934

APA Style

Zhang, L., Li, M., Zhang, P., & Liu, P. (2025). EfficientSegNet: Lightweight Semantic Segmentation with Multi-Scale Feature Fusion and Boundary Enhancement. Sensors, 25(19), 5934. https://doi.org/10.3390/s25195934

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop