Next Article in Journal
On the Identification of Orthotropic Elastic Stiffness Using 3D Guided Wavefield Data
Next Article in Special Issue
Robust 6-DoF Pose Estimation under Hybrid Constraints
Previous Article in Journal
GAN-FDSR: GAN-Based Fault Detection and System Reconfiguration Method
Previous Article in Special Issue
A Novel Memory and Time-Efficient ALPR System Based on YOLOv5
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unified DeepLabV3+ for Semi-Dark Image Semantic Segmentation

by
Mehak Maqbool Memon
1,
Manzoor Ahmed Hashmani
1,*,
Aisha Zahid Junejo
1,
Syed Sajjad Rizvi
2 and
Kamran Raza
3
1
High Performance Cloud Computing Center (HPC3), Department of Computer and Information Sciences, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Malaysia
2
Department of Computer Science, Shaheed Zulfiqar Ali Bhutto Institute of Science and Technology, Karachi 75600, Pakistan
3
Faculty of Engineering Science and Technology, Iqra University, Karachi 75600, Pakistan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(14), 5312; https://doi.org/10.3390/s22145312
Submission received: 9 May 2022 / Revised: 20 June 2022 / Accepted: 23 June 2022 / Published: 15 July 2022
(This article belongs to the Special Issue Machine Learning in Robust Object Detection and Tracking)

Abstract

:
Semantic segmentation for accurate visual perception is a critical task in computer vision. In principle, the automatic classification of dynamic visual scenes using predefined object classes remains unresolved. The challenging problems of learning deep convolution neural networks, specifically ResNet-based DeepLabV3+ (the most recent version), are threefold. The problems arise due to (1) biased centric exploitations of filter masks, (2) lower representational power of residual networks due to identity shortcuts, and (3) a loss of spatial relationship by using per-pixel primitives. To solve these problems, we present a proficient approach based on DeepLabV3+, along with an added evaluation metric, namely, Unified DeepLabV3+ and S 3 c o r e , respectively. The presented unified version reduced the effect of biased exploitations via additional dilated convolution layers with customized dilation rates. We further tackled the problem of representational power by introducing non-linear group normalization shortcuts to solve the focused problem of semi-dark images. Meanwhile, to keep track of the spatial relationships in terms of the global and local contexts, geometrically bunched pixel cues were used. We accumulated all the proposed variants of DeepLabV3+ to propose Unified DeepLabV3+ for accurate visual decisions. Finally, the proposed S 3 c o r e evaluation metric was based on the weighted combination of three different accuracy measures, i.e., the pixel accuracy, IoU (intersection over union), and Mean BFScore, as robust identification criteria. Extensive experimental analysis performed over a CamVid dataset confirmed the applicability of the proposed solution for autonomous vehicles and robotics for outdoor settings. The experimental analysis showed that the proposed Unified DeepLabV3+ outperformed DeepLabV3+ by a margin of 3% in terms of the class-wise pixel accuracy, along with a higher S 3 c o r e , depicting the effectiveness of the proposed approach.

1. Introduction

The scene comprehension of urban environments is the most crucial component of the autonomous vehicle industry. Since the comprehension and understanding of visual environments include dynamic visual environments, the deployment of scene comprehension demands the optimal discovery of all the objects present in the incoming visual scene, even if the incoming image is captured in low-light conditions. The rapid advancements in the domain of the deep learning (DL) paradigm illustrate promising solutions that focus on different problem domains. The methods proposed in the DL paradigm present satisfactory performance for the classification, detection, and segmentation of normal (bright) images. However, these methods display performance degradation when it comes to semi-dark imagery [1,2]. Furthermore, for visual image comprehension in an urban setting, the automated solution will face diversity in terms of lightning conditions. For scene comprehension in autonomous vehicles, a widely used alternative is semantic segmentation. Semantic segmentation results in a pixel-wise understanding of a visual image. Each pixel of an image receives an object label, which then helps to create an object mask that presents the global context of the image, and hence, the autonomous application (vehicle or robot) can make an informed decision. Recent research showed that DL-based semantic segmentations can be successfully employed for remote sensing imagery, including radar images [3] and high-resolution (HR) images from indoor, outdoor, agriculture, and industrial settings [4,5,6]. By leveraging the advancements of deep neural networks (DNNs) in the DL paradigm, more focused solutions can be proposed to deal with the diversity (including semi-dark imagery) involved in visual scene understanding. For the sematic segmentation of semi-dark imagery, dark pixel values pose great challenges to DNNs for the classification of HR images. Traditional fully convolution neural networks identify the object class to a certain extent but cannot handle variable image sizes and suffer from the retainment of the spatial component of a classified pixel [7]. To deal with these problems, a different architectural mechanism was presented, namely, an encoder–decoder approach [8]. This approach tackles problems associated with earlier CNNs that use an encoder, which encapsulates the image information by extracting image features. This feature extraction is accomplished by using any pretrained network. The encoder is then followed by a decoder, which projects the discriminative features (with a low resolution) coming from the encoder over a high-resolution pixel mask [9]. Over time, a lot of different variations of the encoder–decoder architecture were presented [10,11,12]. Encoder–decoders were widely used for automated scene comprehension. Specifically, the DeepLab family has evolved rapidly and has made innovative achievements [10,13,14]. However, even with the recent developments of DeepLab, the optimal semantic segmentation of semi-dark images remains an open area of research. The functionality of the state-of-the-art semantic segmentation modules is constrained by several factors. First, restricted multiscale feature extraction is witnessed due to the usage of biased centric exploitations of the receptive field. The receptive field is the portion of the image that is under consideration by the DCNN for the feature extraction. Centric exploitation results in a greater weightage being assigned to the center pixels and less weightage assigned to the corner pixels for the calculation of output features. This implies that the multiscale information residing in the corners of the receptive field is likely to be skipped by the network. Furthermore, the pretrained network used for an encoder, such as ResNet [12], results in reduced representational power and suffers from network training problems. These problems are attributed to the presence of identity shortcuts in the DCNN and the occurrence of an internal covariate shift during the training phase [15]. Finally, the loss of spatial relationships is also observed as the pixels are classified at the pixel level by using per-pixel primitives, due to which the relationship between the global and local context of the image is lost. Hence, all these problems contribute to inaccurate semantic segmentation of semi-dark imagery. In summary, the existing DL pool of knowledge is limited in the following aspects regarding the semantic segmentation of semi-dark imagery in outdoor settings:
  • The loss of spatial relationships by using per-pixel primitives, resulting in poorly detected object classes over boundary pixels.
  • The limited representational power due to the use of identity shortcuts and the occurrence of an internal covariant shift during the training of DNNs.
  • The constrained multiscale feature extraction caused by biased centric exploitations.
Although the encoder–decoder architectures have brought promising progress, the existing state-of-the-art methods fail to produce exact pixel-wise classification in semi-dark image scenarios. All the above-mentioned problems contribute to an inherent loss of information. On the other hand, the increasing demand for optimal automated solutions points to the critical need for generalized solutions in terms of lightning conditions (dark, semi-dark, and bright). To overcome the effect of all the stated problems regarding semi-dark images, we proposed a combined approach using ML-based bunched pixel primitives with enhanced DeepLabV3+ for semantic segmentation. Alongside this, we also proposed an accurate representative measure for accuracy computation. Overall, the presented research produced the following contributions:
1.
Proposal of a novel version of DeepLabV3+ for semi-dark imagery—Unified DeepLabV3+
An enhanced version of DeepLabV3+ that provides a unified solution for handling the challenges associated with the comprehension of semi-dark imagery is presented.
Geometrical pixel abstraction as the input: To keep the spatial relationship intact, the image is first passed through a machine learning (ML)-based preprocessor to highlight the local pixel structure based on the color and proximity. This generates a high level of pixel abstraction to retain the local spatial information, which in turn helps to map the local context (local pixel structure) to the global context of the image to correctly identify the object class in the final semantic segmentation map.
Encoder enhancement: To increase the representational power of the base DeepLab version, the ResNet encoder was updated. The enhancement was focused on boosting the network’s feature extraction by replacing identity shortcuts with non-linear (ReLU) shortcuts. Moreover, to deal with the network training problem, grouped normalization layers are followed by non-linear shortcuts. The usage of group normalization provides better generalization, solving the problem of the internal covariate shift that occurs during the network-training phase.
Decoder enhancement: To induce neutrality for all the pixels of the receptive field, dilated convolution layers are usually stacked, which reduces the effect of centric exploitations. Dilated convolutions are also termed atrous spatial pyramid pooling in DeepLabV3+. Following the same idea of pyramid pooling by means of hole (zero) insertion in the convolution filter, we present customized dilation rates (which decide the rate/number of zeros) for semi-dark images. The customized dilation rates are carefully chosen based on the rationale of extracting as much information as possible for optimal semantic segmentation in terms of the resulting class-wise accuracy.
2.
Proposal of a novel evaluation metric— S 3 c o r e
The existing evaluation metric relating to the accurate identification of network semantic segmentation does not provide a detailed understanding of the presented visual scene. The state-of-the-art evaluation metric presents incomplete representation since it produces inaccurate pixel class prediction. Thus, we present a balanced evaluation metric that can be used as a representative metric for semantic segmentation, providing a robust evaluation for all scenarios.
The following sections of this article are organized as follows: Section 2 introduces the feature-encoder-based methods, the increased resolution of the feature-encoder-based methods closely related to the presented work, and these methods’ relevance to the applicability for semi-dark imagery. Then, we comprehensively describe our proposed encoder–decoder architecture in Section 3. In Section 4, we present the experimental setup, evaluation metrics, and the discussion of the creation of robust evaluation criteria. Same section presents the results of extensive experimental analysis and comparison with other state-of-the-art methods. Section 5 discusses the major findings and pertinence of the presented research for real-world applications. Finally, Section 6 concludes the research and points out the future directions.

2. Related Work

Semantic segmentation is also known as a dense prediction. Semantic segmentation is considered a dense prediction because all the pixels of the image obtain a certain level that creates a resultant pixel mask that highlights the precise object boundaries. The difference between semantic segmentation and instance segmentation is the addition of object semantics in terms of object labels (car, tree, bus, etc.) and object instance semantics in terms of the occurrence of a certain object (car1, car2, …, bus1, bus2, …, etc.) [8]. Figure 1 shows an image and its subsequent semantic segmentation.
Figure 1 shows that the prediction labels are inferred for each pixel, hence this makes it a dense prediction. For semantic segmentation, different methods are available in the DL domain. We focused on two classes of feature encoders, both of which show outstanding progress with comprehensive results and ease of applicability, as well as increasing the resolution of feature-encoder-based methods.

2.1. Feature-Encoder-Based Methods

The widely used DL networks that implement feature-encoder-based architectures are VGG [9] and ResNet [12]. The base mechanism employs stacked convolution layers, ReLU, and pooling layers to extract features from the image. The currently used networks take inspiration from these base state-of-the-art encoder networks for presenting their variants by performing transfer learning or training their networks with additional architectural changes.
VGG contributes to a network architecture by incrementing the convolutional layers, which increases the depth of a network. VGG can take only fixed-sized images due to the presence of fully connected layers. Moreover, it uses the subtraction of RGB values and suffers from centric exploitations. Since all the pixel values in semi-dark scenarios reach maximum values, the mean calculation ends up with a loss of information by providing the same numeric value for the local pixels. Furthermore, the implemented design rules preserve complexity per layer [16]. In turn, all these constraints result in the poor localization of objects in terms of misclassified boundary pixels. ResNet contributes in terms of depth and identity connections from initial network layers to the later layer to reduce the spatial localization problem of the object. However, the invariance of the spatial transformation still leads to a loss of fine details of the visual scene. ResNet presents some other challenges, including limited representational power due to the usage of identity shortcuts and the occurrence of an internal covariant shift [15,17]. These challenges further increase the possibility of lower network performance for the focused image problem. Moreover, neither mentioned network was tested for semi-dark images exclusively, implying worse semantic segmentation results for such visual scenarios.

2.2. Increased Resolution of Feature-Encoder-Based Methods

To deal with the problems of feature encoder methods, increased resolution of feature-encoder-based methods has been proposed in the literature. The enhancement is suggested to enhance the spatial resolution by incorporating different ML and DL concepts.
First, enhancement is used as a step to deal with the problem associated with the field of view (FoV) to obtain high-resolution maps. To increase the FoV and to deal with the problem of centric exploitations, ‘Atrous convolutions’ are used. An atrous convolution involves the addition of holes/zeros in the convolution filter. Since there are holes in the filter, the focus is diverted from the center pixels to there being equal weightage for all the pixels. The key is to stack atrous convolution layers and not to increase the network’s memory strain. This concept is widely used in the DeepLab network family (further critically analyzed in Section 2.3), where the DeepLab networks use the concept of an encoder–decoder architecture. The encoder takes the input image, generates a high-dimensional feature vector, and aggregates them at multiple levels. The decoder network generates a semantic segmentation mask by upsampling the feature vector generated by the encoder. For the encoder, the DeepLab networks have used different existing networks, including VGG and ResNet. The best performance was obtained with a ResNet-based encoder, followed by a decoder that incorporated atrous convolutions for upsampling [10,13,14]. This resulted in feature maps at higher sampling rates. Another enhancement is the so-called ‘Spatial pyramid Pooling’, where the idea is to pool the network together with atrous convolutions of different rates, such as 6, 12, and 18, since the rate controls the number of holes and the filter rates are chosen in the pattern of a pyramid to extract as much information as possible. However, the DeepLab family also fails to deal with low-contrast or semi-dark images. Finally, the last enhancement is achieved by using ML concepts or graphical probabilistic methods. A lot of methods were presented using super-pixel creation methods [18] for preprocessing or conditional random fields (CRFs) for post-processing [10,18,19,20]. The CRFs are used in a fully connected manner, which again imposes architectural constraints that require using same-sized images. However, they provide better object boundary localization with an additional network overhead. Furthermore, the super-pixel methods are used to pool the boundary information as an extra feature for the network, which increases the network overhead.
Table 1 presents a critical analysis of the closely related research that further supports the rationale and need for a generalized semantic segmentation solution for all scenarios (including semi-dark images).
The critical analysis presented in Table 1 indicates the need for DL mechanisms that hold for semi-dark images as well. The existing mechanisms solely focus on the design of deeper structures, which further increase the processing load, along with other complexities. Some of the solutions point to the usage of pre- and post-processing modules to sharpen the boundary pixels only, which is only one aspect of handling semi-dark images. Finally, only a few studies mentioned the occurrence of semi-dark images in the training and testing protocols. Thus, those methods cannot be generalized without further extended studies.

2.3. Critical Analysis of DeepLab Versions

DeepLab semantic segmentation networks are state-of-the-art models proposed by the Google research group. This model follows the increased-resolution feature-encoding-methods flow. The dense predictions are accomplished using atrous convolutions for the upsampling.
  • DeepLabV1
Chen et al. [10] proposed a semantic segmentation network using VGG-16 as the backbone network, and on top of the VGG feature map, probabilistic graphical methods, particularly conditional random fields (CRFs), are used. However, the network performance is not high enough since some failure modes were reported regarding dynamic light conditions and occlusions. Another study claimed that the proposed methods failed to integrate the features of CNN with CRF, limiting its overall functionality, though they could be efficiently integrated using architectural schemes presented in [19].
2.
DeepLabV2
DeepLab network V2 incorporates spatial pyramid pooling into the previous architecture for better semantic segmentation accuracy. This version also replaces VGG-16 with ResNet [12] to tackle features with reduced spatial resolution caused by upsampling and downsampling; meanwhile, the atrous algorithm was used so that the DCNN works in the fully convolutional mode. The atrous algorithm upsamples the filters in subsequent convolutional layers, which results in feature maps at higher sampling rates by means of the insertion of holes between non-zero taps. The proposed version of the model still fails to capture the details of the object boundaries if they are sensitive to some of the results. DeepLabV2 uses atrous convolution in a post hoc manner; however, it is believed that integrating atrous convolution into the encoder would also affect the overall training by reducing the computational overhead. Moreover, the serial application of atrous convolution, as in [38], is expected to fetch accurate object semantic classes.
3.
DeepLabV3
This version continues to use atrous convolution to provide a better receptive field. The enhancement is in terms of cascading the extra layers to include the global context by pooling the image-level features. The proposed network still fails to generate optimal results; as such, one aspect of necessary further investigation is performance analysis over different dilation rates. The study reported some of the misclassified results of the network where sofa/chair, dining table/chair, and the rear view of the objects were incorrectly classified in dynamic lighting conditions [13].
4.
DeepLabV3+
The latest version of DeepLab ensures the fetching of rich contextual features, along with sharp object boundaries [14]. The network architecture still utilizes the concepts of encoder–decoder methods. The enhancement is in terms of bilinear upsampling of the encoder output, which is from DeepLabV3 in this case. The upsampling is performed using an upsampling factor value of 4 and then concatenating it with the low-level image features from the network base. The network performs better than the earlier versions; however, it still fails for some complex scenarios (congested with small objects of different scales), the rear views of objects, and for dynamic lighting conditions [14]. Moreover, the interpolation method can be further changed to bicubic interpolation to fetch most of the visual image information.

2.4. Conclusive Deductions

Indeed, every new version of DCNN has brought about better results in terms of accuracy. However, the enhancements focus on bringing deeper structures, which is acceptable if the results hold for all scenarios. However, many of the surveyed studies explicitly declared semi-dark images as a failure mode for these works. The improvements were made using unsupervised methods; however, the internal network problems still exist, such as slow training, covariate shifts, and the biased exploitation of the receptive field. These issues can be tackled by using appropriate atrous convolution and using appropriate normalization techniques within the DCNN as a part of a focused solution for semi-dark images. Inspired by the DeepLab architecture’s base rationale of controlling and neutralizing the exploitation throughout the image pixels (receptive field) by using atrous convolutions, Unified DeepLab was formulated. We proposed to extend this concept by presenting customized atrous rates, along with appropriate normalization techniques and non-linear connections to alleviate the effect of a covariate shift. Moreover, it is also recommended to use preprocessing for the creation of super-pixels so that local and global contexts are correlated to create the final segmentation mask.

2.5. Preliminary Hypothesis Validation

Based on the literary analysis, it was seen that most of the existing semantic segmentation solutions reported semi-dark images as the operational failure mode. Moreover, the conclusion of an extensive literature review resulted in hypothesis formulation based on our understanding of the problem at hand. It was hypothesized that lower image visibility increases the possibility of inaccurate semantic segmentation. To confirm this hypothesis, we trained the existing DeepLabV3+ version with the same data that was used for the benchmarking of the proposed solution and undertook related detailed qualitative analysis using DeepLabV3+ implementation in MATLAB. The results for semi-dark images were recorded using the addition and subtraction of offset values. These offset values served the purpose of providing image content visibility by increasing or decreasing the brightness. The examined images were processed to increase or decrease the brightness and then used for network training and testing, where all brightness value networks were independently trained and tested. The final performance results in terms of accuracy were then compiled and presented as graphical visualization. The final analysis results are shown in Figure 2.
Figure 2 shows that all the images with decreased brightness resulted in poor semantic segmentation results using DeepLabV3+. Up to an approximately 50–60% semantic segmentation accuracy was achieved for an offset value of minus 30 (Offset-30). However, the accuracy trend increased for all images with increased brightness. For an offset value of plus 30, the accuracy reached up to 80% for test case 6 (image: 001TP_007470), whose accuracy with a minus 30 offset was 60%.
Since the performance was substantially affected by the visibility of the image content, we further proposed a full-fledged framework to check the visibility of image content and then accordingly process the presented image for semantic segmentation. The framework was composed of three different layers, one of which identified the visibility of the image by using the RPLC (relative perceived luminance classification) algorithm [1], and the images were accordingly selected and propagated ahead for further processing. If RPLC classified any image as a semi-dark image, then it was passed to the proposed Unified DeepLab; otherwise, it was passed to the existing workflow based on DeepLabV3+.

3. Materials and Methods

3.1. Overview of Proposed Architecture

As stated in Section 2.3, the existing DCNNs used for the semantic segmentation provides constrained functionality for handling semi-dark images. These functionality limitations are attributed to several problems in the existing workflows. First, centric exploitations of the receptive field cause not all pixels of the image to contribute equally to the final calculated segmentation map. Second, the residual connections, particularly identity connections, result in an internal covariate shift, which results in a loss of representational power. Third, neighboring pixels have a higher probability of having the same class label due to holding closely related information. However, the classes of the pixels are still calculated independently based on per-pixel primitives. Finally, the backbone network can also increase or decrease the performance in terms of accurate segmentations and increased or decreased network loads. These problems in turn generate poor semantically segmented masks for semi-dark images. To solve these issues, the proposed framework consists of four components: a customized pyramid module for semi-dark images to reduce the effect of centric exploitations. The customized atrous convolution layers are selected in the form of a pyramid structure to extract as much information as possible without increasing the network’s processing load. Non-linear shortcuts are used for the encoder since the backbone network ResNet uses identity shortcuts, resulting in limited representational power. The identity shortcuts are replaced by non-linear ReLU connections and group normalization is used to reduce the effect of internal covariate shifts while training the network. A MobileNet encoder is used as a lightweight model to provide concise depth-wise separable convolutions. Local context segmentation is used to keep track of the local and global context of the image pixels by grouping neighborhood pixels based on certain criteria. Figure 3 illustrates the organization and flow of each component. In the first step, we used local context segmentation as a preprocessing layer for creating super-pixels, where the super-pixeled image used for the experiments was processed using 5000 super-pixels. In the second step, the locally grouped pixels were passed to the proposed parallel DCNN for semantic segmentation. Finally, statistical calculations were performed over the resultant map to create a high-definition accurate semantic segmentation map based on the parallel processing.

Rationale for Proposing an Ensemble Approach

The existing semantic segmentation solutions provide constrained functionality due to several intrinsic problems of convolutional neural networks. Therefore, in this research and based on the literary analysis, we first identified the problems that hinder the fetching of the optimal semantic class of an image pixel. Based on the identified problems, this research proposed an ensemble-based approach as a unified solution to tackle all the identified problems. Although the ensemble approach is believed to produce complexity, the proposed solution produces better semantic segmentation results for semi-dark images by keeping a good balance between complexity and accuracy. The introduction of non-linear shortcuts in the ResNet encoder was chosen to decrease the effect of a covariate shift arising during the training phase. Second, the MobileNet encoder was chosen to fetch smaller details of the visual scene due to the presence of greater network layers, which led to the attainment of richer image information. Finally, a customized decoder was proposed for the creation of image understanding at multiple spatial levels (lower to higher) by using customized atrous rates specifically for semi-dark images.
We further introduce the details of each component of the proposed framework in the following subsections.

3.2. Encoder Enhancement—Non-Linear Shortcuts

In encoder–decoder architectures, encoders play a crucial role in extracting the features from a visual image. One of the encoders that is widely used in the DeepLab architectures is ResNet. ResNet is known for its identity shortcuts in DCCNs. These shortcuts basically pool the information from the initial layers for the later layers to retain better spatial information of the extracted features. However, problems related to representational power persist. To solve this problem, non-linear shortcuts were proposed [15], along with a group normalization layer [17] after the non-linear ReLU connections. Figure 4 shows the existing ResNet connections, along with the updated ResNet layering arch, where the encoder was named RGSNet (ReLU-Group Norm ResNet).
To improve the performance, identity shortcuts were proposed for the DCCNs. Although the identity shortcuts improve the gradient stability, thus resulting in improved performance, the representational power is reduced due to the occurrence of covariate shifts during the network training. Due to this phenomenon, the network learns less during training. To find out the balance between representational power and gradient stability, along with the residual shortcuts, nonlinear group normalization shortcuts were introduced. The group normalization is better for non-linear shortcuts, as it does not require normalization along the batch direction. Thus, it gives an advantage for lightweight memory demands as opposed to batch normalization [39]. Following the architectural updates proposed in [15], a non-linear ReLU activation function plus group normalization is employed for the ResNet-based encoder, ultimately producing RGSNet. RGSNet normalizes the contribution of extra gradient stability in the presence of non-linear connections by using group normalization while involving minimal reengineering effort. The hyperparameter G (number of groups) was set to 16 in our experiments, as in [17]. This connection can be mathematically presented as
y l = h x l + F x l , W l   and   x l   +   1 = f y l
Here, in Equation (1), F represents the residual block and W l represents the learnable weights. x l and x l   +   1 represent the inputs to layers l and l   +   1 , respectively, while h and f represent the identity mapping function of each layer, respectively. This follows the direct path of information propagation throughout the network.

3.3. MobileNetV2 Encoder

As the goal was to extract optimal information from the visual scene with high accuracy while keeping the mathematical operations as low as possible, this same concept was used for the implementation of the residual networks, such as ResNet. However, there exists another architecture that is based on the integration of inverted residual connections and ensures a lightweight nature for limited-memory applications. The problem with residual networks is that they work using the wide → narrow → wide approach, focusing the number of channels. These channels are compressed using a 1 × 1 convolution, followed by a 3 × 3 convolution with fewer parameters, and followed by a 1 × 1 convolution to again increase the number of channels. In contrast, MobileNet (specifically version 2) works using a narrow → wide → narrow approach to limit the number of parameters involved. MobileNetV2 accomplishes this task by widening the network using 1 × 1 a convolution and afterward uses a 3 × 3 depth-wise convolution to reduce the number of parameters involved. This entire scenario ensures accurate information extraction from semi-dark images, as the other components of the parallel DCNN are built using less deep network bases (ResNet18-depth18, no of parameters = 11.7 million). We used MobileNetV2 as the DeepLab encoder due to its lightweight nature with deeper and richer feature extraction; it also provides features to deal with linear bottlenecks (using ReLU6 rather than Simple ReLU) and inverted residual connections with a smaller number of parameters to be handled with only one additional hyperparameter t , which is the expansion rate of the channels and is set to a value of 6 (by default). For the semi-dark images, as the objective was to retrieve as much information as possible, this can be achieved with deeper networks. However, if we increase the number of layers in the network (such as by using ResNet), it increases the overhead, providing overall slow training and testing. For this reason, to benefit from the deep network and lightweight feature of MobileNet, it was employed to increase the detailed information of the visual scene. Eventually, it was found that incorporating MobileNetV2 with a depth of 53 (no. of parameters = 3.5 million) did not increase the memory footprint by much compared to the deeper networks, e.g., ResNet-50 with a depth of 50 (no. of parameters = 25.6 million) and ResNet-101 with a depth of 101 (no. of parameters = 44.6 million).

3.4. Decoder Enhancement—Customized Pyramid Module

The DeepLab network architecture family uses spatial pyramid pooling over the receptive field to mitigate the effect of centric exploitations. The existing pyramid pooling layers are created by inserting holes into the kernel to widen the receptive field; this phenomenon is called atrous convolutions [14]. The idea is to apply the atrous convolutions in parallel with different dilation rates, which controls the size of the filter. For the simple convolutions, a 3-by-3 convolution filter has nine parameters and a resultant receptive field of 3 by 3. However, when using dilated convolutions with a 3-by-3 convolution filter and a dilation rate of 2, the resultant receptive field turns out to be of shape 5 by 5, thus we obtain a larger perspective for the feature being calculated. By choosing the correct dilation rate and pooling a couple of atrous convolution layers, the effect of centric exploitations is mitigated. Finally, each feature map obtained from the atrous convolution is concatenated and propagated ahead in the network for the pixel classification. Figure 5 illustrates the customized decoder module for semi-dark images to optimally manipulate the pixel information.
Using the concept of atrous convolution, the updated kernel after the insertion of holes can be calculated using
k ^ = k + k 1 d 1
Equation (2) represents the calculation of the resultant filter/receptive field. k ^ is the resultant filter, k is the actual filter size, and d is the dilation rate [40]. For the proposed decoder module, we used rate values of 3, 8, 13, 18, and 23 with receptive fields (updated kernels) of 7, 17, 27, 37, and 47, respectively. However, the recent DeepLab architecture uses dilation rates of 6, 12, and 18 with receptive fields (updated kernels) of 13, 25, and 37, respectively. These values are acquired by setting the existing filter size k equal to 3. From these numbers, it can be expected that the DeepLab loses pixel information due to the smaller receptive field, which was further demonstrated in the results given in Section 4.

3.5. Statistical Class-Wise Fusion

The statistical fusion performed on the network’s output is employed in such a way that it retains the final optimal class-wise segmentation map. For any DCNN, the objective function is given as Equation (3).
J θ = log P y | x ; θ = m = 1 M log P y m | x ; θ
where θ is the parameter vector for DCNN. The pixel label distributions are calculated using Equation (4).
P y m | x ; θ exp f m y m | x ; θ
where f m y m | x ; θ represents the output of the DCNN at pixel m . Given this information, our statistical module follows the steps mentioned in Algorithm 1 for the proposed parallel DCNN for semi-dark images.
Algorithm 1: Proposed Parallel DCNN for Semi-Dark Images
Input: Initial DCNN parameters θ   RGSNet-DeepLab (Net1), MobNet-DeepLab (Net2), Custom-DeepLab (Net3), customization parameters b l (including expansion rate, dilation rate, and group normalization number), segmented image I s with M pixels, pixel label set to y (1 → n, no of classes).
Parallel DCNN Processing Steps:
1: For each image pixel m, all three DCNN variants perform f m l ) = f m ( l | x ; θ + b l given y l = 1 n .
2: Every pixel (m → M, where M is the no. of image pixels) of image I s receives a semantic label y l from Net1, Net2, and Net3.
Statistical Class-wise Fusion Step:
3: Class-wise mapping is achieved using
f u n i , m = max f m 1 l , f m 2 l , f m 3 l
where f m 1 l is a class-wise label generated from Net1, f m 2 l is a class-wise label generated from Net2, and f m 3 l is a class-wise label generated from Net3.
Output: Semantic segmentation pixel mask representing n object classes.
In particular, the final semantic segmentation masks resulting from each proposed network component are further analyzed to capture the optimal class-wise object prediction performance. The parallel DCNN component values are identified using the following algorithm (Algorithm 2).
Algorithm 2: Parallel DCNN Class-Wise Fusion
1:Provided class-wise results of RGSNet ( f m 1 l ), CustomDecoder-DeepLabV3+ ( f m 2 l ) , and MobileNet-DeepLabV3+ ( f m 3 l )
2:Compare ( f m 1 l ), ( f m 2 l ) , and ( f m 3 l )
3:For each object class
4:     If ( f m 1 l ) > ( f m 2 l ) && ( f m 1 l ) > ( f m 3 l )
5:           Store f m 1 l in stack
6:   Else If ( f m 2 l ) > ( f m 1 l ) && ( f m 2 l ) > ( f m 3 l )
7:           Store f m 2 l in stack
8:     Else If ( f m 3 l ) > ( f m 1 l ) && ( f m 3 l ) > ( f m 2 l )
9:           Store f m 3 l in stack
10:Create a final semantic segmentation vector by fetching the stack values
The final semantic segmentation map is the result of statistical class-wise fusion based on the maximum achieved value of the network performance.
The proposed system receives the preprocessed image in the form of locally grouped pixels via local context segmentation to keep track of the local pixel information in the form of super-pixels. All the DCNN hyperparameters were set to the default values, except the group normalization parameter and the filter dilation rates. The values of these parameters are the expansion rate t = 6, number of groups G = 16, and the dilation rates = 3, 8, 13, 18, and 23. Parallel DCNN processing steps are performed for the identification of each pixel label. After all the pixels obtain their class label, the results are fused to find the maximum of each class’s accuracy. Finally, the maximum of each class is fetched to generate the final semantic segmentation mask representing the optimal class labels for each pixel of the semi-dark image.

4. Experiments

In this section, we present the details of the dataset, implementation, evaluation criteria, and finally, the experimental analysis results.

4.1. Dataset and Evaluation Metrics

For the training and testing of the proposed Unified DeepLab version, we used the CamVid dataset [25,41]. This database was considered due to its complex nature, as it provides images of road settings where the dynamic lightning conditions are an intrinsic property. The dataset originally had 32 classes; however, for simplicity, we merged the class labels and undertook our analysis based on 11 classes (namely, sky, building, pole, road, pavement, tree, sign symbol, fence, car, pedestrian, and bicyclist). The merging was performed solely for the purpose of inducing simplicity in the overall experiments. However, the merging was performed based on the condition that none of the categories were incorrectly merged. The actual categories, along with the final merged categories, are shown in Table 2.
The classes/categories were merged based on visual relevance and similarity of the base classes, such as the bicyclist merged category being based on the bicyclist and motorcycle/scooter classes since both are two-wheeled objects. Therefore, to reduce the network training burden, similar categories were merged.
The dataset consisted of high-quality and high-resolution images of ego-motive scenes. The captured images were of resolution 960 × 720. Given that the emphasis of the presented research was the optimal semantic segmentation of semi-dark images, we used the concept of related perceived luminance classification presented in [1] to identify only the problem domain images. The resultant dataset contained 549 images. The filtered images were randomly divided into training, testing, and validation sets with proportions of 60:20:20, i.e., 329, 110, and 110 images for training, testing, and validation, respectively. The random selection of images was achieved so that none of the images was selected twice by using MATLAB’s ‘randperm’ method. Moreover, the images were selected by ensuring that the new random values were greater than 3, meaning that even if the random number turned out to be close to the previous number, it was rejected. The rejection of a random number was done to make sure none of the sequenced images were selected. The reason for not selecting the immediate sequenced image was to ensure that no similar image with a minor change was selected, as the data was coming from an image sequence generated from consecutive video frames. Moreover, if the same images were selected, then the network might not produce the desired result and end up with an overfitted model. However, as seen in Figure 7 for all the proposed parallel networks, the training (blue line) and the validation curves (black lines) were close to each other, implying that the network was not overfitted. Figure 1 shows a sample raw image, along with the ground truth mask and the respective class labels. Some of the sample images, particularly locally grouped super-pixeled images, along with the ground truth labels, are presented in Table 3.

4.1.1. Proposal of Novel Evaluation Criteria

For the evaluation and benchmarking of the proposed solution, we report the overall accuracy (OA), intersection over union (IoU), and MeannBFScore.
OA = TP + TN TP + FP + TN + FN
IoU = Area   of   Overlap Area   of   Union
Mean   BFScore = Area   of   Overlap Total   number   of   Pixels
In Equation (5), TP, FP, TN, and FN represent ‘True Positive’, ‘False Positive’, ‘True Negative’, and ‘False Negative’, respectively. These criteria do not suit the scenarios where there is a class imbalance since they produce unreliable results. Equations (6) and (7) are better representatives of the semantic segmentation problem. In Equation (6), the ‘area of overlap’ shows the overlap between the predicted segmentation and the ground truth, and the ‘area of union’ shows the union of the prediction and the ground truth. The values range from 0–100 (no overlap to perfectly overlapped prediction). In Equation (7), the ‘area of overlap’ is the same as in Equation (6), whereas the ‘total number of pixels’ represents the number of pixels present in the image.

Significance of Proposed Evaluation Criteria

The existing semantic segmentation evaluation criteria have both strengths and limitations. Many studies are presented in the literature regarding reasoning and applicability in all scenarios. The semantic segmentation score for problem-centric semantic class prediction was employed due to the questionable performance of a single evaluation metric, i.e., OA [42,43]. The unreliability of results is attributed to imbalanced datasets. However, a literature analysis showed that OA can be used in accordance with the other metrics in scenarios where the class imbalance is mitigated. To the best of our knowledge, there has not been any evaluation metric that takes into account the information of three widely used evaluation criteria to ensure the optimal retrieval of a pixel class. Thus, we present a new metric of evaluation for semantic segmentation, i.e., S 3 c o r e . This novel metric is a weighted combination of the OA, IoU, and Mean BFScore. Since it is a combination of three evaluation criteria, we named it ‘S cube core’ and it can be calculated using Equation (8).
S 3 c o r e = W 1 × O A + W 2 × I o U + W 3 × M e a n   B F S c o r e
where W 1 , W 2 , and W 3 are the weights associated with the metrics. For the weight analysis, we initially hypothesized randomly dividing the weights between the three evaluation criteria but this did not help much. However, then we expanded our experimental analysis by incrementing the weight of only one of the metrics and equally dividing the remaining weight among the other metrics. The experimental analysis of S 3 c o r e is presented in Section 4.4, along with the recommended weights.

Experimental Protocol for Weight Identification of S 3 c o r e

The existing segmentation analysis was solely based on one of the above-mentioned evaluation criteria in Section 4.1.1. However, considering the need for accurate semantic class identification, generalized metric evaluation criteria were required. The overall accuracy (OA) metric failed to provide an accurate representation of the image pixel semantics in the presence of class imbalance. However, for the situations where the class imbalance is mitigated properly, it was found to be a useful analysis criterion. Meanwhile, the IoU and Mean BFScore remained better representations in terms of low false positives and false negatives. Considering these aspects, a novel semantic segmentation evaluation criterion, namely, S 3 c o r e , is presented. S 3 c o r e takes into account all three of these existing evaluation criteria and generates a generalized weighted score. The obtained score value is applicable to all applications, where even small details of the visual scene are crucial for the final decision. The value of S 3 c o r e is calculated using Equation (8).
For the weight optimization and analysis, different manipulation strategies were used for weight identification. A total of one thousand random weight numbers were analyzed for each of the selected strategies. The weights were selected in such a way that the sum of the weight values was always one for all the strategies. First, the strategy calculated the weights using incremental values of weights for the MeanBfScore and dividing the weights equally between the OA and IoU. The second and third strategies calculated the weights using incremental values of the IoU and OA, respectively, and assigned equal weights to the other two criteria in each case. The fourth strategy calculated randomized weights. The algorithms that were used for the identification of weights are presented in Algorithms 3–5. Another weight identification method employed was based on randomization. Here, again, the condition that the sum of the weights, i.e., ‘W’, should not exceed 1 was imposed and 1000 different random numbers were checked for the final optimized weight proposal. The weights in this strategy were identified using steps 4, 5, and 6 stated in Algorithm 6. For the random value generation, the Mersenne Twister algorithm (MT19937) [44] was used, which generates values greater than 0 and less than 1. Again, for each iteration, S 3   c o r e values were computed.
Algorithm 3: Incremental IoU
1:Initialize weights W 1 (weight for OA), W 2 (weight for IoU), W 3 (weight for MeanBFScore), and Overall_Weight W
2:Set W 2 = 0.001
3:Repeat until W2 == 1
4:Set W 1 = ( 1 W 2 ) / 2
5:Set W 3 = W 1
6: W 2 + = 0.001
7: W = W 1 + W 2 + W 3
8:If W > 1
9:  Break
Algorithm 3 presents the flow of the experiment for weight identification after initializing the variables. W 2 , which is the weight for the IoU, was set to 0.001 and was incrementally changed up until reaching a value of 1; in this way, the final set of iterations produced 1000 different weight values. Based on the value of W 2 , the weights W 1 and W 3 were identified for each iteration using steps 4 and 5. For each iteration, the weight sum W was also checked such that the sum of all weights was always one. Moreover, for each of these weight values, the semantic segmentation performance metric (obtained using the DCNNs) was checked to calculate the final S 3 c o r e value using Equation (8). The final semantic segmentation scores are presented in Table 9. A similar weight identification strategy was used for the incremental MeanBfScore and OA; for detailed steps, refer to Algorithms 4 and 5.
Algorithm 4: Incremental MeanBFScore
1:Initialize weights W 1 (weight for OA), W 2 (weight for IoU), W 3 (weight for MeanBFScore), and Overall_Weight W
2:Set W 3   = 0.001
3:Repeat until W3 == 1
4:Set W 1 = ( 1 W 3 ) / 2
5:Set W 2 = W 1
6:    W 3 + = 0.001
7:    W = W 1 + W 2 + W 3
8:If W > 1
9:    Break
Algorithm 5: Incremental OA
1:Initialize weights W 1 (weight for OA), W 2 (weight for IoU), W 3 (weight for MeanBFScore), and Overall_Weight W
2:Set W 1 = 0.001
3:Repeat until W1 == 1
4:Set W 2 = ( 1 W 1 ) / 2
5:Set W 3 = W 2
6:    W 1 + = 0.001
7:    W = W 1 + W 2 + W 3
8:If W > 1
9:    Break
Another weight identification that was employed was based on randomization. Here, again, the condition that the sum of the weights, i.e., W , should not exceed 1 was imposed and 1000 different random numbers were checked for the final optimized weight proposal. The weights in this strategy were identified using steps 4, 5, and 6 stated in Algorithm 6. For the random value generation, the Mersenne Twister algorithm (MT19937) [44] was used, which generates values greater than 0 and less than 1. Again, for each iteration, S 3 core values were computed.
Algorithm 6: Randomized Weights
1:Initialize weights W 1 (weight for OA), W 2 (weight for IoU), W 3 (weight for MeanBFScore), Overall_Weight W and counter C
2:Set C = 1
3:Repeat until C <= 1000
4:   Set W 1 = 0.3 Rand ( )
5:Set W 2 = 0.3 + 0.3 Rand ( )
6:Set W 3 = 1 ( W 2 + W 3 )
7:    C + = 1
8:    W = W 1 + W 2 + W 3
9:If W > 1
10:    Break
The final proposed weights for S 3 core were based on the analysis of four thousand different weight values. The final optimal weights were found via the incremental IoU weighting strategy, where the numerical analysis is presented in Table 9.

4.2. Implementation Details

4.2.1. Experimental Setup

To conduct the experiments, we used a machine with the following specifications: Intel® Core™ i7-10750 H CPU @2.60 GHz with 16.0 GB RAM on a 64-bit operating system powered using NVIDIA GeForce GTX. The code was written in MATLAB version R2021a, specifically using the Deep Network Designer toolbox for network design and conceptualization. To train the system, the parameter values were set as shown in Table 4.

4.2.2. Handling Class Imbalance

Before training the network, we first checked the dataset statistics to identify whether any class imbalance existed. Since class imbalance can be detrimental to the training process due to unbiased behavior toward the dominant classes, it is imperative to identify and mitigate the effect of class imbalance. To determine the class imbalance, we counted the pixels by class labels. For this purpose, we used the countEachLabel method, which takes in a datastore (images) and returns pixel labels and a count for the input datastore; the result was named ‘ f r e q u e n c y ’, which is basically the average of the pixel counts. The pixel label visualization is shown in Figure 6, and it was seen that sky, building, road, and tree were the dominant classes, whereas pole, pavement, sign symbol, fence, car, pedestrian, and bicyclist had smaller numbers of class labels.
To balance the class labels, we used a weighting strategy. The weighting strategy was accomplished by identifying the median of the frequencies presented in Figure 4. The class weights were calculated using Equation (9).
C l a s s   w e i g h t = m e d i a n f r e q u e n c y n . / f r e q u e n c y n ,   n = 1 , 2 ,     ,   n
Here, n represents the number of classes and the class weight was identified using element-wise division, through which we obtained weights for each class. The final weights are given in Table 5.
Finally, the balanced class labels in terms of the weighting strategy were used for the network training.

4.2.3. Training the Proposed Model Components

In this section, we report the training and validation performances that were obtained after setting the training parameters to the values specified in Table 2. For the benchmarking of the proposed Unified DeepLab, we trained ResNet-DeepLab, where different components of Unified DeepLab were trained separately and the accuracy and loss curves of each of the components are presented in Figure 7.
Figure 7a–c shows that the base ResNet-DeepLabV3+ reached the maximum validation accuracy after 16 epochs and completed 650 iterations. Similar results were observed with RGSNET-DeepLab and Customized Decoder-DeepLab provided that both the networks had extra layers and customization operators at the backend. Here, we could conclude that the memory footprint was not increased by enhancing the network architecture. Meanwhile, the peak performance (validation accuracy) in the training cycle of MobileNet-DeepLab was achieved after 21 epochs in 850 iterations, which was acceptable, as MobileNet is three times deeper than ResNet-18 (refer to Section 3.3 for details). Integrating these distinct network components, we further elaborate on the results of the proposed Unified DeepLab in upcoming sections.

Ablation Study

In this section, we report the individual component performance based on the class-wise OA, IoU, and Mean BFScore for the dataset images. The results presented in Table 5 are for the entire dataset used for the analysis. The RGSNet-based encoder successfully fetched the small details of the visual scene, as it achieved the maximum performances for the classes pedestrian, sign symbol, and pavement. MobileNet-ResNet achieved the maximum accuracies for only two classes, namely, sky and pole, which can be further optimized in later research. The Customized Decoder DeepLab variant produced the best performance by achieving the maximum accuracies for six classes, namely, building, road, tree, fence, car, and bicyclist. The promising performance of this customized decoded variant was attributed to the better extraction and generalized manipulation of receptive field pixels. Ultimately, all these DCNN network components created an optimal ‘Unified’ semantic segmentation solution for semi-dark images by correctly handling different aspects of the DCNN mechanics.
In Table 6, bold numeric values show the best class-wise performer for each network component. Moreover, the novel metric values were seen to boost the performance up to a margin of 10% compared with the singular metric Mean BFScore (such as the classes fence and car). Finally, the overall average for each network component was also seen to increase by using the proposed criteria. The ablation study showed that the best performer was the Customized Decoder DeepLab version due to its relevance with the correctly chosen rate values, which effectively increased the FOV for the feature extraction.
To validate the results visually, Figure 8 shows the image areas where the base ResNet-DeepLab misclassified the image pixels, whereas other variants perform better in terms of visual results.
The image portions highlighted with pink boxes over the tested image show that ResNet-DeepLabV3+ misclassified road (class) pixels as pavement, whereas our proposed variants classified the pixels accurately in the corresponding regions. Table 7 shows some more visual results of semantic segmentation over the tested images.

4.3. Comparison with the State-of-the-Art Method

We conducted a comparative analysis of the CamVid dataset on the base ResNet- DeepLabV3+ against the proposed DCNN, i.e., Unified DeepLab. Table 8 shows that Unified DeepLab performed better than ResNet-DeepLabV3+ in terms of greater class-wise accuracy for up to seven classes out of the eleven predefined classes. Even for a sleek object, such as a bicyclist, a greater MeanBfScore was achieved with a margin of 4%. Overall, in all the other classes, the proposed network exceeded the existing network’s performance with a margin of 3%. Finally, the averaged class-wise values showed that the proposed network performed better than the existing network with margins of 3% for the overall accuracy (OA), approximately 1% for the IoU, and 2% for the Mean BFScore.
Table 8 also shows the analysis results of the existing evaluation metrics versus the proposed criteria. Again, a similar performance boost was witnessed for the state-of-the-art semantic segmentation solution and the proposed Unified DeepLab.

4.4. Semantic Segmentation Analysis Based on S 3 c o r e

The segmentation analysis was solely based on each of the above-mentioned evaluation criteria. However, considering the need for accurate semantic class identification, generalized metric evaluation criteria are required. The overall accuracy (OA) metric fails to accurately represent the image pixel semantics in the presence of class imbalance, though for the situations where the class imbalance is mitigated properly, it can be a useful analysis criterion. However, the IoU and Mean BFScore remain the better representations in terms of low false positives and false negatives. Considering these aspects, we present a novel semantic segmentation evaluation criterion, namely, S 3 c o r e . S 3 c o r e takes into account all three of these existing evaluation criteria and generates a generalized weighted score. The obtained score value is applicable to all situations where even small details of the visual scene are crucial for the final decision. The value of S 3 c o r e is calculated using Equation (8). For the weight optimization and analysis, we conducted experiments using different manipulation strategies for weight identification. A total of one thousand weight numbers were analyzed for each selected strategy. The weights were selected in such a way that the sum of the weight values was always one for each strategy. The first strategy calculated weights using incremental values of weights for MeanBfScore and divided the remaining weight equally between the OA and IoU. The second and third strategies calculated weights using incremental values of the IoU and OA, respectively, and assigned equal weights for the other two criteria in each case. The fourth strategy calculated randomized weights. The final weight analysis is presented in Table 9.
From Table 9, it can be seen that the existing accuracies could be integrated and the accuracy margins can likely be boosted by up to 10% if the correct weighting scheme is employed. The maximum S 3 c o r e value was attained for the incremental weighting OA scheme, i.e., 85% for Unified DeepLab and 82% for ResNet-DeepLabV3+. However, based on the literary survey conducted, we recommend the usage of an incremental IoU weighting scheme as the IoU is a better representative of the semantic segmentation pixel-wise classification, but it still incorporates the information of the OA. Hence, we propose an incremental IoU weighting scheme, which led to boosted margins of up to 10% with the weight values W 1 = 0.4995, W 2 = 0.001, and W 3 = 0.4995 for the OA, IoU, and MeanBFscore, respectively.

5. Discussion

In this paper, we present a unified approach to solving different problematic aspects of DCNN fundamentals. Since semantic segmentation applications are not confined to only the autonomous vehicle industry, our proposed solution holds for all the related applications. The comprehensive analysis was based on high-resolution images of dynamic visual scenes that incorporated different complexities (including different-sized objects and relatively low-luminance images). Incorporating all these complexities, the proposed DCNN still managed to produce a high performance margin. The optimal performance requisition pointed to the applicability of this solution to other problem domain areas, as well where image complexities are supposed to be an intrinsic feature.

Novelty and Contributions

In our considered opinion, the novelty and contribution of this research are of moderate nature, as highlighted below.
  • Novel ensemble model approach
The focused solution base incorporated different network architectural changes that were otherwise used for intelligent visual decisions (simple object detection) but not specifically semantic segmentation problems. The proposed ensemble was based on a modified model that was backed by some existing architectures. To the best of our knowledge, this model does not exist in the current literature. Moreover, the proposed solution is expected to benefit from various automated domain applications that require optimal information extraction for all scenarios (irrespective of object scale, color, and luminance scenarios).
2.
Novel semantic segmentation evaluation criterion
A semantic segmentation criterion was proposed as a representative criterion for all the scenarios where data imbalance strategies are employed. Moreover, the higher values of S 3 c o r e implied a good ranking of semantic segmentation performance against the ground truth labels and perceptual analysis. This perceptual analysis can be further comprehensively investigated by a subjective user study. The criterion is generic in nature and can be used by other researchers in the domain in either its current form or as a basis to create an appropriate extension to suit the needs of semantic segmentation applications.

6. Conclusions

We provided a unified approach based on statistical class-wise fusion, bringing together different problem solutions to facilitate optimal pixel class identification in semi-dark images. We improved the existing solution with minimal DCNN architectural changes. The four key changes made involved the usage of the following:
(1)
Preprocessed super-pixeled images, which are locally grouped pixel images that keep track of the local contexts of images;
(2)
Non-linear shortcuts followed by group normalization layers in a residual network encoder (ResNet-18) to increase the feature representational power and normalization to ensure improved training accuracy;
(3)
A MobileNet-based encoder to increase the depth of the network for fetching fine-grained details using a deeper but less complex network that involves 8.2 million fewer parameters compared with the base ResNet-18 encoder;
(4)
A customized pyramid decoder (customized dilated convolution layers) to provide focused control of the receptive field and mitigate the effect of centric exploitations in semi-dark images.
All these changes in an encoder–decoder-based DCNN provided a unified perspective that allowed for a generalized proposed solution, i.e., Unified DeepLab. The proposed Unified DeepLab was evaluated on the CamVid dataset and was benchmarked against the most recent and effective state-of-the-art counterpart, namely, DeepLabV3+. Our experimental analysis showed that the proposed DCNN outperformed DeepLabV3+ by a margin of 3% for the overall accuracy and 2% for the Mean BFScore. Finally, we also presented a novel evaluation criterion for a semantic segmentation score called ‘ S 3 c o r e ’ that was based on the weighted combination of the overall accuracy, IoU, and Mean BFScore, which further boosted the class-wise pixel performance by 10%.

Author Contributions

Conceptualization, M.M.M. and M.A.H.; methodology, M.M.M.; software, S.S.R.; validation, M.M.M., M.A.H. and A.Z.J.; formal analysis, A.Z.J.; investigation, M.M.M. and A.Z.J.; resources, S.S.R.; data curation, S.S.R.; writing—original draft preparation, M.M.M.; writing—review and editing, A.Z.J.; visualization, M.M.M.; supervision, M.A.H.; project administration, K.R.; funding acquisition, M.A.H. and K.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research study was part of a funded project under a matching grant scheme supported by Iqra University, Pakistan, and Universiti Teknologi PETRONAS (UTP), Malaysia (grant number: 015MEO-227).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The following Camvid dataset was used: http://web4.cs.ucl.ac.uk/staff/g.brostow/MotionSegRecData/ (accessed on 29 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Memon, M.M.; Hashmani, M.A.; Junejo, A.Z.; Rizvi, S.S.; Arain, A. A Novel Luminance-Based Algorithm for Classification of Semi-Dark Images. Appl. Sci. 2021, 11, 8694. [Google Scholar] [CrossRef]
  2. Chen, C.; Chen, Q.; Xu, J.; Koltun, V. Learning to see in the dark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3291–3300. [Google Scholar]
  3. Ouyang, S.; Li, Y. Combining deep semantic segmentation network and graph convolutional neural network for semantic segmentation of remote sensing imagery. Remote Sens. 2021, 13, 119. [Google Scholar] [CrossRef]
  4. Yu, J.; Zeng, P.; Yu, Y.; Yu, H.; Huang, L.; Zhou, D. A Combined Convolutional Neural Network for Urban Land-Use Classification with GIS Data. Remote Sens. 2022, 14, 1128. [Google Scholar] [CrossRef]
  5. Senthilnathan, R. Deep Learning in Vision-Based Automated Inspection: Current State and Future Prospects. In Machine Learning in Industry; Springer: Berlin/Heidelberg, Germany, 2022; pp. 159–175. [Google Scholar]
  6. Chen, L.-C.; Yang, Y.; Wang, J.; Xu, W.; Yuille, A.L. Attention to scale: Scale-aware semantic image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3640–3649. [Google Scholar]
  7. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  8. Lateef, F.; Ruichek, Y. Survey on semantic segmentation using deep learning techniques. Neurocomputing 2019, 338, 321–348. [Google Scholar] [CrossRef]
  9. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  10. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  12. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  13. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  14. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  15. Zhang, C.; Rameau, F.; Lee, S.; Kim, J.; Benz, P.; Argaw, D.M.; Bazin, J.-C.; Kweon, I.S. Revisiting residual networks with nonlinear shortcuts. In Proceedings of the BMVC, Cardiff, UK, 9–12 September 2019. [Google Scholar]
  16. McAllister, R.; Gal, Y.; Kendall, A.; Van Der Wilk, M.; Shah, A.; Cipolla, R.; Weller, A. Concrete Problems for Autonomous Vehicle Safety: Advantages of Bayesian Deep Learning. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence AI and Autonomy Track, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  17. Zhou, X.-Y.; Yang, G.-Z.; Letters, A. Normalization in training U-Net for 2-D biomedical semantic segmentation. IEEE Robot. Autom. Lett. 2019, 4, 1792–1799. [Google Scholar] [CrossRef] [Green Version]
  18. Zhao, W.; Fu, Y.; Wei, X.; Wang, H. An improved image semantic segmentation method based on superpixels and conditional random fields. Appl. Sci. 2018, 8, 837. [Google Scholar] [CrossRef] [Green Version]
  19. Zheng, S.; Jayasumana, S.; Romera-Paredes, B.; Vineet, V.; Su, Z.; Du, D.; Huang, C.; Torr, P.H. Conditional random fields as recurrent neural networks. In Proceedings of the IEEE International Conference on Computer Vision, Washington, DC, USA, 7–13 December 2015; pp. 1529–1537. [Google Scholar]
  20. Plath, N.; Toussaint, M.; Nakajima, S. Multi-class image segmentation using conditional random fields and global classification. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; pp. 817–824. [Google Scholar]
  21. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet Classification with Deep Convolutional Neural Networks; Association for Computing Machinery: New York, NY, USA, 2012; Volume 25. [Google Scholar]
  22. Everingham, M.; Eslami, S.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  23. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  24. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  25. Brostow, G.J.; Fauqueur, J.; Cipolla, R.J.P.R.L. Semantic object classes in video: A high-definition ground truth database. Pattern Recognit. Lett. 2009, 30, 88–97. [Google Scholar] [CrossRef]
  26. Song, S.; Lichtenberg, S.P.; Xiao, J. Sun rgb-d: A rgb-d scene understanding benchmark suite. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 567–576. [Google Scholar]
  27. Cogswell, M.; Lin, X.; Purushwalkam, S.; Batra, D. Combining the best of graphical models and convnets for semantic segmentation. arXiv 2014, arXiv:1412.4313. [Google Scholar]
  28. Farabet, C.; Couprie, C.; Najman, L.; LeCun, Y. Learning hierarchical features for scene labeling. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1915–1929. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Liu, C.; Yuen, J.; Torralba, A.; Sivic, J.; Freeman, W.T. Sift flow: Dense correspondence across different scenes. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; pp. 28–42. [Google Scholar]
  30. Tighe, J.; Lazebnik, S. Superparsing: Scalable nonparametric image parsing with superpixels. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; pp. 352–365. [Google Scholar]
  31. Gould, S.; Fulton, R.; Koller, D. Decomposing a scene into geometric and semantically consistent regions. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 27 September–4 October 2009; pp. 1–8. [Google Scholar]
  32. Papandreou, G.; Chen, L.-C.; Murphy, K.P.; Yuille, A.L. Weakly-and semi-supervised learning of a deep convolutional network for semantic image segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1742–1750. [Google Scholar]
  33. Saito, S.; Kerola, T.; Tsutsui, S. Superpixel Clustering with Deep Features for Unsupervised Road Segmentation. 2017. Available online: https://www.arxiv-vanity.com/papers/1711.05998/ (accessed on 29 May 2022).
  34. He, Y.; Chiu, W.-C.; Keuper, M.; Fritz, M. Std2p: Rgbd semantic segmentation using spatio-temporal data-driven pooling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4837–4846. [Google Scholar]
  35. Zhou, L.; Fu, K.; Liu, Z.; Zhang, F.; Yin, Z.; Zheng, J.J.N. Superpixel based continuous conditional random field neural network for semantic segmentation. Neurocomputing 2019, 340, 196–210. [Google Scholar] [CrossRef]
  36. Kae, A.; Sohn, K.; Lee, H.; Learned-Miller, E. Augmenting CRFs with Boltzmann machine shape priors for image labeling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2019–2026. [Google Scholar]
  37. Smith, B.M.; Zhang, L.; Brandt, J.; Lin, Z.; Yang, J. Exemplar-based face parsing. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 3484–3491. [Google Scholar]
  38. Fisher Yu, V.K. Multi-Scale Context Aggregation by Dilated Convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  39. Wu, Y.; He, K. Group normalization. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  40. Dumoulin, V.; Visin, F. A guide to convolution arithmetic for deep learning. arXiv 2016, arXiv:1603.07285. [Google Scholar]
  41. Brostow, G.J.; Shotton, J.; Fauqueur, J.; Cipolla, R. Segmentation and recognition using structure from motion point clouds. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; pp. 44–57. [Google Scholar]
  42. Csurka, G.; Larlus, D.; Perronnin, F.; Meylan, F.J.I.P. What is a good evaluation measure for semantic segmentation? In Proceedings of the British Machine Vision Conference; BMVA Press: London, UK, 2013; Volume 26. [Google Scholar]
  43. Fernandez-Moral, E.; Martins, R.; Wolf, D.; Rives, P. A new metric for evaluating semantic segmentation: Leveraging global and contour accuracy. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (iv), Suzhou, China, 26–30 June 2018; pp. 1051–1056. [Google Scholar]
  44. Saito, M.; Matsumoto, M. SIMD-oriented fast Mersenne Twister: A 128-bit pseudorandom number generator. In Monte Carlo and Quasi-Monte Carlo Methods 2006; Springer: Berlin/Heidelberg, Germany, 2008; pp. 607–622. [Google Scholar]
Figure 1. Semantic Segmentation Demonstration (a) Actual Image (b) Semantic Segmentation Ground-truth.
Figure 1. Semantic Segmentation Demonstration (a) Actual Image (b) Semantic Segmentation Ground-truth.
Sensors 22 05312 g001
Figure 2. Hypothesis Validation.
Figure 2. Hypothesis Validation.
Sensors 22 05312 g002
Figure 3. Proposed Parallel DCNN for Semi-dark Images.
Figure 3. Proposed Parallel DCNN for Semi-dark Images.
Sensors 22 05312 g003
Figure 4. (a) ResNet with Identity shortcuts. (b) RGSNet (ReLU-Group Norm ResNet) with non-linear connections.
Figure 4. (a) ResNet with Identity shortcuts. (b) RGSNet (ReLU-Group Norm ResNet) with non-linear connections.
Sensors 22 05312 g004
Figure 5. Customized Decoder Module for Semi-dark Images.
Figure 5. Customized Decoder Module for Semi-dark Images.
Sensors 22 05312 g005
Figure 6. Class Distribution in the CamVid Dataset (Semi-dark Images).
Figure 6. Class Distribution in the CamVid Dataset (Semi-dark Images).
Sensors 22 05312 g006
Figure 7. DCNN Accuracy and Loss Graphs. (a) ResNet-DeepLabV3+, (b) RGSNet-DeepLabV3+, (c) Customized Decoder-DeepLabV3+, (d) MobileNet-DeepLabV3+.
Figure 7. DCNN Accuracy and Loss Graphs. (a) ResNet-DeepLabV3+, (b) RGSNet-DeepLabV3+, (c) Customized Decoder-DeepLabV3+, (d) MobileNet-DeepLabV3+.
Sensors 22 05312 g007aSensors 22 05312 g007bSensors 22 05312 g007c
Figure 8. Visual Analysis of a Tested Image. (a) ResNet-DeepLabV3+, (b) RGSNet-DeepLab, (c) MobileNet-DeepLab, (d) Customized Decoder-DeepLab.
Figure 8. Visual Analysis of a Tested Image. (a) ResNet-DeepLabV3+, (b) RGSNet-DeepLab, (c) MobileNet-DeepLab, (d) Customized Decoder-DeepLab.
Sensors 22 05312 g008aSensors 22 05312 g008b
Table 1. Comparative Analysis of Semi-dark Image-Centric Semantic Segmentation.
Table 1. Comparative Analysis of Semi-dark Image-Centric Semantic Segmentation.
Deep Learning ArchitectureOriginal
Architecture
Testing
Benchmark
Semi-Dark Image HandlingObservations
VGG [9]AlexNet [21]ILSVRC-2012
Pascal VOC [22]
MS Coco [23]
🗙
  • Poor localization due to the final DCNN layer [10]
  • Does not use any unsupervised pre-training scheme to aid performance [21]
ResNet [12]VGG [9]ILSVRC-2012🗙
  • Invariance to spatial transformations results in the loss of fine details [10]
  • Limited representational power due to the usage of identity shortcuts [15,17]
Pascal VOC [22]
MS Coco [23]
DeepLab [10]VGG [9]Pascal VOC [22]
Cityscapes [24]
🗙
  • Fails to extract fine-grained details in semi-dark scenarios [10]
SegNet [11]VGG [9]Cam Vid Dataset [25]
Sun RGB-D [26]
  • Fails to capture pixel information and results in coarse segmentation results.
  • Peak performance is achieved in the presence of handcrafted features [27]
DeepLabV2 [10]VGG [9]Pascal VOC [22]
MS Coco [23]
Cityscapes [24]
🗙
  • Better accuracy than the previous version but still cannot handle semi-dark images
  • No preprocessing module is used for highlighting boundaries
DeepLabV3 [13]DeepLabV2 [10]Pascal VOC [22]
Cityscapes [24]
🗙
  • Several different rates need to be further investigated, as the system still fails to handle complex images
DeepLabV3+ [14]DeepLabV3 [13], ResNet [12]Pascal VOC [22]
Cityscapes [24]
🗙
  • Attains better accuracy than the previous version; however, it still fails to work for darker regions, complex images (congested with small objects of different scales), and the rear views of objects [14]
Attention Model [6]Super-pixel creation (SLIC) DeepLab [10]Pascal VOC [22]
MS Coco [23]
🗙
  • Trained and tested to generate a person’s body parts, and fails to segment in scenarios of difficult human poses [6]
  • Uses a preprocessing module for highlighting boundaries
Super-Pixel and CRF-Based FCN [18] Cityscapes [24]🗙
  • Uses CRFs as post-processing to aid the performance accuracy [18]
  • Presents promising research results that can be extended to semi-dark images
Super-Pixel-Based Hierarchical Network [28] Super-pixel creation (tree/clustering)
ConvNet (2-layer MLP) CRF
SIFT Flow dataset [29]
Barcelona dataset [30]
Stanford background [31]
🗙
  • Focuses on the extraction of small details
  • Pools the network with super-pixel boundary information
Semi-Supervised Convolution Neural Network [32]DeepLab-CRF [10]Pascal VOC [22]
MS Coco [23]
🗙
  • Incorporates super-pixel information to provide better results
  • No mention of performance for semi-dark images
Higher-Order CRF in DNNFCN VGG [9]Pascal VOC [22]🗙
  • Fast RCNN for object detection
  • Parallel implementation of an object-detection CNN with CRF and a graph-based method for super-pixel extraction can increase the complexity
Super-Pixel-Based DCNN for Road Segmentation [33]VGG [9], ResNet [12]Cityscapes [24]🗙
  • Trained and tested in controlled road environments
  • No mention of semi-dark images
  • Uses both ResNet and VGG, which increases the memory consumption
Super-Pixel and Statistically Learned DCNNSimple fully convolution neural networkPascal VOC [22]
Sun RGB-D [26]
  • Super-pixels are pooled in the network as extra information
Spatio-Temporal Data-Driven Pooling DCNN [34]CNN (No specifics written)Sun RGB-D [26]
  • Pro: spatio-temporal data-driven pooling can receive multiple images and their correspondence as input
  • Uses prior super-pixels
  • Unknown network base raises questions about the generic applicability
Super-Pixel- and CRF-Based DCNN [35]SEGNET-VGG16 [11], DeepLabV2-ResNet [10]LFW-PL dataset [36],
HELEN dataset [37] (facial images)
Pascal VOC [22]
🗙
  • Complex architecture with different base components
🗙→Absence of Semi-dark image handling, → Presence of Semi-dark image Handling.
Table 2. Actual CamVid Classes vs. Merged Classes.
Table 2. Actual CamVid Classes vs. Merged Classes.
S#Merged ClassActual CamVid Dataset Class
1SkySky
2BuildingBridge
Building
Wall
Tunnel
Archway
3PoleColumn_Pole
TrafficCone
4RoadRoad
LaneMkgsDriv
LaneMkgsNonDriv
5PavementSidewalk
ParkingBlock
RoadShoulder
6TreeTree
VegetationMisc
7Sign SymbolSignSymbol
Misc_Text
TrafficLight
8FenceFence
9CarCar
SUVPickupTruck
Truck_Bus
Train
OtherMoving
10PedestrianPedestrian
Child
CartLuggagePram
Animal
11BicyclistBicyclist
MotorcycleScooter
Table 3. Super-pixeled Images and ground truth masks in the CamVid Dataset.
Table 3. Super-pixeled Images and ground truth masks in the CamVid Dataset.
Sample 1Sample 2Sample 3Sample 4Sample 5
Images Sensors 22 05312 i001 Sensors 22 05312 i002 Sensors 22 05312 i003 Sensors 22 05312 i004 Sensors 22 05312 i005
Ground truth Sensors 22 05312 i006 Sensors 22 05312 i007 Sensors 22 05312 i008 Sensors 22 05312 i009 Sensors 22 05312 i010
Table 4. Network Training Parameters.
Table 4. Network Training Parameters.
ParameterSolverInitial Learning RateValidation FrequencyMaximum EpochsMini-Batch SizeL2
Regularization
Gradient Threshold MethodValidation PatienceShuffle
ValueSgdm (Stochastic Gradient Descent with Momentum)0.01502580.0001L2 Norm5Every-epoch
Table 5. Class Weights for Handling Biased Network Training.
Table 5. Class Weights for Handling Biased Network Training.
ClassWeight Value
Sky0.2933
Building0.1674
Pole4.2693
Road0.1480
Pavement0.5942
Tree0.3766
Sign Symbol4.0753
Fence1.5294
Car1.0000
Pedestrian5.6283
Bicyclist4.2795
Table 6. Class-wise Accuracy Analysis.
Table 6. Class-wise Accuracy Analysis.
NetworkClassSkyBuildingPoleRoadPavementTreeSign SymbolFenceCarPedestrianBicyclistAverage
Metrics
RGSNet-DeepLabOA 0.94080.74910.57610.92410.90280.85090.81520.71190.86120.85060.80010.8166
IoU0.90460.72230.16570.91610.69830.75020.22200.51800.70090.29140.51840.5825
Mean BFScore0.87230.54570.46130.76160.68790.63430.33080.43370.60200.44520.46750.5675
S 3 c o r e 0.90660.64750.51830.84290.79520.74260.57260.572780.73160.64750.63360.6919
MobileNet-DeepLabOA 0.96400.76210.67520.95160.89010.85230.72580.73370.88190.77030.78520.8175
IoU0.91280.73430.16050.93780.74050.74740.30640.56380.73150.35510.59970.6173
Mean BFScore0.90150.53740.46090.83050.75100.63750.42910.50460.64890.51140.59430.6188
S 3 c o r e 0.93270.64980.56760.89110.82050.74490.57720.61910.76530.64050.68960.7180
Customized Decoder-DeepLab OA0.94980.84380.64270.95190.86950.88990.68280.76040.89420.77390.81270.8247
IoU0.91240.80470.21150.93830.74840.77600.36630.60910.75580.41980.59220.6486
Mean BFScore0.89130.64050.55790.82950.77060.68580.49140.52720.68340.60070.61790.6633
S 3 core 0.92050.74220.59990.89070.81990.78780.58680.64370.78870.68700.71510.7438
Table 7. Visual Semantic Segmentation Results.
Table 7. Visual Semantic Segmentation Results.
Test CaseNetwork
ResNet-DeepLabV3+RGSNet-DeepLabMobileNet-DeepLabCustomized Decoder-DeepLab
1 Sensors 22 05312 i011 Sensors 22 05312 i012 Sensors 22 05312 i013 Sensors 22 05312 i014
2 Sensors 22 05312 i015 Sensors 22 05312 i016 Sensors 22 05312 i017 Sensors 22 05312 i018
3 Sensors 22 05312 i019 Sensors 22 05312 i020 Sensors 22 05312 i021 Sensors 22 05312 i022
4 Sensors 22 05312 i023 Sensors 22 05312 i024 Sensors 22 05312 i025 Sensors 22 05312 i026
Table 8. Class-wise Comparative Analysis with the State-of-the-Art DCNN.
Table 8. Class-wise Comparative Analysis with the State-of-the-Art DCNN.
NetworkClassSkyBuildingPoleRoadPavementTreeSign SymbolFenceCarPedestrianBicyclistAverage
Metrics
ResNet-DeepLabV3+OA0.94860.84180.65570.95360.87790.85330.67820.77240.88120.80640.82610.8268
IoU0.91000.79760.19380.94060.75360.76250.3990.57360.74020.39700.59370.6420
Mean BFScore0.89420.62530.52990.83480.77790.66740.51910.49780.65670.56210.57780.6494
S 3 core 0.92140.73360.59230.89420.82780.76030.59840.63500.76890.68390.70180.7380
Unified DeepLab OA 0.96400.84380.67520.95190.90280.88990.81520.76040.89420.85060.81270.8510
IoU0.91280.80470.21150.93830.74840.77600.36630.60910.75580.41980.59970.6493
Mean BFScore0.90150.64050.55790.83050.77060.68580.49140.52720.68340.60070.61790.6643
S 3 core 0.93270.74220.61610.89120.83660.78780.65300.64370.78870.72530.71510.7575
Table 9. Comparative Analysis with S 3 c o r e (different weighting schemes).
Table 9. Comparative Analysis with S 3 c o r e (different weighting schemes).
Network S 3 c o r e   Weighting   Scheme Min   S 3 c o r e   ( with   Weights ) Max   S 3 c o r e   ( with   Weights )
ResNet-DeepLabV3+Incremental MeanBFScore64
W 1 = 0, W 2 = 0, W 3 = 1
72.99
W 1 = 0.4995, W 2 = 0.4995, W 3 = 0.001
Unified DeepLabIncremental MeanBFScore66
W 1 = 0, W 2 = 0, W 3 = 1
74.49
W 1 = 0.4995, W 2 = 0.4995, W 3 = 0.001
ResNet-DeepLabV3+Incremental IoU64
W 1 = 0, W 2 = 1, W 3 = 0
72.99
W1 = 0.4995, W 2 = 0.001, W 3 = 0.4995
Unified DeepLabIncremental IoU64
W 1 = 0, W 2 = 1, W 3 = 0
75.48
W 1 = 0.4995, W 2 = 0.001, W 3 = 0.4995
ResNet-DeepLabV3+Incremental OA64.02
W 1 = 0.0010, W 2 = 0.4995, W 3 = 0.4995
82
W 1 = 1, W 2 = 0, W 3 = 0
Unified DeepLabIncremental OA65.02
W 1 = 0.0010, W 2 = 0.4995, W 3 = 0.4995
85
W 1 = 1, W 2 = 0, W 3 = 0
ResNet-DeepLabV3+Randomized64.00
W 1 = 0.00008, W 2 = 0.48808, W 3 = 0.51184
69.38
W 1 = 0.29931, W 2 = 0.51779, W 3 = 0.18290
Unified DeepLabRandomized64.97
W 1 = 0.00863, W 2 = 0.59849, W 3 = 0.39288
71.02
W 1 = 0.2993822, W 2 = 0.33354, W 3 = 0.36708
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Memon, M.M.; Hashmani, M.A.; Junejo, A.Z.; Rizvi, S.S.; Raza, K. Unified DeepLabV3+ for Semi-Dark Image Semantic Segmentation. Sensors 2022, 22, 5312. https://doi.org/10.3390/s22145312

AMA Style

Memon MM, Hashmani MA, Junejo AZ, Rizvi SS, Raza K. Unified DeepLabV3+ for Semi-Dark Image Semantic Segmentation. Sensors. 2022; 22(14):5312. https://doi.org/10.3390/s22145312

Chicago/Turabian Style

Memon, Mehak Maqbool, Manzoor Ahmed Hashmani, Aisha Zahid Junejo, Syed Sajjad Rizvi, and Kamran Raza. 2022. "Unified DeepLabV3+ for Semi-Dark Image Semantic Segmentation" Sensors 22, no. 14: 5312. https://doi.org/10.3390/s22145312

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop