Next Article in Journal
Effect of Skin Pigmentation and Finger Choice on Accuracy of Oxygen Saturation Measurement in an IoT-Based Pulse Oximeter
Next Article in Special Issue
Design and Implementation of Dongba Character Font Style Transfer Model Based on AFGAN
Previous Article in Journal
Assessment of Panorama Photogrammetry as a Tool for Long-Range Deformation Monitoring
Previous Article in Special Issue
Image Filtering to Improve Maize Tassel Detection Accuracy Using Machine Learning Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Low-Light Sparse Polarization Demosaicing Network (LLSPD-Net): Polarization Image Demosaicing Based on Stokes Vector Completion in Low-Light Environment

1
Electronics and Information Engineering Institute, Changchun University of Science and Technology, Changchun 130022, China
2
Space Opto-Electronics Technology Institute, Changchun University of Science and Technology, Changchun 130022, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(11), 3299; https://doi.org/10.3390/s24113299
Submission received: 1 March 2024 / Revised: 25 March 2024 / Accepted: 26 March 2024 / Published: 22 May 2024
(This article belongs to the Special Issue AI-Driven Sensing for Image Processing and Recognition)

Abstract

:
Polarization imaging has achieved a wide range of applications in military and civilian fields such as camouflage detection and autonomous driving. However, when the imaging environment involves a low-light condition, the number of photons is low and the photon transmittance of the conventional Division-of-Focal-Plane (DoFP) structure is small. Therefore, the traditional demosaicing methods are often used to deal with the serious noise and distortion generated by polarization demosaicing in low-light environment. Based on the aforementioned issues, this paper proposes a model called Low-Light Sparse Polarization Demosaicing Network (LLSPD-Net) for simulating a sparse polarization sensor acquisition of polarization images in low-light environments. The model consists of two parts: an intensity image enhancement network and a Stokes vector complementation network. In this work, the intensity image enhancement network is used to enhance low-light images and obtain high-quality RGB images, while the Stokes vector is used to complement the network. We discard the traditional idea of polarization intensity image interpolation and instead design a polarization demosaicing method with Stokes vector complementation. By using the enhanced intensity image as a guide, the completion of the Stokes vector is achieved. In addition, to train our network, we collected a dataset of paired color polarization images that includes both low-light and regular-light conditions. A comparison with state-of-the-art methods on both self-constructed and publicly available datasets reveals that our model outperforms traditional low-light image enhancement demosaicing methods in both qualitative and quantitative experiments.

1. Introduction

Polarization is an outward manifestation of the transverse wave nature of light and carries information about the structure, material, and reflection angle of the target surface, which can complement the texture information that is passively obtained from the scene. Many polarization imaging techniques have been devised and polarization information is used in a wide range of military and civilian applications, such as in photometrically challenging object attitude estimation [1], the segmentation identification of objects with special materials (metallic, highly exposed or transparent) [2], reflection removal [3], polarization information color constancy [4], and material classification [5].
Traditional DoFP polarimetric imaging devices possess significant advantages in real-time imaging, and thus are widely employed at present. The structure of the polarizer and color filter distribution form of its sensor is shown in Figure 1a, and the existing color polarization imaging instruments are shown in Figure 1b. Although the real-time performance of this structure can be guaranteed, the Bayer structure + polarizing filter structure degrades the quality of the RGB image, especially in low-light environment with less photon transmission. As a result, polarized imaging produces severe noise and artifacts, in addition to a degradation in the quality of the RGB image. For advanced applications of polarized images (e.g., semantic segmentation and target detection), the effect of noise makes the model complex and difficult to converge. It is important to emphasize here that the source of the noise is mostly Gaussian noise, which is further amplified in the subsequent processing and, hence, is reflected in the DoLP and AoP images [6]. Although increasing the ISO of the camera can make the image brighter, the ensuing noise and chromatic aberration are unavoidable, and such noise and distortion have a more serious impact on the subsequent resolution of polarized images. Additionally, increasing the exposure time will inevitably reduce the frame rate, which can also lead to blurring and artifacts [7].
In recent years, there have been many studies related to low-light image enhancement; however, there are currently no studies that can guarantee low-light image enhancement while realizing a distortion-free resolution of polarization information for special polarization sensor structures. The most straightforward approach to enhance low-light images is to linearly amplify each pixel. A typical example is shown in Figure 2, where it can be seen that although the method results in an increase in S0 brightness, at the same time, the noise is amplified. Moreover, due to the amplification of the intensity of image noise, the polarization image (DoLP, AoP) noise, which is then solved by the nonlinear operation, is even more amplified. Figure 3 shows an example of the effect of noise; in the first row, we simulate a non-polarized environment, i.e., the pixel value for acquiring the DoLP is 0. In the second line, we simulate the introduction of a small amount of Gaussian noise ( σ = 0.04 ) into the camera component, and it can be seen that the DoLP at this point has a large value of DoLP in the regions of the image with low intensity. This shows that a weak noise can present a large number of errors in DoLP images. Therefore, it is essential for our method to consider the effect of noise.
For the enhancement of low-light images, there are currently two main types of methods, i.e., traditional methods, including non-local block matching methods [8] and hypothesis-based methods [9], and deep learning methods. The research papers [10,11] adopted the deep learning approach. Although the above traditional or deep learning-based low-light image enhancement methods excel in image denoising and bias color correction, these methods do not take into account the peculiarities of low-light polarized images; thus, these methods are not applicable to low-light polarized image enhancement tasks. The reason is as follows: Stokes vector is obtained for the sum or difference operation of intensity images in different directions, and none of the traditional methods in the process of intensity image enhancement consider the light intensity change characteristics between polarized intensity images in different directions, i.e., the correlation of the intensity values of the corresponding pixel points of the images with different polarized intensities. This causes the resolved polarized image to be distorted and produces a lot of noise and artifacts. In addition, in the polarization image solving task, the traditional polarization demosaicing methods [6,12,13] also only solve the polarization image by the enhanced polarization intensity image, and the interpolation process still results in a large amount of noise; moreover, all of them fail to take into account the quality degradation of the RGB image due to the polarization filters.
Based on the above, this paper designs a multi-task convolutional neural network structure, LLSPD-Net, in terms of both low-light polarization intensity image enhancement and low-noise high-quality imaging of polarized images. The structure fully takes into account the correlation between image RGB channels and polarization intensity images at 0°, 45°, 90°, and 135°, as well as the advantage of Stokes vector complementation, which solves the problems of low-light polarization image enhancement and high-quality polarization simultaneous imaging.
In summary, this paper addresses two key issues. The first is the problem of intensity image enhancement in a low-light environment. The second is the problem of noise generated by polarization imaging in a low-light environment with serious artifacts. The main contributions of this paper can be summarized in the following threeaspects:
(1)
An imaging model for polarized images in low-light environments, LLSPD-Net, is designed to generate polarized images with less noise while enhancing low-light polarized intensity images.
(2)
In order to obtain high-quality RGB images and polarized images at the same time, we designed a Stokes complementary method to acquire polarized images with the help of the hourglass network structure, and simulated the sparse arrangement of polarization filters.
(3)
We collected a low-light polarization dataset, L-Polarization, containing different materials for indoor, outdoor, and different scenes, which is a paired dataset containing 300 sets of low-light and constant-light environments.

2. Related Works

2.1. Low-Light Image Enhancement Methods

For low-light image enhancement methods, related research can be broadly categorized into two main groups: traditional methods as well as deep learning-based methods. These two main categories of methods in low-light enhancement at the same time are inevitably considered the work of image denoising, and denoising is essential. One of the more typical traditional methods is the C-BM3D [14], which utilizes highly sparse local 3D transform domains in each channel of the luminance–chrominance color space for the design of filtering to achieve the denoising of low-light images and visual quality enhancement. With the development of deep learning in recent years, Jiang et al. [15] used an unsupervised generative adversarial network architecture to solve the problem of low-light image enhancement in the absence of low-light–normal-light image pairs. Li et al. [16] designed a trainable convolutional neural network (CNN) structure for microlight image enhancement to obtain enhanced images using the Retinex model. Ma et al. [17] established a cascading illumination learning process with weight sharing for low-light image enhancement and defined the associated unsupervised training loss to improve the generalization of their model. The above low-light processing models have achieved good results in image denoising and color correction, but they are not satisfactory for low-light polarized images. Since the polarization information reflects the difference between the intensity images [18], the enhancement process without considering the correlation between the four polarization directions of intensity images— I 0 , I 45 , I 90 , I 135 —will result in a low quality of the computed color S0 as well as problems such as DoLP and AoP noises, distortions, and severe artifacts.

2.2. Polarization Demosaicing and Depth Completion Methods

Current demosaicing methods for DoFP sensors can be broadly categorized into three groups, namely sparse representation-based, interpolation-based, and end-to-end deep learning-based methods. Typical studies of sparse representations include the research of Luo et al. [19], who combined the demosaicing task into an optimization problem to be solved by introducing RGB polarization channel correlation, adaptive subdictionary, and non-local self-similarity constraints. In addition, traditional methods broadly use interpolation, for example, Li et al. [20] proposed a technique for polarization demosaicing using Newtonian polynomial interpolation, which takes into account the interpolation error and chooses to interpolate in the polarization difference domain. For deep learning, Zeng et al. [21] designed an end-to-end fully convolutional neural network architecture, which aims to directly improve the imaging quality of the three polarization characteristics of intensity (S0), degree of line polarization (DoLP), and angle of polarization (AoP), and does not interpolate the polarization intensity images.
All of the above demosaicing methods for DoFP sensors have achieved some significant results, but none of them considered the problem of RGB image quality degradation due to the polarization filter, i.e., the polarization filter reduces the photon transmittance problem. In this paper, a sparsity DoFP sensor polarization demosaicing method based on Stokes vector complementation is designed in conjunction with previous studies, taking into account the polarization imaging in low-light conditions and the structural properties of DoFP sensors. We were inspired mainly by the idea of deep complements, such as the work of Zhang et al. [22]. Our method takes an RGB image as input and predicts dense surface normals and occlusion boundaries to solve the problem of missing pixels in the original observation. Hegde et al. [23] utilized an exact sparse depth as input to the RGB image to generate a dense depth map. The method focuses on a quadtree decomposition modeling approach. However, all of the above complementation methods are designed to be specific and, therefore, cannot be directly applied to the task of polarization information complementation.

3. Method

3.1. Overview

The aim of this paper is to enhance low-light images while ensuring that the polarization information is accurate and not true imaging. The calculation of the polarization Stokes vector is shown by Equation (1) [24], which is closely related to the enhanced polarized intensity image. Therefore, it is our aim to try to avoid the loss of polarization information while ensuring the enhancement of low-light images. Our second aim concerns the fact that DoFP sensors have a low photon transmittance due to their own structure, and polarization demodulation often requires a process of “sampling + interpolation + computation”. This results in severe noise and distortion, so we compensate for missing pixels in the Stokes vector map directly through a deep learning network with the help of a sparse sensor structure proposed in the literature [25]. This method is able to ensure high-quality RGB images while obtaining polarized images with less noise.
S 0 = 1 2 I 0 + I 45 + I 90 + I 135 S 1 = I 0 I 90 S 2 = I 45 I 135
To solve the above two problems, we constructed a multi-task convolutional neural network architecture: LLSPD-Net. As shown in Figure 4, our entire network architecture consists of two sub-networks, the intensity network and the polarization complementation network (PCN), where the input of the PCN is the output of the intensity network after enhancing the low-light image; the specific structure of its two-part network is described in detail in Section 3.2.1 and Section 3.2.2.
The raw input to our network is a low-light DoFP sensor RAW image, which is decomposed by data to obtain an RGB image and four polarization direction intensity images. The intensity image is then fed into the intensity network for low-light enhancement. Finally, the enhanced RGB image is used as the high-resolution image input to guide the compensation of the Stokes vector using PCN. The enhanced I 0 , I 45 , I 90 , and I 135 are used to generate sparse Stokes vectors, which are then predicted by the PCN network for the missing pixels of the sparse Stokes components to obtain the high-resolution Stokes vectors. S0 is acquired by utilizing the gain absorption sensitivity difference for the enhanced RGB image and thus obtaining the high-resolution S0 component.

3.2. Network Architecture

3.2.1. Intensity Network

As shown in the first branch of Figure 4 for the intensity network, we feed the mosaiced RGB image, as well as the polarized light intensity images in the four polarization directions of 0°, 45°, 90°, and 135° into this network to enhance the intensity image. Inspired by the literature [26], our network architecture consists of three main components. First, the low-light image is decomposed into an illuminance map and a reflectance map by three convolutional layers; then, the decomposed image is used as input, and the noise in the reflectance map is suppressed by using the illuminance map as a constraint through six convolutional layers to realize the denoising purpose. Finally, the illuminance map obtained in the first part along with the reflectance map obtained in the second part are fed to the DRM module, through which the detail reconstruction module is finally used to obtain the ortho-light image with better visual quality. Due to the special jump connection structure of residual networks, it can make deep neural networks easier to optimize in the training phase without causing gradient vanishing or explosion. Therefore, inspired by the literature [27], we used multiple residual modules (RMs) in the network to obtain better decomposition and denoising results. The specific structure of the RM module is shown in Figure 5, where each RM contains five convolutional layers with convolutional kernel sizes of 1, 3, 3, 3, and 1, respectively. We added a 64 × 1 × 1 convolutional layer at the shortcut connection. Each RM has a 64 × 3 × 3 convolutional layer before and after it. Due to some of the correlation between our four polarized intensity images and the RGB images, we chose to process the correlation channels jointly. In addition, our ultimate goal was to achieve high-quality polarization imaging, so we added the intensity network-enhanced RGB images and the four polarization intensity images into the polarization information complementary network (PCN), which further ensured high-quality, high-fidelity imaging of the polarization images by means of Stokes vectors.

3.2.2. Polarization Completion Network

As shown in Figure 6, our PCN network is a Stokes vector complementary network, and the overall structure contains two hourglass network branches, i.e., the RGB image guidance branch and the polarization Stokes vector complementary branch. The inputs of the network are the high-quality RGB images output from the intensity network and the sparse Stokes components S1,2 after enhancement, and the output of the network is the complementary full-resolution dense S1,2.
The RGB image is fed into the RGB image guidance branch as a full-resolution high-quality guidance image. The image is first fed into the hourglass network by a 5 × 5 convolutional coding, and the first hourglass structure is used to extract the RGB image information. The aim is to provide guidance for the second stage of the densification complement of polarized Stokes, whereby the second stage is the Stokes completion branch; this stage densifies the input sparse Stokes vector. The overall network consists of two hourglass units similar to the symmetric structure of Unet, and we also introduced two modules, channel attention block (CAB) [28], and repetitive guidance module (RG) [29], into the network. Among them, the CAB module is used to extract global features and visual features in complex scenes, using encoder-side important feature weighting, which is added with decoder features to achieve a more accurate information guidance for RGB branches.
The RG module has the function of repetitive guidance, which enhances the complementation of S1,2 information through gradual refinement. The specific structure is shown on the right side of Figure 6, which takes D ij and e 1 j as inputs, and finally obtains the refined d jk through the gradual refinement process. Then, the final polarization d j is obtained by adaptive fusion. The RG structure mainly consists of an efficient guidance (EG) module and an adaptive fusion (AF) module [29], where EG first stitches RGB images and Stokes vector maps together by 3 × 3 convolution, and then generates C × 1 × 1 features by global average pooling. Finally, a point-by-point operation is performed between the features and the Stokes input. The AF module is the union of many coarse Stokes features ( d j 1 , , d jk ) obtained by repeated guidance to generate refined Stokes vector maps. This structure introduces an adaptive fusion mechanism to refine the Stokes vector map.
In Figure 6, E i j in the encoder takes E i ( j 1 ) and D ( i 1 ) j as inputs. In the decoder, D i j takes E i j and D i ( j + 1 ) as inputs. The process when i > 1 is as follows:
E i j = Conv D ( i 1 ) j , j = 1 , Conv E i ( j 1 ) + D ( i 1 ) j , 1 < j 5 , D i j = Conv E i 5 , j = 5 , Deconv D i ( j + 1 ) + E i j , 1 j < 5 ,
In the above equation, Deconv ( · ) is the inverse convolution operation and E 1 j = Conv E 1 ( j 1 ) .

3.3. Loss Function

Our loss is divided into two parts, the first part is the intensity loss that constrains the image enhancement of low-light intensity, and the second part is the Stokes loss that realizes the Stokes vector complement. Intensity loss mainly consists of two parts: content loss and perception loss. For Stokes loss, we mainly use the mean-squared error (MSE) to calculate the loss.
Specific definitions are given below:
(1)
In the intensity loss, we use 1 loss as the content loss, and compute the perceptual loss using the features extracted from the VGG-16 pre-trained model before the activation layer. The specific definitions are as follows:
L con = 1 N i = 1 N G ^ low i G gt i
L per = 1 N i = 1 N 1 C j H j W j ϕ j G ^ low i ϕ j G gt i
where i [ 1 , N ] denotes the ith sample used to compute the loss and j denotes the jth layer of the VGG-16 pre-trained model. That is, the strength loss expression is as follows:
L G ( G ^ ) = L con + λ 1 L per
(2)
Stokes loss is defined as
L S ( S ^ ) = S ^ 1 , 2 S 1 , 2 gt 2
L Stokes = L S ( S ^ ) + λ 2 L S S ^ 1 st + L S S ^ 2 nd
where s 1 , 2 gt is the ground truth.
In summary, the overall loss function formula is as follows:
f = arg min λ 3 L Stokes + L G ( G ^ )
In Equation, λ 1 is used as a balance factor between content loss and perceptual loss, and its optimal value was empirically determined to be 0.1. λ 2 is a hyperparameter that decreases with the number of epochs, and the initial value was empirically set to 0.2. And λ 3 is a balance parameter between Stokes loss and intensity loss, which was empirically set to 0.6.

4. Experimental Section

4.1. Experimental Configurations

(1)
Dataset:
In this paper, the L-Polarization dataset is acquired using LUCID PHX050S-QC camera, manufactured by LUCCID Canada, and used to train, validate, and test our LLSPD-Net. The dataset takes into full consideration the large polarization difference between metal and dielectric materials, etc., and collects indoor and outdoor scene data with different shapes and materials, such as metal, plastic, wood, and fabric. Among them, the sparse polarization images used in this paper are acquired from reference [25]. In order to more clearly reflect the diversity of the training data in the dataset of this paper, we crop the images of the dataset into 64 × 64 subimages and map these subimages to a point within the unit circle. As shown in Figure 7, we compare the polarization dataset under two low-light conditions: L-Polarization and LLCP [30]. The x-axis and y-axis of the graph are defined as the median pixel values of s 1 / s 0 and s 2 / s 0 , respectively. Each subimage is then grouped into one of ten classes. We classify them according to the magnitude of the a = s 1 / s 0 2 + s 2 / s 0 2 values, where we classify the values of a < = P L 2 as low-polarization subimages, and those of a > = P U 2 as high-polarization subimages, where P L and P U denote the threshold boundary values for the first and the tenth categories, respectively, and the rest of the eight categories are identified by dividing the unit element centroid ring region into equal areas, which correspond to the different regions of the AoP. The P L and P U in the first column of Figure 7 are set to 0.1 and 0.4, respectively, and the distribution of the data can be seen through the point mapping of the two datasets. The vertical axis in the second column indicates the median intensity value of S0, while the vertical coordinate in the third column is the median local binary pattern (LBP) [31] of S0, reflecting the texture diversity of the subimage.
As can be seen from the figure, our L-Polarization contains a large number of highly polarized subimages (nearly 2 % ), reflecting more polarization diversity. Moreover, the L-Polarization dataset exhibits greater texture diversity in terms of median LBP compared to the LLCP data.
(2)
Training details:
We constructed training, testing, and validation datasets from 300 sets of low-light data in an 8:1:1 ratio. Our network architecture was implemented on a PC equipped with an NVIDIA A100 GPU. The epoch for the whole model training was set to 50, the batch size was 3, the initial learning rate was 0.003, and for the adaptive optimizer, we chose the Adam optimizer with better noise immunity.

4.2. Experimental Results

4.2.1. Comparison with Low-Light Image Enhancement Methods

We chose three mainstream low-light enhancement models to compare with our method. C-BM3D [14] is the traditional method, EnlightenGAN [15] and LightenNet [16] are the deep learning methods. We first compared the enhanced RGB images in terms of color, contrast, and brightness, and we also compared the enhanced RGB and polarization results with their respective ground truth (GT). Figure 8 shows the qualitative result analysis of the enhanced RGB image, the complementary Stokes vectors S1 and S2, and the generated DoLP and AoP images, which show that our model is the closest to the ground truth, and that our model works well for the polarization imaging of different materials with large polarization differences. The most positive thing is that our LLSPD-Net suppresses the obvious noise of DoLP and AoP images while guaranteeing the robust characterization of the edge information of polarized images. As shown in the figure, C-BM3D has a poor brightness effect for low-light image enhancement, but it has a strong ability to maintain the polarization state of polarization intensity images in different directions. Unfortunately, the method introduces a large amount of noise, which makes the generated DoLP and AoP produce more serious noise. Although EnlightenGAN enhances the brightness, color, and other information of the RGB image, it produces obvious distortion compared to the ground truth. Due to the existence of this distortion, it will have a great impact on the subsequent compensation of the Stokes vector and the acquisition of DoLP and AoP; furthermore, it can be seen that the polarization imaging results of EnlightenGAN noise, artifacts, and distortion phenomena are very obvious. For the LightenNet method, it is not ideal in terms of RGB image enhancement, but the method is better in terms of polarization state preservation; and although the polarization imaging effect is better, it is still far from the ground truth. In conclusion, it can be seen that our method is closest to the ground truth, and the enhanced RGB image is the best in terms of visual effect and polarization state preservation.
Qualitative comparison: We selected 20 sets of images from each dataset, the public dataset LLCP and the self-constructed dataset L-Polarization, for quantitative testing. We chose four quantitative metrics such as structural similarity (SSIM), peak signal-to-noise ratio (PSNR), patch-based contrast quality index (PCQI) [32], and angle error [33], which is a measure of the angle error of the polarization angle image to judge the performance of low-light image enhancement and polarization imaging. Among them, SSIM mainly evaluates the contrast and structural fidelity of the generated results, and a higher value indicates a higher image quality. PSNR measures the distortion of polarization and RGB images, and a higher value indicates that the image is richer in information and is less noisy. PCQI measures the contrast information of the results, and a higher value indicates that the contrast of the image is better. Angle error denotes the generating error of the AoP image. The smaller the angular error, the better the performance of the model. The average metrics of the test results are shown in Table 1, and the test results show that LLSPD-Net reflects the optimal effect in several metrics, proving that the method of this paper has a good information fidelity advantage in terms of RGB image enhancement in low-light environments and polarization image generation, and is able to maximize the high-contrast, low-noise imaging of polarization images while ensuring the enhancement of RGB images.

4.2.2. Comparison with the Basic Polarization Demosaicing Method

In order to reduce the noise of polarization imaging, we discard the traditional interpolation idea, but draw on the idea of deep completion, and use the end-to-end mapping of deep learning models to directly complement the Stokes vector for polarization imaging. Therefore, we use the LLSPD-Net Part I intensity network-enhanced intensity images with different polarization directions as source images, and compare them with several existing typical polarization demosaicing (PDM) models. Among them, Newton’s [20] and bicubic [34] interpolations are traditional methods, while ForkNet [21] is a deep learning method. The comparison results in Figure 9 evidently show that Newton’s interpolation can effectively enhance the edge profile information of DoLP and AoP, and is very sensitive to the gradient information, but the method introduces artifacts, the distortion of polarized images, and is accompanied by certain meshing results. Bicubic, on the other hand, can guarantee the gradient information of polarized images, but the method introduces serious noise, which is particularly noticeable in AoP images. Compared to the above two traditional methods, ForkNet has some advantages in noise suppression, but the AoP images generated by this method have significant distortion. In conclusion, compared to the above methods, LLSPD-Net has better stability for polarization gradient characterization with less noise and higher contrast.
The quantitative comparison of the PDM models is shown in Table 2, which shows that LLSPD−Net is the best in all metrics except the PCQI metric of AoP. Moreover, the reason why the ForkNet contrast is high is because the method introduces a contrast constraint term. Although this method reduces noise, it also leads to the loss of polarization information. Therefore, overall, our LLSPD-Net has the best performance.

4.2.3. Ablation Experiments

In order to verify the effect of each component in LLSPD−Net on the whole model, we performed ablation experiments. First, we removed the intensity network, as shown in Figure 10. It can be seen that without the enhancement of the low-illumination source image, the polarized image produced severe noise, which was caused by the lack of a high-quality RGB bootstrap image, resulting in artifacts in the complementary S1,2. In order to further explore the visual effect of the detail reconstruction module (DRM) in the intensity network for the enhancement of low-light images and the effect of polarization preservation, we removed the DRM module in the intensity network, as shown in Figure 10. Due to the module lacking in detail preservation and color correction, the enhanced RGB image produced color distortions. At the same time, from the solved Stokes vectors, i.e., the AoP and the DoLP images, it can be seen that the polarization information is seriously lost, and the polarization information generated is very far away from the GT, i.e., it generates polarization distortions. In addition, in order to explore the role of the repeat guidance (RG) module for the network in a polarization complementary network (PCN), we removed the module, and the results show that the generated polarized images are monotonically smoothed, which introduces less noise but loses the detailed information such as edges, gradients, etc., of the AoP and DoLP images.
The quantitative comparison of the ablation experiments is shown in Table 3, and it is clear that the LLSPD-Net combining the components of each part is the best in each index. It further proves that our method, using Stokes complementation, can significantly suppress the noise of polarized images. It also proves that our intensity network for low-illumination image enhancement does not cause any loss of polarization information, and that the polarization imaging is more accurate, at a high quality.

5. Conclusions

In this work, we propose a new sparse polarization imaging model for low-light environments based on Stokes vector completion. Specifically, we establish an incremental and synergistic polarization imaging model. This model contains two parts: the low-light image enhancement network and the Stokes complementary network. In addition, we acquired a paired polarization dataset (L-Polarization) that contains low- and constant-light field imaging for supervised network training. Both qualitative and quantitative results demonstrate that LLSPD-Net can enhance the realism and effectiveness of polarization image scene representation, and that our polarization imaging model for low-light environments is capable of acquiring high-contrast, high-fidelity polarization images as well as high-quality RGB images. In future work, we plan to design this model in a more concise manner to facilitate hardware piggybacking and implementation.

Author Contributions

All authors contributed significantly to the writing of this study. G.C. and J.L. conceived the study, designed the experiments, analyzed the results, and wrote the manuscript. J.D. refined the language. Y.H. , J.S. and L.J. made insightful suggestions and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 62127813, 61890960 and also by the Jilin Scientific and Technological Development Program, grant number 20210203181SF.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gao, D.; Li, Y.; Ruhkamp, P.; Skobleva, I.; Wysocki, M.; Jung, H.; Wang, P.; Guridi, A.; Busam, B. Polarimetric pose prediction. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 735–752. [Google Scholar]
  2. Kalra, A.; Taamazyan, V.; Rao, S.K.; Venkataraman, K.; Raskar, R.; Kadambi, A. Deep polarization cues for transparent object segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8602–8611. [Google Scholar]
  3. Pang, Y.; Yuan, M.; Fu, Q.; Ren, P.; Yan, D.M. Progressive polarization based reflection removal via realistic training data generation. Pattern Recognit. 2022, 124, 108497. [Google Scholar] [CrossRef]
  4. Ono, T.; Kondo, Y.; Sun, L.; Kurita, T.; Moriuchi, Y. Degree-of-linear-polarization-based color constancy. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19740–19749. [Google Scholar]
  5. Kupinski, M.; Li, L. Evaluating the Utility of Mueller Matrix Imaging for Diffuse Material Classification. J. Imaging Sci. Technol. 2020, 64, 060409-1–060409-7. [Google Scholar] [CrossRef]
  6. Qiu, S.; Fu, Q.; Wang, C.; Heidrich, W. Linear polarization demosaicking for monochrome and colour polarization focal plane arrays. Comput. Graph. Forum. 2021, 40, 77–89. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  8. Buades, A.; Coll, B.; Morel, J.M. A non-local algorithm for image denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; IEEE: Piscataway, NJ, USA, 2005; Volume 2, pp. 60–65. [Google Scholar]
  9. Priyanka, S.A.; Wang, Y.K.; Huang, S.Y. Low-light image enhancement by principal component analysis. IEEE Access 2018, 7, 3082–3092. [Google Scholar] [CrossRef]
  10. Liu, J.; Xu, D.; Yang, W.; Fan, M.; Huang, H. Benchmarking low-light image enhancement and beyond. Int. J. Comput. Vis. 2021, 129, 1153–1184. [Google Scholar] [CrossRef]
  11. Lv, F.; Li, Y.; Lu, F. Attention guided low-light image enhancement with a large scale low-light simulation dataset. Int. J. Comput. Vis. 2021, 129, 2175–2193. [Google Scholar] [CrossRef]
  12. Wen, S.; Zheng, Y.; Lu, F.; Zhao, Q. Convolutional demosaicing network for joint chromatic and polarimetric imagery. Opt. Lett. 2019, 44, 5646–5649. [Google Scholar] [CrossRef] [PubMed]
  13. Kiku, D.; Monno, Y.; Tanaka, M.; Okutomi, M. Beyond color difference: Residual interpolation for color image demosaicking. IEEE Trans. Image Process. 2016, 25, 1288–1300. [Google Scholar] [CrossRef] [PubMed]
  14. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Color image denoising via sparse 3D collaborative filtering with grouping constraint in luminance-chrominance space. In Proceedings of the 2007 IEEE International Conference on Image Processing, San Antonio, TX, USA, 16 September–19 October 2007; IEEE: Piscataway, NJ, USA, 2007; Volume 1, pp. I-313–I-316. [Google Scholar]
  15. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  16. Li, C.; Guo, J.; Porikli, F.; Pang, Y. LightenNet: A convolutional neural network for weakly illuminated image enhancement. Pattern Recognit. Lett. 2018, 104, 15–22. [Google Scholar] [CrossRef]
  17. Ma, L.; Ma, T.; Liu, R.; Fan, X.; Luo, Z. Toward fast, flexible, and robust low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5637–5646. [Google Scholar]
  18. Le Teurnier, B.; Boffety, M.; Goudail, F. Error model for linear DoFP imaging systems perturbed by spatially varying polarization states. Appl. Opt. 2022, 61, 7273–7282. [Google Scholar] [CrossRef] [PubMed]
  19. Luo, Y.; Zhang, J.; Tian, D. Sparse representation-based demosaicking method for joint chromatic and polarimetric imagery. Opt. Lasers Eng. 2023, 164, 107526. [Google Scholar] [CrossRef]
  20. Li, N.; Zhao, Y.; Pan, Q.; Kong, S.G. Demosaicking DoFP images using Newton’s polynomial interpolation and polarization difference model. Opt. Express 2019, 27, 1376–1391. [Google Scholar] [CrossRef] [PubMed]
  21. Zeng, X.; Luo, Y.; Zhao, X.; Ye, W. An end-to-end fully-convolutional neural network for division of focal plane sensors to reconstruct s 0, dolp, and aop. Opt. Express 2019, 27, 8566–8577. [Google Scholar] [CrossRef] [PubMed]
  22. Zhang, Y.; Funkhouser, T. Deep depth completion of a single rgb-d image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 175–185. [Google Scholar]
  23. Hegde, G.; Pharale, T.; Jahagirdar, S.; Nargund, V.; Tabib, R.A.; Mudenagudi, U.; Vandrotti, B.; Dhiman, A. Deepdnet: Deep dense network for depth completion task. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 2190–2199. [Google Scholar]
  24. Chipman, R.; Lam, W.S.T.; Young, G. Polarized Light and Optical Systems; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar]
  25. Liu, J.; Duan, J.; Hao, Y.; Chen, G.; Zhang, H.; Zheng, Y. Polarization image demosaicing and RGB image enhancement for a color polarization sparse focal plane array. Opt. Express 2023, 31, 23475–23490. [Google Scholar] [CrossRef] [PubMed]
  26. Hai, J.; Xuan, Z.; Yang, R.; Hao, Y.; Zou, F.; Lin, F.; Han, S. R2rnet: Low-light image enhancement via real-low to real-normal network. J. Vis. Commun. Image Represent. 2023, 90, 103712. [Google Scholar] [CrossRef]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. Learning a discriminative feature network for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1857–1866. [Google Scholar]
  29. Yan, Z.; Wang, K.; Li, X.; Zhang, Z.; Li, J.; Yang, J. RigNet: Repetitive image guided network for depth completion. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 214–230. [Google Scholar]
  30. Xu, X.; Wan, M.; Ge, J.; Chen, H.; Zhu, X.; Zhang, X.; Chen, Q.; Gu, G. ColorPolarNet: Residual dense network-based chromatic intensity-polarization imaging in low-light environment. IEEE Trans. Instrum. Meas. 2022, 71, 5025210. [Google Scholar] [CrossRef]
  31. Ahonen, T.; Hadid, A.; Pietikäinen, M. Face recognition with local binary patterns. In Proceedings of the Computer Vision-ECCV 2004: 8th European Conference on Computer Vision, Prague, Czech Republic, 11–14 May 2004; Proceedings, Part I 8. Springer: Berlin/Heidelberg, Germany, 2004; pp. 469–481. [Google Scholar]
  32. He, R.; Guan, M.; Wen, C. SCENS: Simultaneous contrast enhancement and noise suppression for low-light images. IEEE Trans. Ind. Electron. 2020, 68, 8687–8697. [Google Scholar] [CrossRef]
  33. Morimatsu, M.; Monno, Y.; Tanaka, M.; Okutomi, M. Monochrome and color polarization demosaicking using edge-aware residual interpolation. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 2571–2575. [Google Scholar]
  34. Gao, S.; Gruev, V. Bilinear and bicubic interpolation methods for division of focal plane polarimeters. Opt. Express 2011, 19, 26161–26173. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Polarization sensor structure. (a) An illustration of the polarization filter and Bayer filter distribution for the DoFP sensor. (b) The color polarization sensor used for the dataset acquisition in this paper.
Figure 1. Polarization sensor structure. (a) An illustration of the polarization filter and Bayer filter distribution for the DoFP sensor. (b) The color polarization sensor used for the dataset acquisition in this paper.
Sensors 24 03299 g001
Figure 2. (a) The 0° image; the lower left is the low-light image, and the upper right is the image after direct luminance magnification. (b) The S0 image obtained from linearly amplified low-intensity images at 0°, 45°, 90°, and 135°. (c,d) The DoLP and AoP images, respectively, under low-light conditions.
Figure 2. (a) The 0° image; the lower left is the low-light image, and the upper right is the image after direct luminance magnification. (b) The S0 image obtained from linearly amplified low-intensity images at 0°, 45°, 90°, and 135°. (c,d) The DoLP and AoP images, respectively, under low-light conditions.
Sensors 24 03299 g002
Figure 3. Noise creates polarization artifacts. Photo credit: Tim Caro.
Figure 3. Noise creates polarization artifacts. Photo credit: Tim Caro.
Sensors 24 03299 g003
Figure 4. The inputs to this network are the RGB image of the raw image after demosaicing, the sparse polarization components S1,2, and the masked image M.
Figure 4. The inputs to this network are the RGB image of the raw image after demosaicing, the sparse polarization components S1,2, and the masked image M.
Sensors 24 03299 g004
Figure 5. Specific structure diagram of the residual module (RM). It contains a total of 6 convolutional layers. One shortcut connection convolutional layer is included.
Figure 5. Specific structure diagram of the residual module (RM). It contains a total of 6 convolutional layers. One shortcut connection convolutional layer is included.
Sensors 24 03299 g005
Figure 6. Polarization complementation network, which contains an RGB image guidance branch and a Stokes vector generation branch, where the repetitive guidance module RG is used to refine the Stokes Vector.
Figure 6. Polarization complementation network, which contains an RGB image guidance branch and a Stokes vector generation branch, where the repetitive guidance module RG is used to refine the Stokes Vector.
Sensors 24 03299 g006
Figure 7. The dataset sample richness and completeness are represented in 2D and 3D forms by normalizing the Stokes parameter values of the polarization dataset. The first row is a categorical representation of the LLCP dataset samples. The second row is the categorical representation of the self−constructed L−Polarization dataset in this paper. The second column’s vertical coordinate indicates the median intensity value S0. The third column plot vertical coordinate indicates the median LBP value of S0.
Figure 7. The dataset sample richness and completeness are represented in 2D and 3D forms by normalizing the Stokes parameter values of the polarization dataset. The first row is a categorical representation of the LLCP dataset samples. The second row is the categorical representation of the self−constructed L−Polarization dataset in this paper. The second column’s vertical coordinate indicates the median intensity value S0. The third column plot vertical coordinate indicates the median LBP value of S0.
Sensors 24 03299 g007
Figure 8. The results of RGB image enhancement, Stokes vector complementation, and DoLP and AoP imaging in low−light environment. (a) Indoor environment; (b) outdoor environment.
Figure 8. The results of RGB image enhancement, Stokes vector complementation, and DoLP and AoP imaging in low−light environment. (a) Indoor environment; (b) outdoor environment.
Sensors 24 03299 g008aSensors 24 03299 g008b
Figure 9. Results of qualitative comparison with several classical PDM models.
Figure 9. Results of qualitative comparison with several classical PDM models.
Sensors 24 03299 g009
Figure 10. Ablation experimental validation of different components of the LLSPD−Net model. W/O Intensity Network denotes the result of direct polarization imaging without introducing a low-light image enhancement process, and W/O DRM denotes the effect of removing the detail enhancement module in the intensity network. W/O RG denotes the result of polarization imaging when removing the repetition guidance module in the Stokes completion network.
Figure 10. Ablation experimental validation of different components of the LLSPD−Net model. W/O Intensity Network denotes the result of direct polarization imaging without introducing a low-light image enhancement process, and W/O DRM denotes the effect of removing the detail enhancement module in the intensity network. W/O RG denotes the result of polarization imaging when removing the repetition guidance module in the Stokes completion network.
Sensors 24 03299 g010
Table 1. Quantitative comparison for average PSNR, SSIM, PCQI, and angle error.
Table 1. Quantitative comparison for average PSNR, SSIM, PCQI, and angle error.
MethodRGBS1,2DoLPAoP
PSNR SSIM PCQI PSNR SSIM PCQI PSNR SSIM PCQI PSNR SSIM PCQI Error [°]
Input26.320.630.5136.810.580.4228.420.530.4112.410.330.1115.32
C-BM3D28.590.810.8338.260.790.6930.160.570.4612.520.260.3719.51
EnlightenGAN38.230.890.6355.370.830.5336.260.680.6919.130.310.3812.33
LightenNet39.910.910.7155.460.860.6236.630.720.6120.260.390.4111.05
LLSPD-Net (ours)41.210.940.8856.210.890.7737.150.790.6922.430.490.479.14
Table 2. Quantitative Comparison for average PSNR, SSIM, and PCQI.
Table 2. Quantitative Comparison for average PSNR, SSIM, and PCQI.
MethodDoLPAoP
PSNR SSIM PCQI PSNR SSIM PCQI Error [°]
Newton’s27.420.510.4112.210.340.1315.33
Bicubic31.050.530.4613.640.260.3618.59
ForkNet35.870.630.7320.030.380.4510.86
LLSPD-Net36.950.780.7621.430.440.429.13
Table 3. Ablation experiments for quantitative metrics comparisons.
Table 3. Ablation experiments for quantitative metrics comparisons.
MethodRGBS1,2DoLPAoP
PSNR SSIM PCQI PSNR SSIM PCQI PSNR SSIM PCQI PSNR SSIM PCQI Error [°]
W/O Intensity Network34.970.890.8133.170.830.7735.260.670.6817.130.410.4315.25
W/O DRM31.470.750.7631.410.780.7931.390.580.6118.460.420.3920.23
W/O RG35.680.940.8534.380.850.8135.570.690.7322.870.470.4511.73
LLSPD-Net36.840.960.8735.470.880.8436.870.740.7623.850.510.4710.15
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, G.; Hao, Y.; Duan, J.; Liu, J.; Jia, L.; Song, J. Low-Light Sparse Polarization Demosaicing Network (LLSPD-Net): Polarization Image Demosaicing Based on Stokes Vector Completion in Low-Light Environment. Sensors 2024, 24, 3299. https://doi.org/10.3390/s24113299

AMA Style

Chen G, Hao Y, Duan J, Liu J, Jia L, Song J. Low-Light Sparse Polarization Demosaicing Network (LLSPD-Net): Polarization Image Demosaicing Based on Stokes Vector Completion in Low-Light Environment. Sensors. 2024; 24(11):3299. https://doi.org/10.3390/s24113299

Chicago/Turabian Style

Chen, Guangqiu, Youfei Hao, Jin Duan, Ju Liu, Linfeng Jia, and Jingyuan Song. 2024. "Low-Light Sparse Polarization Demosaicing Network (LLSPD-Net): Polarization Image Demosaicing Based on Stokes Vector Completion in Low-Light Environment" Sensors 24, no. 11: 3299. https://doi.org/10.3390/s24113299

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop