Next Article in Journal
Active Multimodal Sensor System for Target Recognition and Tracking
Previous Article in Journal
Review of Portable and Low-Cost Sensors for the Ambient Air Monitoring of Benzene and Other Volatile Organic Compounds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Colorization-Based RGB-White Color Interpolation using Color Filter Array with Randomly Sampled Pattern

1
Department of Electrical and Electronic Engineering, Yonsei University, 50 Yonsei-ro, Seodaemun-Ku, Seoul 03722, Korea
2
Department of Software Engineering, Dongseo University, 47 Jurye-ro, Sasang-Ku, Busan 47011, Korea
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(7), 1523; https://doi.org/10.3390/s17071523
Submission received: 26 April 2017 / Revised: 23 June 2017 / Accepted: 24 June 2017 / Published: 28 June 2017
(This article belongs to the Section Physical Sensors)

Abstract

:
Recently, several RGB-White (RGBW) color filter arrays (CFAs) have been proposed, which have extra white (W) pixels in the filter array that are highly sensitive. Due to the high sensitivity, the W pixels have better SNR (Signal to Noise Ratio) characteristics than other color pixels in the filter array, especially, in low light conditions. However, most of the RGBW CFAs are designed so that the acquired RGBW pattern image can be converted into the conventional Bayer pattern image, which is then again converted into the final color image by using conventional demosaicing methods, i.e., color interpolation techniques. In this paper, we propose a new RGBW color filter array based on a totally different color interpolation technique, the colorization algorithm. The colorization algorithm was initially proposed for colorizing a gray image into a color image using a small number of color seeds. Here, we adopt this algorithm as a color interpolation technique, so that the RGBW color filter array can be designed with a very large number of W pixels to make the most of the highly sensitive characteristics of the W channel. The resulting RGBW color filter array has a pattern with a large proportion of W pixels, while the small-numbered RGB pixels are randomly distributed over the array. The colorization algorithm makes it possible to reconstruct the colors from such a small number of RGB values. Due to the large proportion of W pixels, the reconstructed color image has a high SNR value, especially higher than those of conventional CFAs in low light condition. Experimental results show that many important information which are not perceived in color images reconstructed with conventional CFAs are perceived in the images reconstructed with the proposed method.

1. Introduction

Up to the present, most digital imaging systems obtain a full color image using a single sensor to reduce the cost and size of the system. The surface of the sensor is covered by a patterned color filter, called the color filter array (CFA), where each pixel in the pattern passes through only a certain color corresponding to a particular spectral band. The most widely used CFA pattern is the Bayer CFA pattern [1], where each pixel passes through only one of the primary colors (Red, Green and Blue), and the ratio of the numbers of the R, G, and B pixels is 1:2:1. Since each pixel captures only one of the primary colors, the other missing colors have to be obtained by interpolating neighboring colors. A lot of research has been done to find a good interpolation technique that reconstructs the missing colors as good as possible with respect to the given color filter array [2,3,4,5,6,7,8,9,10,11].
In order to use spectral information other than the RGB channel spectrum, a lot of study have been done on multispectral filter arrays [12,13,14,15,16]. By utilizing the additional spectral information obtained by the multispectral filter array, it is possible to reduce color artifacts [12,13], discriminate objects more precisely [14], and obtain invisible information such as that of NIR [15,16]. Simultaneously with these researches, demosaicing methods for various patterns have been also developed [17,18,19,20,21].
Recently, new CFA patterns [22,23,24,25] and demosaicing methods [18,19,20,21] have been proposed which also contain panchromatic “white” (W) pixels in the pattern. The W pixel has a much wider spectral band than the R, G, and B pixels as shown in Figure 1, and therefore can absorb more photons than other color pixels, making it more robust against the image noise. There are many causes of image noise, which is produced by the sensors and circuitry of a digital imager [26,27]. The fact that the W pixels are more robust against the noise is due to the fact that the signal to noise ratio (SNR) increases as the number of captured photons increases. This is especially true in low light condition, since in low light condition, the energies of the signal and the noise are similar and increasing the energy of the signal has a great effect in the SNR value. With conventional methods, the conversion from the acquired RGBW pattern to the color image is usually done in two steps: first, the conversion of the RGBW pattern to the widely used Bayer CFA pattern, and second, the conversion of the Bayer CFA pattern to the color image. The reason that the RGBW pattern is first converted to the Bayer CFA pattern is that this conversion is relatively easy and that numerous demosaicing and denoising algorithms [2,3,4,5,6,7,8,9,10,11,28,29] exist for the Bayer CFA pattern. Furthermore, conventional Bayer-oriented imaging signal processors(ISP) can be utilized with this approach. However, the reconstructed color image is degraded by this two-step conversion, since both steps introduce aliasing artifact and color distortion, which aggravates when combined together.
In this paper, we propose a new RGBW CFA pattern that has a large ratio of W pixels in the pattern and also propose the corresponding demosaicing method for this pattern. Due to the large ratio of W pixels in the CFA pattern, a large amount of light can be absorbed by the CFA pattern, resulting in a reconstructed color image with a high SNR value. Therefore, the proposed RGBW CFA pattern shows a great advantage over conventional ones in low light condition, since in low light condition, the quality of the reconstructed color image depends critically on the energy of the light absorbed by the CFA. The proposed demosaicing method for the proposed CFA pattern is based on the colorization technique which was initially proposed for automatic colorization of a gray image by the computer graphic society. The proposed demosaicing method is capable of recovering the RGB color channels from a small set of color pixels, which is the main reason that the proposed CFA pattern can have a relatively small set of RGB pixels while leaving a large space left for W pixels. Furthermore, the demosaicing method directly converts the proposed RGBW CFA pattern image to the color image, and therefore, the aliasing artifact introduced by the two step conversion in conventional RGBW CFA methods is reduced.

2. Preliminaries

To understand the proposed method, the following preliminaries have to be understood.

2.1. RGB-White CFA

A conventional Bayer CFA pattern contains twice as many G pixels as the R and B pixels as shown in Figure 2a. This is due to the fact that the spectral band of the G channel lies between the R and B channels, and therefore, has a relatively high spectral correlation with those channels. Furthermore, the G channel is highly related with the luminance channel of the image, and therefore, has a large influence on the perceptual quality of the color image, thus, it is advantageous that the G channel possesses a large portion of the CFA pattern.
However, as can be seen in Figure 1, compared to the G channel, the W channel has a much wider spectral band and also a larger overlapping region with the R and B spectral bands. Due to the wide spectral band, the W channel absorbs more light than other channels, and therefore, has a larger SNR value than other channels, which is especially true in low light condition. Furthermore, the W channel can be regarded as the luminance channel, and contains by itself most of the perceptual information in the image. Due to those facts, recently, it has been considered to use the W channel as the major channel in the CFA pattern, and many different RGBW CFAs have been proposed. For example, Figure 2b shows the RGBW CFA proposed in [22].
Compared with other RGBW CFAs [22,23,25] where the W pixels occupy about 50% of the CFA pattern, the proportion of the W channel in the proposed CFA pattern is much higher. This is possible because we use a color interpolation method based on the colorization technique, which can recover the colors in the large proportion of white pixels.

2.2. Levin’s Colorization

The color interpolation method for the proposed CFA is based on the colorization method [30]. In this method, a color image is estimated from a monochrome image with a small number of color seeds which contain the chrominance information as shown in Figure 3. In [30], Levin et al. propose how to formulate the colorization process as an optimization problem. Let M denote the number of pixels in the color image, r be the pixel position index in raster-scan order ( 1 r M ) , y be an M × 1 vector denoting the luminance channel, u , an M × 1 vector denoting the chrominance channel (U or V) to be solved, and x , an M × 1 sparse vector containing the chrominance values only in certain positions (called as the representative pixels) and zeros in all other positions. Furthermore, let Ψ be the set of all r where x ( r ) 0 , and N ( r ) be the set of the 8-neighborhood pixels of the r-th. The colorization process is performed by minimizing the following energy function:
J ( u ) = p Ψ { x ( p ) u ( p ) } 2 + r Ψ { u ( r ) s N ( r ) ω r s u ( s ) } 2 .
Here, ω r s is a weighting value defined by
ω r s e ( y ( r ) y ( s ) ) 2 / 2 σ r 2 .
Define the weighting matrix W of size M × M as follows
W ( r , s ) = ω r s if   r Ψ and   s N ( r ) 0 otherwise ,
and define A = I W , where I is an M × M identity matrix. Then, (1) can be expressed as
J ( u ) = x A u 2 .
The minimizer u of J ( u ) is the color channel constructed from the small number of color seeds, i.e., from x . The colorization technique brought about the idea that the R, G, and B channels can be reconstructed by a few number of sensed color pixels which contain the true color information, and the full-resolution white channel.

3. Proposed Method

3.1. Proposed Randomly Sampled RGBW CFA Pattern

The proposed RGBW CFA pattern differs from other RGBW patterns in two aspects: it has a larger ratio of W pixels, and the positions of the R, G, and B pixels are random. Figure 4 shows the proposed randomly sampled RGBW pattern. pattern is 75%, and thus receives more light energy than other RGBW CFA patterns. This makes it stronger against the noise. Furthermore, since the W channel corresponds to the luminance channel which determines the resolution of the color image, a high resolution image can be obtained with the proposed CFA. The remaining 25% area of the CFA is composed of RGB pixels, which are randomly distributed over the pattern. The reason that we use a random RGB pixel pattern rather than a periodic one is that the aliasing artifact in the color channels can be reduced. In other RGBW CFAs, the periodic RGBW pattern is converted into the color image by directional interpolation. In this case, some pixels with certain directions cannot be included in the interpolation process, which results in the aliasing artifact.
Figure 5a shows the case of estimating the G channel value at an R pixel in a RGBW pattern which has same proportion of R, G, B, and W pixels, but whose R, G, and B pixels are periodically distributed over the pattern. Here, no G pixels are in the lower-left to upper-right diagonal direction. Therefore, the true G values in this direction cannot be referenced. As a result, aliasing artifact occurs if high frequency components of the G channel exist in this direction. Figure 5c shows the color interpolation result of the RGBW pattern in 5a, where the aliasing artifact can be observed in the upper right and lower left areas. In comparison, the proposed random RGBW pattern has G pixels in all directions, and therefore, the R pixel in Figure 5b can refer to those pixels in estimating the G channel value. Furthermore, the method in referring to those pixels is non-directional, i.e., based on a diffusion method as will be explained later. As a result, the aliasing artifact in the upper-right and lower-left areas is reduced as compared to the periodically sampled RGBW CFA, as can be seen in Figure 5d.

3.2. W Channel Interpolation

Most of the demosaicing methods for the Bayer CFA pattern first interpolate the G channel, since it contains most of the spatial information. The information of the interpolated G channel is then used in the subsequent R and B channel interpolation by referring to the inter-channel spectral correlations. Meanwhile, with conventional RGBW CFAs, the RGBW pattern is first converted to the RGB Bayer pattern, and then to the color image. Therefore, the demosaicing method is similar to the one with Bayer CFA pattern, including only an extra process of converting the RGBW pattern to the RGB pattern.
In comparison, with the proposed method, we reconstruct all the W pixels first, and then recover the colors by the colorization-based interpolation. This is due to the fact that the W channel has a higher spectral correlation with the R and B channels than the G channel, and therefore, is more suitable as a reference channel for recovering the color information. Furthermore, only 25% of the W channel has to be recovered with the proposed RGBW pattern, thus the recovery becomes highly reliable. Figure 6 shows a 3 × 3 local region of the RGBW CFA pattern where the W value of the central pixel has to be estimated. Here, w denotes the W channel, and c denotes one of the primary color channels, i.e., c { r , g , b } . With the Bayer CFA pattern, only four pixels, i.e., the pixels in the horizontal and vertical directions, can be referred to interpolate the missing G value. In comparison, the 8-neighborhood pixels can be used to estimate the missing W pixel values ( w ^ ) with the proposed RGBW CFA pattern. Let { i , j } denote the coordinates of the missing W pixel, N ( i , j ) represent the set of the 8 neighborhood of { i , j } , and { u , v } be the coordinates of the pixels belonging to N ( i , j ) . We define a weighting function α u , v which determines how much the white pixel value in position { u , v } , i.e., w u , v , should contribute to the reconstruction of the missing value at { i , j } , i.e., w i , j as follows:
α u , v = 1 β i , j Δ u , v i n t e r + Δ u , v i n t r a , u , v N ( i , j ) .
Here, α u , v consists of two terms, the inter-channel term Δ u , v i n t e r and the intra-channel term Δ u , v i n t r a , where β i , j acts as a balance between these two terms. Then, reconstruction of w ^ is performed by the following equation:
w ^ i , j = u , v N ( i , j ) α u , v w u , v u , v N ( i , j ) α u , v .
Equation (6) is a weighted interpolation, where the weights are α u , v . As will be seen later, the term β i , j in (5) is designed to have a large value in achromatic regions, and a small value in colorful regions. If the difference between w u , v and w i , j is large, this means that the pixels { u , v } and { i , j } belong to different regions in the image, and therefore, the value w u , v should not contribute to the reconstruction of w i , j . Therefore, the weight α u , v should be inversely proportional to the difference between w u , v and w i , j .
In achromatic regions, the color value at { i , j } itself is similar to the missing value w i , j , and thus, the difference between w i , j and w u , v can be approximately measured by the difference between c i , j and w u , v , where c i , j refers to one of the primary color values at { i , j } . Therefore, the inter-channel term is defined as
Δ u , v i n t e r = | c i , j w u , v | , u , v N ( i , j ) .
On the contrary, in colorful regions, c i , j is not similar to w i , j , and therefore, the difference between w u , v and w i , j should be measured by the differences of adjacent pixels. For example, if { u , v } = { i 1 , j 1 } , the difference between w i , j and w u , v is measured by the average of | w i , j 1 w i + 1 , j | and | w i 1 , j w i , j + 1 | as illustrated in the top left image in Figure 7.
The intra-channel terms Δ u , v i n t r a for different positions of { u , v } in N ( i , j ) are defined respectively as
Δ u = i 1 , v = j 1 i n t r a = ( | w i 1 , j w i , j + 1 | + | w i , j 1 w i + 1 , j | ) / 2 , Δ u = i , v = j 1 i n t r a = ( | w i 1 , j 1 w i 1 , j | + | w i + 1 , j 1 w i + 1 , j | ) / 2 , Δ u = i 1 , v = j + 1 i n t r a = ( | w i 1 , j w i , j 1 | + | w i , j + 1 w i + 1 , j | ) / 2 , Δ u = i , v = j + 1 i n t r a = ( | w i 1 , j + 1 w i 1 , j | + | w i + 1 , j + 1 w i + 1 , j | ) / 2 , Δ u = i + 1 , v = j 1 i n t r a = ( | w i 1 , j w i , j 1 | + | w i , j + 1 w i + 1 , j | ) / 2 , Δ u = i 1 , v = j i n t r a = ( | w i 1 , j 1 w i , j 1 | + | w i 1 , j + 1 w i , j + 1 | ) / 2 , Δ u = i + 1 , v = j + 1 i n t r a = ( | w i 1 , j w i , j + 1 | + | w i , j 1 w i + 1 , j | ) / 2 , Δ u = i + 1 , v = j i n t r a = ( | w i + 1 , j 1 w i , j 1 | + | w i + 1 , j + 1 w i , j + 1 | ) / 2 .
The value β i , j which determines whether the pixel { i , j } belongs to an achromatic region or a colorful region is defined as
β i , j = 0.5 if m i , j i n t e r m i , j i n t r a 2 if m i , j i n t e r 2 m i , j i n t r a 1.5 ( m i , j i n t e r m i , j i n t r a ) m i , j i n t r a + 0.5 otherwise ,
and has a large value in achromatic regions and a small value in colorful regions. Here, m i , j i n t e r and m i , j i n t r a are the minimum values related with Δ u , v i n t e r and Δ u , v i n t r a , and are defined as
m i , j i n t e r = min ( Δ i 1 , j 1 i n t e r + Δ i + 1 , j + 1 i n t e r , Δ i 1 , j + 1 i n t e r + Δ i + 1 , j 1 i n t e r , Δ i , j 1 i n t e r + Δ i , j + 1 i n t e r , Δ i 1 , j i n t e r + Δ i + 1 , j i n t e r ) m i , j i n t r a = min ( Δ i 1 , j 1 i n t r a + Δ i + 1 , j + 1 i n t r a , Δ i 1 , j + 1 i n t r a + Δ i + 1 , j 1 i n t r a , Δ i , j 1 i n t r a + Δ i , j + 1 i n t r a , Δ i 1 , j i n t r a + Δ i + 1 , j i n t r a ) .

3.3. Primary Color Channel Interpolation

As described in Section 3.1, the number of color seeds, i.e., pixels which are sensing the primary colors, is very small, and their positions are randomly distributed. After the W channel is fully interpolated with the aforementioned method, the primary color channels can be obtained with a colorization scheme. However, in low light conditions, the primary color channels are more degraded by the noise than the W channel. Therefore, in this work, we design a colorization matrix which simultaneously diffuse the original color seeds to other pixels and remove the noise in the original color seeds.
Let w and c represent the lexicographically ordered vectors corresponding to the 2-dimensional images w and c, i.e.,
w = [ w 1 , 1 , w 2 , 1 , , w N r , 1 , w 1 , 2 , , w 1 , N c , , w N r , N c ] T ,
and
c = [ c 1 , 1 , c 2 , 1 , , c N r , 1 , c 1 , 2 , , c 1 , N c , , c N r , N c ] T ,
where N r and N c denote the height and the width of the image. Here, c represents the color channel which we want to recover from the sensed RGBW pattern image.
While other sophisticated residual images such as that proposed in [11] can also be used, we use a simple color difference channel. The color difference channel u c between the primary color channel c { r , g , b } and the W channel is defined as:
u c = c w .
Only 25% of the components in the u c vector are sensed by the proposed RGBW CFA, and the problem of the primary color channel reconstruction, i.e., the color channel interpolation, is to recover the true u c vector with proper color components at every position. We define by x c the vector which contains the u c values only at the positions where the corresponding c value is sensed by the proposed RGBW CFA and has zero values at all other positions, i.e.,
x c ( m ) = c ( m ) w ^ ( m ) if m Ψ c 0 otherwise ,
where w ^ is the full W channel reconstructed by the method in Section 3.2, m is the position index of the pixel in the lexicographically ordered vector x c , and Ψ c represents the set of the pixel positions where the channel c value is sensed by the RGBW CFA. In accordance with [30], we call the pixels in the set Ψ c the representative pixels or the color seeds.
The proposed colorization-based color interpolation solves the interpolation problem by minimizing the following functional with respect to u c given x c and w ^ :
J ( u c ) = D ( u c ) + F ( u c ) .
The functional J ( u c ) consists of two energy terms: the diffusion term D ( u c ) and the noise-suppressing fidelity term F ( u c ) . Let r denote the position index of the pixels not belonging to Ψ c , and N ( r ) the neighborhood pixels of r. The diffusion term is defined as
D ( u c ) = r Ψ c { u c ( r ) s N ( r ) ϵ r s u c ( s ) } 2 .
The minimization of D ( u c ) can be seen as weighted diffusion shown in Figure 8a, where the amount of diffusion between r and s is determined by the weight ϵ r s . The weight ϵ r s is controlled by an edge directional ellipsoidal kernel as will be explained later. Let p denote the position index of the pixels belonging to Ψ c , and N ( p ) the neighborhood pixels of p. The noise-suppressing fidelity term F ( u c ) is defined as
F ( u c ) = p Ψ c [ u c ( p ) x ( p ) + γ { u c ( p ) q N ( p ) ϵ p q u c ( q ) } ] 2 .
It tries to not only make u c similar to x c , but also subtract the weighted Laplacian of u c , i.e., γ { q N ( p ) ϵ p q u c ( q ) u c ( p ) } from u c at the seed pixel. It is different from the classical cost function defined as follow:
J o l d ( u c ) = p Ψ c [ u c ( p ) x c ( p ) ] 2 + γ p [ { u c ( p ) q N ( p ) ϵ p q u c ( q ) } ] 2 .
Minimizing (18), the neighborhood pixels tend to follow the noisy seed pixels and result in low frequency noise.
The first term in (17) preserves the fidelity of u c ( p ) to the sensed value at p, i.e., x c ( p ) , while the second term tries to smooth it as a weighted average of the values of the neighborhood pixels as shown in Figure 8b, where the weights ϵ p q are obtained in the same manner as ϵ r s as will be explained later. The second term is necessary because, unlike the original colorization problem, the color seeds in the proposed method are degraded by the noise, especially in low light condition. This leads to spotted color noise after applying the colorization-based interpolation process. The value γ controls the balance between the two terms, thus controlling the degree of diffusion. When γ is large, the values of u c at the position of the color seeds are influenced much by the nearby color values, whereas if γ = 0 , they are not influenced at all by the nearby values. Here, γ is proportional to the noise variance. Let σ n 2 denote the noise variance and τ represent a control constant, γ is defined as
γ = τ σ n 2 .
The noise variance σ n 2 can be estimated by calculating the variance of the flat region in the image. In [30], the weight parameters were determined based on the differences of the luminance values between adjacent pixels, so that the color seeds were easily spread out in flat regions, while not in edge regions. In low light conditions, however, the difference between adjacent pixel values could not be estimated correctly due to the noise. As some edge regions were weakened by the noise, the color diffused across the edge region which results in the wash-out of the colors in edge regions. As the edge could not be determined exactly using the derivative at the current position only, the derivatives in a neighborhood region should be considered together to determine the edge direction.
The weight parameter is designed to take into account the directional tendency of adjacent pixels [31,32]. In order to estimate the tendency of the edge direction, the covariance matrix C is determined in a small region R ( m ) centered at m
C = n R ( m ) w ^ v ( n ) w ^ v ( n ) n R ( m ) w ^ v ( n ) w ^ h ( n ) n R ( m ) w ^ v ( n ) w ^ h ( n ) n R ( m ) w ^ h ( n ) w ^ h ( n ) ,
where w ^ v and w ^ h represent the derivatives of w ^ in the vertical and the horizontal directions, respectively. Let d m n = [ v m n , h m n ] T be a pointing vector, where v m n and h m n are the vertical and the horizontal distances between m, the current pixel position, and n, a neighborhood pixel of m. The weight value at n is calculated as
ϵ m n = exp { λ d m n T C d m n } .
It is derived from an ellipsoidal kernel function (Figure 9) which has a large value if n lies at the center of the kernel and a small value if n is far from the center. Here, λ represents a parameter which controls the smoothness of the kernel. Furthermore, ϵ m n is large if n lies in a direction orthogonal to the edge, and small if n lies in the direction of the edge. Therefore, ϵ m n can be used as a measure whether a pixel n lies along or across an edge. For the weights in (16) and (17), we use a normalized version of ϵ m n , i.e., ϵ m n = 1 k ϵ m n , where k is the normalizing factor.
The cost function defined in (15) can be re-written in matrix form as
J ( u c ) = D u c 2 + F u c x c 2 ,
where the diffusion matrix D which spreads the color seeds is defined as
D ( r , s ) = 1 if   r Ψ c and   s = r ϵ r s else   if   r Ψ c and   s N ( r ) 0 otherwise .
The noise-suppressing fidelity matrix F is defined as
F ( p , q ) = 1 + γ if   p Ψ c and   q = p γ ϵ p q else   if   p Ψ c and   q N ( p ) 0 otherwise .
Solving (22), the color difference channel can be estimated as follows:
u ˜ c = ( D T D + F T F ) 1 F T x c .
By using the estimated color difference channel u ^ c , the primary color channel can be obtained as
c ˜ = u ˜ c + w ^ .

3.4. Post Processing

The color channels c ˜ { r ˜ , g ˜ , b ˜ } reconstructed by the method explained in Section 3.3 suffer from low frequency noise. This is due to the fact that the color seeds from which they are reconstructed contain noise themselves. In this section, we propose a post-process which subtracts the low frequency noise n L F from c ˜ to obtain a noise removed color channel c ^ :
c ^ = c ˜ n L F .
Here, the low frequency noise n L F is estimated by subtracting w ^ obtained in Section 3.2 from w ˜ which denotes a noisy W channel to be constructed as follow:
n L F = w ˜ w ^ .
This is based on the assumptions that the low frequency noise contained in c ˜ is highly correlated with the noise in w ˜ , and that w ^ is free from this low frequency noise. The first assumption is satisfied since the noisy w ˜ is constructed by a linear combination of the noisy color channels:
w ˜ = λ r r ˜ + λ g g ˜ + λ b b ˜ + λ 1 .
Here, λ r , λ g , λ b , and λ 1 are linear coefficients which are obtained by the following L 2 norm minimization with respect to multiple arguments λ r , λ g , λ b , and λ 1 :
λ r , λ g , λ b , λ 1 = argmin λ r , λ g , λ b , λ 1 w ^ ( λ r r ˜ + λ g g ˜ + λ b b ˜ + λ 1 ) 2 2 .
This is a multiple regression model [33] which makes the L 2 norm difference between w ^ and are w ˜ as small as possible, thus resulting in the coefficients λ r , λ g , λ b and λ 1 that the overall brightness of w ˜ becomes similar to w ^ . This is similar to the guided filter approach proposed in [34], but instead of a single parameter we use multiple parameters, and unlike the guided filter which tries to obtain a noise-free channel, the purpose is to obtain a noisy white image w ˜ which is used afterwards to eliminate the noise in the reconstructed noisy color channels. The constructed channel w ˜ contains the same low frequency noise as in c ˜ since it is constructed from it. The second assumption is satisfied since w ^ is constructed from the sensed W channel, which has a larger SNR value than those of the color channels. Using the linear least square method, (30) can be solved as follow:
λ r λ g λ b λ 1 = μ { r ˜ } μ { g ˜ } μ { b ˜ } 1 μ { r ˜ r ˜ } μ { g ˜ r ˜ } μ { b ˜ r ˜ } μ { r ˜ } μ { r ˜ g ˜ } μ { g ˜ g ˜ } μ { b ˜ g ˜ } μ { g ˜ } μ { r ˜ b ˜ } μ { g ˜ b ˜ } μ { b ˜ b ˜ } μ { b ˜ } 1 μ { w ^ } μ { w ^ r ˜ } μ { w ^ g ˜ } μ { w ^ b ˜ } ,
where a { r ˜ , g ˜ , b ˜ , w ^ } , μ { a } denotes the mean of a vector a , and ∘ denotes the entrywise product operator. Figure 10a–d show w ˜ , w ^ , c ˜ , and c ^ , respectively. As can be seen in Figure 10d, the low frequency noise in c ^ is quite reduced.

4. Experimental Results

We compared the quality of the demosaicing result of the proposed method based on the proposed RGBW CFA pattern with the demosaicing results of other methods based on other CFA patterns. Experimental results show that in low light condition, spatial information which cannot be obtained by the Bayer CFA pattern using demosaicing method can be obtained by the proposed method. Furthermore, we also show that the proposed method produces results of good quality in sufficient light condition. In order to obtain the original full resolution R, G, B and W channel images, we used a filter-wheel-installed camera, where the wheel contains four different optical filters and is driven by a stepping motor. The four optical filters selectively filter the R, G, B, and W bands. We took four photographs of the same scene with the four different filters to obtain the R, G, B, and W channels. Figure 1 shows the spectral responsibilities of the R, G, B, and W filters. Here, the exposure time was set to 30 ms at 100 lux illumination. Using the full-resolution four channel images, we sampled the R, G, B, and W pixels corresponding to the different patterns, i.e., the Bayer CFA pattern, the Sony RGBW CFA pattern, and the proposed RGBW CFA pattern. For the proposed RGBW CFA pattern, the numbers of the R, G, and B pixels are 306,500, 306,697, and 306,489, respectively, which are not equal since there are randomly sampled, but together they cover 25% of the CFA pattern which has a total of 3,686,400 pixels since the size of the image is 1920 × 1920 . The control constant τ is set to a value around 100.
The images in the first column in Figure 11 show the demosaiced results of the Bayer CFA, and those in the second column are the demosaiced results of the Sony RGBW CFA. For the Bayer CFA results, the DLMMSE method [7] was used, while with the Sony RGBW CFA, the RGBW pattern was first converted into the Bayer CFA pattern using the method in [22], and then finally got demosaiced by the DLMMSE method. The images in the third column in Figure 11 show the results of the proposed RGBW CFA which are reconstructed by the proposed colorization-based interpolation method. As described in Section 3.1, the results using the Sony RGBW CFA shows the largest aliasing artifacts due to the extra errors in the conversion from the RGBW to the Bayer CFA. The Bayer CFA based conversion shows less aliasing artifacts in the circle region than the proposed RGBW CFA based conversion in the first row of Figure 11. However, in the vertical stripe region, the result with the proposed method is better than those using other methods in terms of both the aliasing and the noise artifacts as can be seen in the second row of Figure 11.
Next, for an assessment in low light condition, we obtained the full-resolution R, G, B, and W images using the four different optical filters with an exposure time of 1 / 60 s at 1 lux illumination. After that, we sampled the R, G, B, and W pixels corresponding to the Bayer CFA and the proposed RGBW CFA patterns. The image acquired at 1 lux illumination is very dark as shown in Figure 12a. The energy levels of the R, G, B, and W channels are different due to the different spectral responses of illumination and sensor sensitivity. Therefore, we first multiplied a large number (420) to the pixels of the acquired image to stretch the ranges in the pixel values, and then applied a white balancing as a pre-process to obtain Figure 12b. We performed the gray world method [35] which is one of the simplest white balancing method, but other methods [35,36,37,38,39,40] can be applied as well with little difference in the reconstruction performance. Figure 12b illustrates the result and its corresponding histogram after rescaling the intensity range and applying white balancing to the acquired image. The color interpolation is then performed on this white-balanced image, resulting in the color image shown in Figure 12c.
In Figure 13, we compared the results of the proposed method with the images reconstructed from the Bayer CFA sampled images using the state-of-the-art Bayer denoising methods [28,29]. The methods in [28,29] utilize variations of the BM3D(Block-matching and 3D filtering) method for the denoising of the Bayer-patterned image. We show that in low illumination, the proposed method results in better reconstructed images even without any extra sophisticated denoising methods as in [28,29] which is due to the nature of the proposed CFA having a very large number of W pixels. In the case with 20 dB and 23 dB noise simulations, the BM3D denoising removes the noise well, resulting in a reconstructed image of higher SNR values than that obtained by the proposed method not using a sophisticated denoising method. To make a fair comparison, we also applied the BM3D method to the W channel as a post-denoising method. We cannot apply the exactly same BM3D method as in [28,29] to our method, because they are mainly designed for the Bayer pattern, which is different from the proposed RGBW pattern. Therefore, we applied the BM3D denoising [41] on the W channel, and afterward used it as the reference channel. We call the modified proposed method with BM3D denoising the ‘proposed + BM3D’ method. The input parameter for the denoising methods in [28,29,41] is the level of noise, for which we give the standard deviation of the noise. The standard deviation of noise in the RGB channels is 0.25 (40 in 8 bit system), and that of the noise in the W channel is 0.06 (16 in 8 bit system) in 1 lux low light condition.
Figure 13 compares the results reconstructed by the proposed RGBW CFA, the Bayer CFAs [7,11], the denoised Bayer CFAs [28,29], and the Sony RGBW CFA [22]. It can be seen in Figure 13, that the results obtained by the proposed method exhibit superior high frequency information compared with other results. Moreover, the objects are better identified, and the image details, such as letters and lines, are more distinguished, especially in Figure 13p compared with Figure 13a,d,g,j,m. Figure 13p has also less noise than Figure 13a,d,g in flat regions. The results with Bayer CFA denoising show a lot of noise removal in flat regions, but also have lost important detail information. The ‘proposed + BM3D’ method shows the best result both in flat and detailed regions which can be observed in Figure 13s,t,u. Here, the denoising is done on the reconstructed W channel, which is then used for the colorization of the color seeds.
For objective assessments, we used the Kodak 24 test images of size 512 × 768 and measured the FSIM [42] and PSNR of the reconstructed images, where the PSNR is given by
P S N R = 10 × log 10 { 1 / MSE } .
We made two sets of test images, one containing 20 dB noise, and the other 23 dB noise, where the noise follows a Poisson distribution as is the case in low illumination. The noise was added to the R, G, and B channels, respectively. The standard deviations of the noises in the RGB channels are 0.1 (26 in 8 bit system) and 0.07 (18 in 8 bit system), respectively. The standard deviations of the noises in the W channel are 0.1 / 3 = 0.06 (16 in 8 bit system) and 0.07 / 3 = 0.04 (10 in 8 bit system) in 20 dB and 23 dB, respectively. This is due to the fact that the W channel is a superposition of the R, G, and B channels. The standard deviations of the noises go in as the parameters to the BM3D denoising methods. After that, the pixels were sampled according to the Bayer CFA, the Sony RGBW CFA, and the proposed RGBW CFA pattern, respectively, to generate the CFA pattern images. For the proposed RGBW CFA, the numbers of the color seeds for the R, G, and B channels are 32,456, 32,550, and 32,659, respectively, which together sums up to 25% of the whole pixels ( 512 × 768 = 393 , 216 ), and the rest of the pixels (75%) are W pixels. Then, the images sampled by the proposed RGBW CFA were reconstructed by the colorization-based method, while the images sampled by Bayer CFA are demosaiced by the DLMMSE and RI methods. We also added the results using the methods in [28,29]. As shown in Table 1, Table 2, Table 3 and Table 4, the color images reconstructed by the proposed method show significant improvements compared with the results of the Bayer CFA and Sony RGBW CFA. Bold text in the tables are the optimum results for each Kodak images.
The PSNR is improved by an average of 3.2934 dB, 3.2938 dB, and 0.8639 dB, compared to the results of the DLMMSE, RI, and Sony RGBW CFA, respectively, with 20 dB noise. In the case of 23 dB noise, the PSNR is improved by an average of 2.7725 dB, 2.3157 dB, and 0.5639 dB, compared to the results of the DLMMSE, RI, and Sony RGBW CFA, respectively. As shown in Figure 14 and Figure 15, the visual qualities of the images obtained by the proposed method are superior to those by the Bayer CFA and Sony RGBW CFA.
The PSNR and FSIM values of the images obtained by combining the BM3D denoising are higher than the those obtained by the proposed method without BM3D denoising. However, the images reconstructed by the ‘proposed + BM3D’ method show higher FSIM values and similar PSNR values compared with the results of [28,29]. As shown in Figure 14h,p and Figure 15h,p, the results of the ‘proposed + BM3D’ show similar noise levels in flat regions, but preserve more detail information. Also the ‘proposed + BM3D’ introduces fewer aliasing artifacts in high frequency regions as shown in the window bars in Figure 14 and Figure 15. The Bayer CFA denoising methods not only remove the noise in the flat regions but also the details in high frequency regions. This is not reflected effectively in the PSNR measure, but in the FSIM measure which reflects the edge fidelity and the color difference consistency, which is the reason that the FSIM values of the ‘proposed + BM3D’ method are higher in all cases.
Figure 16 illustrates the relation between the modulation transfer function(MTF) and the aliasing artifact in the reconstructed images with the proposed method. The value of the horizontal axis in Figure 16a is equal to the line number shown in Figure 16d, which also represents the spatial frequency value. This is due to the fact that the line becomes thinner and denser as the line number increases as can be seen in the yellow boxes in Figure 16d. When the line number is sufficiently low, aliasing is weak as can be seen in Figure 16e. The aliasing artifact is strongest in the region of line 16 (Figure 16f), and again decreases in the region of line 20 due to the lens diffraction (Figure 16g) which also decreases the MTF value. Meanwhile, the MTF is greatly affected by the noise in low light conditions and color artifacts occur due to the diffusion of false signals as can be observed in Figure 16c which is obtained with low lighting, compared to Figure 16b obtained with high lighting.

5. Conclusions

In this paper, we proposed a colorization-based demosaicing method which suits well with the proposed RGBW CFA pattern, having a large number of white pixels. Using the proposed demosaicing technique, color images can be reconstructed using a small number of sensed color pixels while the majority of the pixels can sample the white channel, which makes it possible to obtain a reconstructed color image with high SNR value. Compared with the Bayer CFA based results, the image details, such as the letters or lines, are better preserved with the proposed method in low illumination condition. This is important especially for surveillance camera systems, and therefore, the proposed method can find its application in such areas.

Acknowledgments

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT and Future Planning (No. 2015R1A2A1A14000912).

Author Contributions

Paul Oh and Sukho Lee designed the experiments, performed the experiments, analyzed the data and wrote the manuscript under the supervision of Moon Gi Kang.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bayer, B. Color Imaging Array. U.S. Patent 3,971,065, 20 July 1976. [Google Scholar]
  2. Kimmel, R. Demosaicing: Image reconstruction from color CCD samples. IEEE Trans. Image Process. 1999, 8, 1221–1228. [Google Scholar] [CrossRef] [PubMed]
  3. Pei, S.C.; Tam, I.K. Effective color interpolation in CCD color filter arrays using signal correlation. IEEE Trans. Circuits Syst. Video Technol. 2003, 13, 503–513. [Google Scholar]
  4. Gunturk, B.K.; Altunbasak, Y.; Mersereau, R.M. Color plane interpolation using alternating projections. IEEE Trans. Image Process. 2002, 11, 997–1013. [Google Scholar] [CrossRef] [PubMed]
  5. Alleysson, D.; Süsstrunk, S.; Hérault, J. Linear demosaicing inspired by the human visual system. IEEE Trans. Image Process. 2005, 14, 439–449. [Google Scholar] [PubMed]
  6. Gunturk, B.K.; Glotzbach, J.; Altunbasak, Y.; Schafer, R.W.; Mersereau, R.M. Demosaicking: Color filter array interpolation. IEEE Signal Process. Mag. 2005, 22, 44–54. [Google Scholar] [CrossRef]
  7. Zhang, L.; Wu, X. Color demosaicking via directional linear minimum mean square-error estimation. IEEE Trans. Image Process. 2005, 14, 2167–2178. [Google Scholar] [CrossRef] [PubMed]
  8. Dubois, E. Frequency-domain methods for demosaicking of Bayer-sampled color images. IEEE Signal Process. Lett. 2005, 12, 847–850. [Google Scholar] [CrossRef]
  9. Pekkucuksen, I.; Altunbasak, Y. Multiscale Gradients-Based Color Filter Array Interpolation. IEEE Trans. Image Process. 2013, 22, 157–165. [Google Scholar] [CrossRef] [PubMed]
  10. Menon, D.; Calvagno, G. Color Image Demosaicking: An Overview. Signal Process. Image Commun. 2011, 26, 518–533. [Google Scholar] [CrossRef]
  11. Kiku, D.; Monno, Y.; Tanaka, M.; Okutomi, M. Beyond Color Difference: Residual Interpolation for Color Image Demosaicking. IEEE Trans. Image Process. 2016, 25, 1288–1300. [Google Scholar] [CrossRef] [PubMed]
  12. Monno, Y.; Kikuchi, S.; Tanaka, M.; Okutomi, M. A Practical One-Shot Multispectral Imaging System Using a Single Image Sensor. IEEE Trans. Image Process. 2015, 24, 3048–3059. [Google Scholar] [CrossRef] [PubMed]
  13. Yasuma, F.; Mitsunaga, T.; Iso, D.; Nayar, S.K. Generalized Assorted Pixel Camera: Postcapture Control of Resolution, Dynamic Range, and Spectrum. IEEE Trans. Image Process. 2010, 19, 2241–2253. [Google Scholar] [CrossRef] [PubMed]
  14. Prasad, D.K. Strategies for Resolving Camera Metamers Using 3+1 Channel. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 954–962. [Google Scholar]
  15. Schaul, L.; Fredembach, C.; Süsstrunk, S. Color image dehazing using the near-infrared. In Proceedings of the 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 1629–1632. [Google Scholar]
  16. Sadeghipoor, Z.; Lu, Y.M.; Süsstrunk, S. A novel compressive sensing approach to simultaneously acquire color and near-infrared images on a single sensor. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 1646–1650. [Google Scholar]
  17. Rafinazari, M.; Dubois, E. Demosaicking algorithm for the Fujifilm X-Trans color filter array. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 660–663. [Google Scholar]
  18. Tian, Q.; Lansel, S.; Farrell, J.E.; Wandell, B.A. Automating the design of image processing pipelines for novel color filter arrays: Local, linear, learned (L3) method. Proc. SPIE 2014, 9023. [Google Scholar] [CrossRef]
  19. Condat, L. A Generic Variational Approach for Demosaicking from an Arbitrary Color Filter Array. In Proceedings of the 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 1605–1608. [Google Scholar]
  20. Gu, J.; Wolfe, P.J.; Hirakawa, K. Filterbank-based universal demosaicking. In Proceedings of the 17th IEEE International Conference on Image Processing (ICIP), Hong Kong, China, 26–29 September 2010; pp. 1981–1984. [Google Scholar]
  21. Park, S.W.; Kang, M.G. Generalized color interpolation scheme based on intermediate quincuncial pattern. J. Electron. Imaging 2014, 23, 030501. [Google Scholar] [CrossRef]
  22. Tachi, M. Image Processing Device, Image Processing Method, and Program Pertaining to Image Correction. U.S. Patent 8,314,863, 20 November 2012. [Google Scholar]
  23. Yamagami, T.; Sasaki, T.; Suga, A. Image Signal Processing Apparatus Having a Color Filter with Offset Luminance Filter Elements. U.S. Patent 5,323,233, 21 June 1994. [Google Scholar]
  24. Gindele, E.; Gallagher, A. Sparsely Sampled Image Sensing Device with Color and Luminance Photosites. U.S. Patent 6,476,865, 5 November 2002. [Google Scholar]
  25. Compton, J.; Hamilton, J. Image Sensor with Improved Light Sensitivity. U.S. Patent 8,139,130, 20 March 2012. [Google Scholar]
  26. Dougherty, G. Digital Image Processing for Medical Applications; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  27. Bottacchi, S. Noise and Signal Interference in Optical Fiber Transmission Systems: An Optimum Design Approach; Wiley: Hoboken, NJ, USA, 2008. [Google Scholar]
  28. Akiyama, H.; Tanaka, M.; Okutomi, M. Pseudo four-channel image denoising for noisy CFA raw data. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 4778–4782. [Google Scholar]
  29. Danielyan, A.; Vehvilainen, M.; Foi, A.; Katkovnik, V.; Egiazarian, K. Cross-color BM3D filtering of noisy raw data. In Proceedings of the 2009 International Workshop on Local and Non-Local Approximation in Image Processing, Tuusula, Finland, 19–21 August 2009; pp. 125–129. [Google Scholar]
  30. Levin, A.; Lischinski, D.; Weiss, Y. Colorization using optimization. ACM Trans. Graph. 2004, 23, 689–694. [Google Scholar] [CrossRef]
  31. Takeda, H.; Farsiu, S.; Milanfar, P. Kernel regression for image processing and reconstruction. IEEE Trans. Image Process. 2007, 16, 349–366. [Google Scholar] [CrossRef] [PubMed]
  32. Feng, X.; Milanfar, P. Multiscale Principal Components Analysis for Image Local Orientation Estimation. In Proceedings of the 36th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 3–6 November 2002; pp. 478–482. [Google Scholar]
  33. Freedman, D. Statistical Models : Theory and Practice; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  34. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef] [PubMed]
  35. Buchsbaum, G. A spatial processor model for object colour perception. J. Frankl. Inst. 1980, 310, 1–26. [Google Scholar] [CrossRef]
  36. Land, E.H.; McCann, J.J. Lightness and Retinex Theory. J. Opt. Soc. Am. 1971, 61, 1–11. [Google Scholar] [CrossRef] [PubMed]
  37. Land, E. The Retinex Theory of Color Vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef] [PubMed]
  38. Barnard, K.; Cardei, V.; Funt, B. A comparison of computational color constancy algorithms. I: Methodology and experiments with synthesized data. IEEE Trans. Image Process. 2002, 11, 972–984. [Google Scholar] [CrossRef] [PubMed]
  39. Barnard, K.; Martin, L.; Coath, A.; Funt, B. A comparison of computational color constancy Algorithms. II. Experiments with image data. IEEE Trans. Image Process. 2002, 11, 985–996. [Google Scholar] [CrossRef] [PubMed]
  40. Cheng, D.; Prasad, D.K.; Brown, M.S. Illuminant estimation for color constancy: Why spatial-domain methods work and the role of the color distribution. J. Opt. Soc. Am. A 2014, 31, 1049–1058. [Google Scholar] [CrossRef] [PubMed]
  41. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  42. Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. FSIM: A Feature Similarity Index for Image Quality Assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) W channel and (b) RGB channel images in the low light condition(1lux); and (c) the transmission graph of the R, G, B, and W color filters in the visible spectrum band.
Figure 1. (a) W channel and (b) RGB channel images in the low light condition(1lux); and (c) the transmission graph of the R, G, B, and W color filters in the visible spectrum band.
Sensors 17 01523 g001
Figure 2. Various CFA patterns: (a) Bayer [1]; (b) Sony RGBW [22]; (c) Yamagami [23]; (d) Gindele [24]; (e) Compton [25]; and (f) Proposed RGBW.
Figure 2. Various CFA patterns: (a) Bayer [1]; (b) Sony RGBW [22]; (c) Yamagami [23]; (d) Gindele [24]; (e) Compton [25]; and (f) Proposed RGBW.
Sensors 17 01523 g002
Figure 3. Comparison of (a) the original image; (b) the gray image with few number of color seeds; and (c) the colorized result of (b).
Figure 3. Comparison of (a) the original image; (b) the gray image with few number of color seeds; and (c) the colorized result of (b).
Sensors 17 01523 g003
Figure 4. (a) Original image; (b) randomly sampled RGBW CFA pattern; (c) sampled image using the pattern in (b) pattern on (a).
Figure 4. (a) Original image; (b) randomly sampled RGBW CFA pattern; (c) sampled image using the pattern in (b) pattern on (a).
Sensors 17 01523 g004
Figure 5. The G pixel estimation at the position of an R pixel with (a) a periodic RGBW CFA; and (b) a random patterned RGBW CFA; (c) Color interpolation result with pattern shown in (a); (d) Color interpolation result with pattern shown in (b).
Figure 5. The G pixel estimation at the position of an R pixel with (a) a periodic RGBW CFA; and (b) a random patterned RGBW CFA; (c) Color interpolation result with pattern shown in (a); (d) Color interpolation result with pattern shown in (b).
Sensors 17 01523 g005
Figure 6. 3 × 3 windows of (a) Bayer CFA pattern and (b) the proposed RGBW CFA pattern.
Figure 6. 3 × 3 windows of (a) Bayer CFA pattern and (b) the proposed RGBW CFA pattern.
Sensors 17 01523 g006
Figure 7. Showing the pixels involved in the calculation of the inter-channel and the intra-channel terms for the 8-neighborhood directions. Red arrows: differences considered in the inter-channel terms Δ u , v i n t e r . Black arrows: differences considered in the intra-channel terms Δ u , v i n t r a .
Figure 7. Showing the pixels involved in the calculation of the inter-channel and the intra-channel terms for the 8-neighborhood directions. Red arrows: differences considered in the inter-channel terms Δ u , v i n t e r . Black arrows: differences considered in the intra-channel terms Δ u , v i n t r a .
Sensors 17 01523 g007
Figure 8. (a) The diffusion and (b) the noise suppression process of the colorization matrix in the edge region of the RGBW sampled image. The contribution of the pixels in a region different from that in which the pixel under consideration lies is weak as visualized by the faint colored arrows.
Figure 8. (a) The diffusion and (b) the noise suppression process of the colorization matrix in the edge region of the RGBW sampled image. The contribution of the pixels in a region different from that in which the pixel under consideration lies is weak as visualized by the faint colored arrows.
Sensors 17 01523 g008
Figure 9. The noisy edge region and an ellipsoidal kernel for the pixel position m. The major axis of the ellipsoidal kernel is parallel to the edge direction.
Figure 9. The noisy edge region and an ellipsoidal kernel for the pixel position m. The major axis of the ellipsoidal kernel is parallel to the edge direction.
Sensors 17 01523 g009
Figure 10. (a) The reconstructed W channel w ˜ by (29); (b) the reconstructed W channel w ^ after applying the post-process in Section 3.2; (c) the reconstructed color channel c ˜ obtained by the method in Section 3.3; and (d) the noise removed color channel c ^ after applying the post-process in Section 3.2.
Figure 10. (a) The reconstructed W channel w ˜ by (29); (b) the reconstructed W channel w ^ after applying the post-process in Section 3.2; (c) the reconstructed color channel c ˜ obtained by the method in Section 3.3; and (d) the noise removed color channel c ^ after applying the post-process in Section 3.2.
Sensors 17 01523 g010
Figure 11. Experimental results and their aliasing artifacts in 100 lux light condition: (a,d) results using the Bayer CFA; (b,e) results using the Sony RGBW CFA; and (c,f) results using the proposed RGBW CFA.
Figure 11. Experimental results and their aliasing artifacts in 100 lux light condition: (a,d) results using the Bayer CFA; (b,e) results using the Sony RGBW CFA; and (c,f) results using the proposed RGBW CFA.
Sensors 17 01523 g011
Figure 12. Images and their histograms: (a) the acquired image with the proposed RGBW CFA pattern before white balancing; (b) the pattern image after white balancing; and (c) color image from reconstructed from (b).
Figure 12. Images and their histograms: (a) the acquired image with the proposed RGBW CFA pattern before white balancing; (b) the pattern image after white balancing; and (c) color image from reconstructed from (b).
Sensors 17 01523 g012
Figure 13. Experimental results with 1 lux low light condition. (a–c): results using the Bayer CFA with DLMMSE [7], (d–f): results using the Bayer CFA with RI [11], (g–i): results using the Sony RGBW CFA [22], (j–l): results using the denoised Bayer CFA with Akiyama [28], (m–o): results using the denoised Bayer CFA with BM3D-CFA [29], (p–r): results using the proposed RGBW CFA, and (s–u): results using the proposed RGBW CFA + BM3D.
Figure 13. Experimental results with 1 lux low light condition. (a–c): results using the Bayer CFA with DLMMSE [7], (d–f): results using the Bayer CFA with RI [11], (g–i): results using the Sony RGBW CFA [22], (j–l): results using the denoised Bayer CFA with Akiyama [28], (m–o): results using the denoised Bayer CFA with BM3D-CFA [29], (p–r): results using the proposed RGBW CFA, and (s–u): results using the proposed RGBW CFA + BM3D.
Sensors 17 01523 g013
Figure 14. Experiment results of Kodak images with 20dB noise. (a,i): original images; (b,j): results using the Bayer CFA with DLMMSE [7]; (c,k): results using the Bayer CFA with RI [11]; (d,l): results using the Sony RGBW CFA [22]; (e,m): results using the proposed RGBW CFA; (f,n): results using the denoised Bayer CFA with Akiyama [28]; (g,o): results using the denoised Bayer CFA with BM3D-CFA [29]; and (h,p): results using the proposed RGBW CFA + BM3D.
Figure 14. Experiment results of Kodak images with 20dB noise. (a,i): original images; (b,j): results using the Bayer CFA with DLMMSE [7]; (c,k): results using the Bayer CFA with RI [11]; (d,l): results using the Sony RGBW CFA [22]; (e,m): results using the proposed RGBW CFA; (f,n): results using the denoised Bayer CFA with Akiyama [28]; (g,o): results using the denoised Bayer CFA with BM3D-CFA [29]; and (h,p): results using the proposed RGBW CFA + BM3D.
Sensors 17 01523 g014
Figure 15. Experiment results of Kodak images with 23dB noise. (a,i): original images; (b,j): results using the Bayer CFA with DLMMSE [7]; (c,k): results using the Bayer CFA with RI [11]; (d,l): results using the Sony RGBW CFA [22]; (e,m): results using the proposed RGBW CFA; (f,n): results using the denoised Bayer CFA with Akiyama [28]; (g,o): results using the denoised Bayer CFA with BM3D-CFA [29]; and (h,p): results using the proposed RGBW CFA + BM3D.
Figure 15. Experiment results of Kodak images with 23dB noise. (a,i): original images; (b,j): results using the Bayer CFA with DLMMSE [7]; (c,k): results using the Bayer CFA with RI [11]; (d,l): results using the Sony RGBW CFA [22]; (e,m): results using the proposed RGBW CFA; (f,n): results using the denoised Bayer CFA with Akiyama [28]; (g,o): results using the denoised Bayer CFA with BM3D-CFA [29]; and (h,p): results using the proposed RGBW CFA + BM3D.
Sensors 17 01523 g015
Figure 16. (a) MTF graph with high lighting (650 lux, blue line) and low lighting (2 lux, red line) condition; (b) reconstruction result with high light condition; (c) reconstruction result with low light condition; (d) reconstruction result (full image) with high light condition; Showing the region of (e) line 3; (f) line 16; and (g) line 20, with the aliasing artifact in the lower parts.
Figure 16. (a) MTF graph with high lighting (650 lux, blue line) and low lighting (2 lux, red line) condition; (b) reconstruction result with high light condition; (c) reconstruction result with low light condition; (d) reconstruction result (full image) with high light condition; Showing the region of (e) line 3; (f) line 16; and (g) line 20, with the aliasing artifact in the lower parts.
Sensors 17 01523 g016
Table 1. Comparison of the PSNR (dB) values of the Kodak RGB images interpolated by the Bayer CFA, Sony RGBW CFA, denoised Bayer CFA, and proposed RGBW CFA with 20 dB noise levels.
Table 1. Comparison of the PSNR (dB) values of the Kodak RGB images interpolated by the Bayer CFA, Sony RGBW CFA, denoised Bayer CFA, and proposed RGBW CFA with 20 dB noise levels.
DataBayer DLMMSE [7]Bayer RI [11]Sony RGBW [22]ProposedAkiyama [28]BM3D-CFA [29]Proposed + BM3D
120.624021.025322.833724.030025.764026.218526.5624
221.021721.577724.144023.259528.756729.025227.2122
320.792821.558223.541423.833330.659330.832729.2358
420.781421.447323.790124.151529.301629.520828.4922
520.914121.334722.766023.852725.229226.060226.0390
620.826721.273022.950024.204227.063827.282927.6730
720.731121.375923.421824.375429.484529.928129.4038
820.727420.995522.779623.921025.455326.190326.7617
920.640621.383923.511524.592130.598230.733930.2494
1020.656321.393623.453024.591130.021530.150530.2441
1120.822721.371823.258624.329827.600528.005428.1005
1220.784121.470523.470224.505630.537330.606830.0079
1320.652720.882621.791923.437023.951624.288424.9340
1420.802221.295323.083523.700226.424127.016025.8387
1521.410221.862823.953124.430629.129129.379728.2387
1620.690321.339323.406324.537829.449929.520929.0873
1721.082521.662623.822624.837929.248429.585129.7193
1820.870021.400122.911423.967326.002226.546626.5046
1920.691421.265023.275724.360028.651629.030828.7008
2021.889922.318224.108724.827726.948827.898628.0698
2120.642921.254323.052124.220427.478927.832928.1341
2220.640021.288323.121924.122027.886428.248327.8049
2320.793421.565023.389623.216531.103630.842728.3076
2420.656121.159322.625823.892625.971326.448426.6763
Avg20.839421.395923.269324.133228.029928.383127.9999
Table 2. Comparison of the PSNR (dB) values of the Kodak RGB images interpolated by the Bayer CFA, Sony RGBW CFA, denoised Bayer CFA, and proposed RGBW CFA with 23 dB noise levels.
Table 2. Comparison of the PSNR (dB) values of the Kodak RGB images interpolated by the Bayer CFA, Sony RGBW CFA, denoised Bayer CFA, and proposed RGBW CFA with 23 dB noise levels.
DataBayer DLMMSE [7]Bayer RI [11]Sony RGBW [22]ProposedAkiyama [28]BM3D-CFA [29]Proposed + BM3D
123.474323.731525.215426.199827.164727.684028.2742
223.755524.059726.578326.085029.829530.167729.1778
323.688824.399526.525226.578431.990932.383531.0654
423.617324.178426.521026.398330.326230.660130.1931
523.691424.041224.911825.899227.062327.876127.7638
623.690224.015725.524826.336028.558828.896529.2019
723.639624.228426.333226.767231.203931.807231.1182
823.498823.625025.001926.081327.106227.803528.3955
923.586524.265926.544127.072932.100332.372031.6528
1023.630924.293526.533627.120531.492231.824731.8130
1123.674224.113125.917826.552028.980929.559329.5306
1223.682024.327626.452826.889331.794432.189631.4717
1323.414023.426823.572625.130925.567225.836426.3775
1423.554623.950825.486725.787327.871328.491527.3360
1524.092924.484626.550926.716430.434330.789729.9927
1623.582624.138726.325827.041030.661330.993231.0541
1723.886424.383526.567027.255030.864531.252331.3186
1823.570724.010225.136625.909927.590528.070728.0735
1923.582024.069626.036526.665129.862630.369630.2485
2024.675925.006526.814527.183127.462628.459230.0495
2123.526524.058625.651826.451328.996029.403529.6244
2223.497624.045025.793526.349629.141729.554429.3926
2323.669024.375726.265126.284932.373832.545730.7539
2423.424423.839324.849925.890627.496828.003728.1189
Avg.23.671124.127925.879626.443629.413929.874829.6666
Table 3. Comparison of the FSIM values of the Kodak RGB images interpolated by the Bayer CFA, Sony RGBW CFA, denoised Bayer CFA, and proposed RGBW CFA with 20 dB noise levels.
Table 3. Comparison of the FSIM values of the Kodak RGB images interpolated by the Bayer CFA, Sony RGBW CFA, denoised Bayer CFA, and proposed RGBW CFA with 20 dB noise levels.
DataBayer DLMMSE [7]Bayer RI [11]Sony RGBW [22]ProposedAkiyama [28]BM3D-CFA [29]Proposed + BM3D
10.86720.87810.89140.92440.92330.92920.9524
20.79780.80440.83130.87010.88780.91560.9382
30.75020.76710.79550.85740.93300.93910.9516
40.79610.81070.83730.88560.92230.93390.9519
50.90670.91520.91900.94180.92970.93370.9595
60.86880.87890.88570.92340.91890.92640.9512
70.82900.84040.85930.90520.94410.94660.9631
80.89910.90780.91710.94090.94620.95000.9687
90.76400.78070.80450.86440.94250.93970.9549
100.80240.81760.83570.89000.93020.93060.9530
110.84060.85120.86780.91030.92060.93100.9504
120.79990.81400.83090.88320.91240.92850.9474
130.91000.91770.91780.94310.91350.92500.9499
140.86840.87830.89130.92410.91700.92520.9500
150.78890.79980.82540.87380.92930.93830.9539
160.79340.80840.82900.88300.91230.92010.9419
170.83680.85050.86890.91270.93080.93380.9564
180.88020.88900.89850.92610.90280.91270.9430
190.82950.84100.86100.90510.92010.93120.9512
200.77830.78910.81950.86910.94400.94720.9589
210.86390.87350.88240.92050.92410.93050.9541
220.82630.83920.85780.90110.90290.91910.9424
230.73930.75700.78080.85040.95220.94320.9554
240.84170.85330.86870.90870.92340.92990.9538
Avg.0.82830.85740.84010.90060.92430.93170.9522
Table 4. Comparison of the FSIM values of the Kodak RGB images interpolated by the Bayer CFA, Sony RGBW CFA, denoised Bayer CFA, and proposed RGBW CFA with 23dB noise levels.
Table 4. Comparison of the FSIM values of the Kodak RGB images interpolated by the Bayer CFA, Sony RGBW CFA, denoised Bayer CFA, and proposed RGBW CFA with 23dB noise levels.
DataBayer DLMMSE [7]Bayer RI [11]Sony RGBW [22]ProposedAkiyama [28]BM3D-CFA [29]Proposed + BM3D
10.91540.92210.93140.94970.94800.95130.9668
20.86570.86890.89240.91790.91460.93780.9535
30.83130.84360.87050.90790.94830.95550.9642
40.86500.87570.89860.92750.94260.95250.9653
50.94130.94650.94780.96110.95580.95790.9733
60.91560.92150.93040.94890.94680.95030.9656
70.88780.89570.91320.93860.96120.96440.9738
80.93450.94010.94650.96090.96350.96680.9779
90.83980.85220.87470.91280.95990.95950.9671
100.87150.88150.90070.93130.95230.95280.9681
110.89500.90230.91610.94070.94330.95050.9653
120.86890.87850.89390.92500.93450.94660.9602
130.94280.94710.94560.95960.94530.95000.9661
140.91480.92110.93070.95010.94300.94770.9645
150.85650.86410.88750.91760.94800.95450.9650
160.86340.87350.89220.92570.93740.94280.9591
170.89360.90200.91860.94400.95180.95290.9694
180.92250.92850.93540.95020.93970.94390.9615
190.88880.89710.91360.93860.94230.94910.9648
200.85100.85760.88820.91660.95860.96270.9695
210.91310.91940.92710.94840.94890.95210.9685
220.88750.89620.91240.93530.92900.94090.9579
230.82410.83750.86060.90500.96570.96110.9665
240.89820.90590.91800.94030.94730.95160.9673
Avg.0.88700.89490.91020.93560.94700.95230.9659

Share and Cite

MDPI and ACS Style

Oh, P.; Lee, S.; Kang, M.G. Colorization-Based RGB-White Color Interpolation using Color Filter Array with Randomly Sampled Pattern. Sensors 2017, 17, 1523. https://doi.org/10.3390/s17071523

AMA Style

Oh P, Lee S, Kang MG. Colorization-Based RGB-White Color Interpolation using Color Filter Array with Randomly Sampled Pattern. Sensors. 2017; 17(7):1523. https://doi.org/10.3390/s17071523

Chicago/Turabian Style

Oh, Paul, Sukho Lee, and Moon Gi Kang. 2017. "Colorization-Based RGB-White Color Interpolation using Color Filter Array with Randomly Sampled Pattern" Sensors 17, no. 7: 1523. https://doi.org/10.3390/s17071523

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop