Next Article in Journal
Reconstruction-Based Change Detection with Image Completion for a Free-Moving Camera
Next Article in Special Issue
FPGA-Based High-Performance Embedded Systems for Adaptive Edge Computing in Cyber-Physical Systems: The ARTICo3 Framework
Previous Article in Journal
Finding the Key Periods for Assimilating HJ-1A/B CCD Data and the WOFOST Model to Evaluate Heavy Metal Stress in Rice
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Energy-Efficient Compressive Image Coding for Green Internet of Things (IoT)

School of Computer and Information Technology, Xinyang Normal University, Xinyang 464000, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(4), 1231; https://doi.org/10.3390/s18041231
Submission received: 20 March 2018 / Revised: 14 April 2018 / Accepted: 14 April 2018 / Published: 17 April 2018
(This article belongs to the Special Issue Green Communications and Networking for IoT)

Abstract

:
Aimed at a low-energy consumption of Green Internet of Things (IoT), this paper presents an energy-efficient compressive image coding scheme, which provides compressive encoder and real-time decoder according to Compressive Sensing (CS) theory. The compressive encoder adaptively measures each image block based on the block-based gradient field, which models the distribution of block sparse degree, and the real-time decoder linearly reconstructs each image block through a projection matrix, which is learned by Minimum Mean Square Error (MMSE) criterion. Both the encoder and decoder have a low computational complexity, so that they only consume a small amount of energy. Experimental results show that the proposed scheme not only has a low encoding and decoding complexity when compared with traditional methods, but it also provides good objective and subjective reconstruction qualities. In particular, it presents better time-distortion performance than JPEG. Therefore, the proposed compressive image coding is a potential energy-efficient scheme for Green IoT.

1. Introduction

1.1. Motivation and Objective

With the proliferation of Internet of Things (IoT), we have witnessed more and more devices become interconnected over the recent years. There are many applications in the IoT framework, e.g., environmental monitoring, surveillance, device tracing, et al. One benefit from the framework of IoT is that a large amount of data can be gathered in a central processing server, so that we can analyze the data and achieve the valuable information in real time. The data transmission between devices consumes lots of energy, so a Green IoT has become the focus of both the academia and industry [1,2,3]. Since visual sensors are a major energy consumer on IoT, many existing works have made great efforts to design energy-efficient ones, e.g., CITRIC [4], MicrelEye [5]. These visual sensors are only the size of a coin, and their battery can last as long as a dozen hours. Some cooperative mechanisms [6,7] are deployed in visual sensors to reduce energy consumption by only transmitting some valuable visual features, and some energy-aware protocols are also used to save energy [8,9]. In spite of that, image coding is still a heavy burden for visual sensors, e.g., the processing part, which often uses JPEG [10] to compress images, is the most consuming component in MicrelEye. Therefore, a challenge that we face for Green IoT is to design an energy-efficient image coding.
For Green IoT, Compressive Sensing (CS) [11,12] can provide a potential scheme to compress images in visual sensor. As a unique advantage of CS, random measurement captures and represents compressible images at a rate that is significantly below the Nyquist rate, bringing about a low-energy encoder. More importantly, the less investment of energy does not affect the compression efficiency. Like traditional image encoders, the CS-base one also provides a low-bandwidth bit stream. However, the price of saving energy is a big computational burden when decoding an image. The architecture of IoT caters to the imbalance between encoder and decoder for CS-based image coding. The cloud on IoT has the sufficient ability to cope with the high-cost decoding task, but we still expect the green decoder to consider both energy saving and real-time display. Therefore, the objective of this paper is to design a CS-based image codec for Green IoT, which has both a light encoder and a light decoder while providing a good recovery quality.

1.2. Related Work

Lots of existing works have tried to improve the unsatisfactory rate-distortion performance of CS-based image codec. Regardless of the computational cost, these works use many complex prior models, e.g., collaborative sparsity [13], auto-regression [14], low rank [15], etc., in order to design a variety of recovery algorithms. However, with an energy-constrained CS encoder, it is difficult to significantly improve recovery quality by investing plenty of computations, because the prior model cannot always reveal the image statistics. A shortcut is to learn some information on image statistics in advance. CS theory indicates that the required number of CS measurement by perfect recovery is related to the image sparsity. Suppose that M is CS measurement number and K is the sparse degree of image. Based on experience, a satisfactory quality can be obtained only when M is roughly equal to 4K [11], so that we can use the known information on image sparsity to properly allocate measuring resources. By learning the information on image sparsity, the recovery quality can be improved significantly at the cost of a few computations. Block CS (BCS) framework [16] is suitable for performing the above scheme. The local areas in image have different waveforms, e.g., some areas have lots of smooth components, and some have strong edges, so different image blocks have different sparse degrees in a fixed space. There is no unique formula to define the sparse degree of a block, and the existing works generally use image features to measure it, e.g., DCT coefficients [17], variance [18], saliency [19], etc. Different methods to measure sparse degree bring about different recovery qualities. With an energy-constrained CS encoder, we still need a lighter method to measure the sparse degree. The non-uniform allocation of CS measurements introduces another advantage, i.e., a light decoder, which can even enable us to abandon the traditional CS recovery algorithm and turn to a linear operator, e.g., Minimum Mean Square Error (MMSE) [20], in order to decode image in real time. To conclude, our objective can be achieved by designing the measuring allocation based on sparse degree under the BCS framework.

1.3. Main Contribution

First, we present an energy-efficient compressive encoder, in which CS measurements are allocated based on the sparse degree of each image block. Different from the existing works, we define a block-based gradient field to measure the sparse degree. The block-based gradient field models the distribution of block sparse degree, and its computation has a low complexity, so that visual sensors are likely to reduce the energy consumption, while guaranteeing a good recovery quality.
Next, we present a real-time decoder, which learns a projection matrix to recover image blocks by matrix-vector product. Projection matrix learning depends on the MMSE criterion, which fuses the statistics of image block into the matrix so as to guarantee a good recovery quality. Different from traditional CS recovery algorithms, our real-time decoder has a low energy cost, so it is a suitable way to achieve Green IoT.

2. Proposed Compressive Image Coding

2.1. Framework Overview

The block diagram of the proposed compressive image coding is depicted in Figure 1. A natural scene is captured by the CMOS sensor array to be a full-sampling image x of N = Ir × Ic in size. The original image x is compressed by the compressive encoder as CS measurements, which are transmitted to the real-time decoder. By using these CS measurements, the real-time decoder accurately reconstructs the estimation x ^ of x.
At the compressive encoder, the original image x is first divided into n blocks of B × B in size, and let xi represent the i-th block, i = 1, 2, …, n, n = N/B2. Second, we compute the block-based gradient field to measure the sparse degrees of image blocks. On the gradient field, xi has a gradient value gi, which reveals the relative level of sparse degree of xi among all of the blocks. The block-based gradient field guides the allocation of measuring resources, and it decides the measurement number Mi of xi. Then, according to Mi, we construct the measurement matrix ΦBi of xi. The elements in measurement matrix obey Gaussian distribution, and they are generated by a pseudorandom sequence [21]. Finally, xi is measured by ФBi as CS measurement vector yi. All of the CS measurement vectors are transmitted to the real-time decoder.
At the real-time decoder, we first reconstruct the measurement matrices of all the blocks. These matrices are required to be the same as those at the compressive encoder. According to the property of pseudorandom sequence, we only need to synchronize the seeds of pseudorandom sequences that are used both at the encoder and at decoder. Then, depending on MMSE criterion, we learn the projection matrix Hi of xi, and obtain the estimation x ^ i u matrix-vector product. Finally, we combine all of the block estimations into the reconstructed image x ^ .
In the following subsections, we describe the flows of compressive encoder and real-time decoder in detail.

2.2. Compressive Encoder

The core of compressive encoder is how to measure the sparse degree of each block. As shown in Figure 2, a block that is marked in blue lies in the smooth region, it has a low sparse degree in DCT space due to its stable pixel variation, and its color is quite close to those of its neighboring blocks; a block that is marked in green lies in the edge region, it has a moderate sparse degree in DCT space due to the existence of strong edge pixels, and it exhibits a strong visual contrast in some direction; a block marked in red lies in the texture region, it has a high sparse degree in DCT space due to its unstable statistics of pixels, and there is a strong visual contrast between any two blocks. We can see from the above that the visual contrast is related to the sparse degree, so it can be regarded as an index to measure the sparse degree.
The visual contrast can be measured by the block-based gradient, which is the maximum energy difference between the current block and its four neighboring blocks. As shown in Figure 3, we select four neighboring blocks xi,j (j = 1, 2, 3, 4) of xi, and calculate the energy differences between xi and them, as follows,
E i , j = 1 B 2 x i x i , j 2 ,
where ||·||2 is the l2 norm. The block-based gradient of xi is defined as
G i = max { E i , 1 , E i , 2 , E i , 3 , E i , 4 } ,
where max{·} is an operator to get the maximum value from the input set. Gi cannot be directly used to represent the block sparse degree, but its spatial distribution is similar to that of block degree. Therefore, we further normalize Gi and generate the block-based gradient field, as follows,
g i = G i i = 1 n G i ,
where gi ranges from [0, 1]. A small gi indicates that the sparse degree of xi is at a low level relative to those of other blocks, and vice verse.
According to gi, we adaptively allocate the measuring resources for each block, which guarantees that more CS measurements are allocated for blocks with a high sparse degree and fewer for blocks with a low sparse degree. At first, in order to obtain the basic recovery quality, we uniformly assign 30% of the measuring resources to all of the blocks, i.e., the initial measurement number M0i of xi is set, as follows:
M 0 i = 0.3 × M n ,
where M is the total measurement number of the whole image. The rest 70% of the measuring resources are adaptively allocated according to the block-based gradient field, i.e., the measurement number Mi of xi is set, as follows,
M i = round [ 0.7 g i M + M 0 i ] ,
where round[·] is rounding operator. Besides, in order to avoid the circumstance where the measurement number exceeds the total number of pixels in a block, we set an upper limit of measurement number B2 for each block, and the measurements that are exceeding the upper limit are evenly distributed to other blocks for the sake of simplicity.
After determining the measurement number Mi of xi, we can construct the block measurement matrix ФBi, and compress xi by the following formula,
y i = Φ B i x i ,
in which the size of ΦBi is Mi × B2, and yi is the CS measurement vector of length Mi. The CS measurement vectors of all the blocks are packed and are transmitted to the decoder.

2.3. Real-Time Decoder

The real-time decoder linearly projects the received CS measurement vectors into image blocks. The projection matrix is learned according to MMSE criterion. We expect to find an estimation x ^ i that approaches the original block xi in the statistical sense, i.e., by MMSE criterion, to obtain the estimation of each block,
min x ^ i E [ x i x ^ i 2 2 ] ,
where E[·] is the expectation function. x ^ i is reconstructed by the following matrix-vector product,
x ^ i = H i y i ,
where Hi is the projection matrix. Plug Equation (8) into the model (7), we get
min H i E [ x i H i y i 2 2 ] .
yi is the measurements on xi, and by plugging Equation (6), the model (9) can be modified as
min H i E [ x i H i Φ B i x i 2 2 ]
Suppose that the objection of the model (10) is as follows,
f [ H i ] = E [ x i H i Φ B i x i 2 2 ] .
The above function is convex and differentiable, so we take a derivative of f[Hi] with respect to Hi, and obtain the requirement
f H i = 2 H i Φ B i E [ x i x i T ] Φ B i T 2 E [ x i x i T ] Φ B i T .
By making Equation (12) to be 0, the solution of the model (10) is obtained as
H i = E [ x i x i T ] Φ B i T { Φ B i E [ x i x i T ] Φ B i T } 1 .
in which E[xixiT] is the auto-correlation matrix of xi, and its constitution is as follows,
E [ x i x i T ] = [ E ( x i , 1 x i , 1 ) E ( x i , 1 x i , 2 ) E ( x i , 1 x i , B 2 ) E ( x i , 2 x i , 1 ) E ( x i , 2 x i , 2 ) E ( x i , 2 x i , B 2 ) E ( x i , B 2 x i , 1 ) E ( x i , B 2 x i , 2 ) E ( x i , B 2 x i , B 2 ) ] .
in which E[xi,pxi,q] represents the correlation between the two pixels xi,p and xi,q in xi, p, q = 1, 2, …, B2. For natural images, we approximate E[xi,pxi,q] using the following formulation,
E [ x i , p x i , q ] = ρ δ p , q ,
where δp,q is the chessboard distance between xi,p and xi,q, and ρ is an empirical value, which is set to be 0.95. We extract smooth, edge, and texture blocks of 16 × 16 in size from 10 test images of 512 × 512 in size, and construct the sets of smooth, edge, and texture blocks, respectively. For each set, we compute the sample values of the auto-correlation matrix, and present the sample auto-correlation values of pixels on diagonal in Figure 4. From Figure 4, it can be seen that the sample values of smooth blocks vary steadily as the distance between the pixels increases, but the sample values of edge and texture blocks drop off. The estimated values of Equation (15) constitute a good fitted curve that synthesizes three different variations of smooth, edge, and texture blocks. Therefore, the estimation from Equation (15) is a good balance among the sample values of various block classes.
The linear projection method has a low computational complexity, and the whole image reconstruction only requires matrix-vector product for n times. Therefore, our decoder can guarantee both energy saving and real-time display.

3. Experimental Results

In this section, various experiments are conducted to evaluate the performance of the proposed compressive image coding. First, we evaluate the encoding complexity, and the execution time of our encoder is compared with H.264/AVC [22], HEVC [23], and DISCOVER [24], which are the traditional video coding systems. Second, the performance of our decoder is evaluated by using Peak Signal-to-Noise Ratio (PSNR) and Structural SIMilarity (SSIM) [25], and the comparison with the two popular CS recovery algorithms, OMP [26] and NESTA [27], are also presented. Finally, the performance of our overall system is compared with that of JPEG. In all of the experiments, the block size B is set to be 16, and we set the measurement rate S (=M/N) to be between 0.1 and 0.5. All of the experiments are conducted under the following computer configuration: Intel(R) Core (TM) i7 @ 3.30 GHz CPU, 8 GB, RAM, Microsoft Windows 7 64 bits, and MATLAB Version 7.6.0.324 (R2008a).

3.1. Encoder Evaluation

It is difficult to accurately evaluate the energy consumption that is required to encode an image, and instead, we use the execution time of encoding video sequence to indirectly reveal the energy consumption of image coding. The first 100 frames of Foreman, Mobile, Highway, and Container sequences with CIF resolution of 352 × 288 pixels are encoded, respectively, by the proposed system, H.264/AVC, HEVC, and DISCOVER, in which our system is written in MATLAB and others are programmed in C++. Our system uses CS to measure each video frame individually, and H.264/AVC, HEVC, and DISCOVER use their common configurations of inter-prediction. Note that: (1) The configurations of our encoder and these video encoders can provide a good reconstruction quality; (2) no matter it is video or image, the CS encoder remains the same, so that we can compare the encoding performance of our system with those of the traditional video encoders. Table 1 presents the time to encode 100 frames of various video coding systems. We can see that the proposed system requires much time as the measurement rate increases; however, it does not take more than 10 s even with a high measurement rate, e.g., the time to encode Foreman sequence requires only about 9.47 s when S is 0.5. H.264/AVC, HEVC, and DISCOVER take longer than that, particularly HEVC has a heavy computational burden. Although these results have a weak comparability due to the tradeoff between encoding complexity and reconstruction quality, it can be testified under the common test conditions that the encoder of the proposed scheme has a very low energy consumption when compared with these traditional video encoders.

3.2. Decoder Evaluation

Table 2 presents PSNR and reconstruction time of our decoder, OMP and NESTA for 512 × 512 Lenna, Barbara, Peppers, Goldhill, and Mandrill test images when S is, respectively, 0.1, 0.3, and 0.5. We achieve the MATLAB source codes of OMP and NESTA from their original authors, and remain their default configurations. We can see that our real-time decoding method achieves higher PSNR values than OMP and NESTA in most cases, e.g., when S is 0.1, our method is 8.52 dB and 6.27 dB higher than OMP and NESTA for Lenna, respectively. For Barbara, our method is 1.11 dB lower than OMP when S is 0.5, and cannot provide a big PSNR gain at any measurement rate, which indicates that our method has limited a ability to recover periodic patterns. Table 2 also lists the reconstruction time of various systems. The results indicate that our system has a low computational complexity when compared with OMP and NESTA, e.g., when S is 0.1, our method requires only 0.88 s to recover Lenna while NESTA takes 198.47 s. The execution time of our system increases with the rising measurement rate, but only slightly. Table 3 presents SSIM values for test images at the measurement rate of 0.1, 0.3, and 0.5. We can see that our decoder outperforms OMP and NESRA in most cases. For Barbara, our decoder is 0.0944, 0.0254, and 0.0038 higher than NESTA at the measurement rate of 0.1, 0.3, and 0.5, respectively. There is still SSIM degradation for our decoder when reconstructing a few of images at a high measurement rate, e.g., when reconstructing Mandrill at S = 0.3, our decoder is 0.1346 less than NESTA, which indicates that our decoder could lose some structural information for rich-texture images, e.g., Barbara, Goldhill, and Mandrill, at a high measurement rate. The last column of Table 3 lists also the average SSIM values on all test images. It can be seen that our decoder achieves the better results than OMP and NESTA at S = 0.1 and 0.5. Figure 5 presents the visual reconstruction results of Lenna by various recovery algorithms at S = 0.1, 0.3 and 0.5, respectively. At any measurement rate, our method provides a pleasing result when compared with those that were reconstructed by OMP and NESTA. With our method, the surfaces and edges of objects are better preserved, and blocking artifacts are reduced significantly. From the above, we can see that our real-time decoding method provides better objective and subjective qualities with low energy consumption.

3.3. Overall Evaluation

Because the most of visual sensor network platforms use JPEG to compress the captured video sequence [28], we compare the performance of our overall system with JPEG, and their time-distortion and rate-distortion curves for the CIF test sequences Foreman and Container are presented in Figure 6. As shown in Figure 6a, the PSNR of our system rises rapidly as the encoding time increases, and its encoding time keeps at a low level, regardless of the PSNR values. When we set various quantization steps, the encoding time of JPEG changes little, and it is higher than that of our system. For JPEG, there is no obvious linear correlation between PSNR and encoding time, so it is difficult to control energy consumption by adjusting the reconstruction quality. From Figure 6a, we can see that our method is a superior energy-efficient image coding scheme to JPEG. However, Figure 6b shows that the rate-distortion performance of our system is not as good as that of JPEG, so we need to further improve the rate-distortion performance of our system in future.

4. Conclusions

In this paper, we propose an energy-efficient compressive image coding system that adaptively measures each block according to the block-based gradient field of image, which reveals the variation of block sparse degree. At the compressive encoder, under the guidance of gradient field, more CS measurements are allocated for blocks with a low sparse degree and fewer for blocks with a high sparse degree. At real-time decoder, according to MMSE criterion, a projection matrix is learned, and it is used to linearly reconstruct all of the image blocks. Experimental results show that the proposed system provides a high decoding quality with a low encoding energy. When compared with JPEG, it also shows a better time-distortion performance. However, the rate-distortion curve of our system is not so satisfactory, and in future, we will develop more efficient decoding schemes to improve the rate-distortion performance.
As the research in this paper is exploratory, there are many intriguing questions that future work should consider. First, our image coding scheme should be deployed into an actual hardware platform. By the energy consumption of our scheme on this hardware platform, we are to verify the trade-off among the number of CS measurements, energy consumption, and the reconstruction quality. Second, we should analyze the trade-off on energy between local codec and transmission. Importantly, by some simulation results, we hope to construct an empirical model to measure this trade-off in the future work.

Acknowledgments

This work was supported in part by the National Natural Science Foundation of China, under Grants Nos. 61501393, 61601396 and 61572417, and in part by Nanhu Scholars Program for Young Scholars of Xingyang Normal University.

Author Contributions

Ran Li (R.L.) contributed to the concept of the research, performed the data analysis and wrote the manuscript; Xiaomeng Duan (X.D.) and Xu Li (X.L.) helped perform the data analysis with constructive discussions and helped performed the experiment. Wei He (W.H.) conducted the experiment and collected the data. Yanling Li (Y.L.) revised the manuscript, and provided the necessary support for the research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mahapatra, R.; Nijsure, Y.; Kaddoum, G.; Hassan, N.U.; Yuan, C. Energy efficiency tradeoff mechanism towards wireless green communication: A Survey. IEEE Commun. Surv. Tutor. 2016, 18, 686–705. [Google Scholar] [CrossRef]
  2. Al-Turjman, F.; Ever, E.; Zahmatkesh, H. Green femtocells in the IoT era: Traffic modeling and challenges—An overview. IEEE Netw. 2017, 31, 48–55. [Google Scholar] [CrossRef]
  3. Liu, Q.; Ma, Y.; Alhussein, M.; Zhang, Y.; Peng, L. Green data center with IoT sensing and cloud-assisted smart temperature control system. Comput. Netw. 2016, 101, 104–112. [Google Scholar] [CrossRef]
  4. Chen, P.; Ahammad, P.; Boyer, C.; Huang, S.I.; Lin, L.; Lobaton, E.; Meingast, M.; Oh, S.; Wang, S.; Yan, P.; et al. CITRIC: A low-bandwidth wireless camera network platform. In Proceedings of the ACM/IEEE International Conference on Distributed Smart Cameras, Stanford, CA, USA, 7–11 September 2008; pp. 1–10. [Google Scholar]
  5. Kerhet, A.; Magno, M.; Leonardi, F.; Boni, A.; Benini, L. A low-power wireless video sensor node for distributed object detection. J. Real Time Image Process. 2007, 2, 331–342. [Google Scholar] [CrossRef]
  6. Mikhaylov, K.; Jamsa, J.; Luimula, M.; Tervonen, J.; Autio, V. Intelligent sensor interfaces and data format. In Intelligent Sensor Networks: Across Sensing, Signal Processing, and Machine Learning; Hu, F., Hao, Q., Eds.; Taylor & Francis LLC, CRC: Boca Raton, FL, USA, 2012. [Google Scholar]
  7. Pooranian, Z.; Shojafar, M.; Naranjo, P.G.V.; Chiaraviglio, L.; Conti, M. A novel distributed fog-based networked architecture to preserve energy in fog data centers. In Proceedings of the IEEE International Conference on Mobile Ad Hoc and Sensor Systems, Orlando, FL, USA, 22–25 October 2017; pp. 604–609. [Google Scholar]
  8. Idzikowski, F.; Chiaraviglio, L.; Cianfrani, A.; Vizcaíno, J.L.; Polverini, M.; Ye, Y. A survey on energy-aware design and operation of core networks. IEEE Commun. Surv. Tutor. 2017, 18, 1453–1499. [Google Scholar] [CrossRef]
  9. Baiocchi, A.; Chiaraviglio, L.; Cuomo, F.; Salvatore, V. Joint management of energy consumption, maintenance costs, and user revenues in cellular networks with sleep modes. IEEE Trans. Green Commun. Netw. 2017, 1, 167–181. [Google Scholar] [CrossRef]
  10. Gonzalez, R.C.; Woods, R.E.; Eddins, S.L. Digital Image Processing Using MATLAB; Pearson Education: Delhi, India, 2004; pp. 197–199. [Google Scholar]
  11. Candè, E.J.; Wakin, M.B. An introduction to compressive sampling. IEEE Signal Process. Mag. 2008, 25, 21–30. [Google Scholar] [CrossRef]
  12. Baraniuk, R.G. Compressive Sensing. IEEE Signal Process. Mag. 2007, 24, 118–121. [Google Scholar] [CrossRef]
  13. Zhang, J.; Zhao, D.; Zhao, C.; Xiong, R.; Ma, S.; Gao, W. Image compressive sensing recovery via collaborative sparsity. IEEE J. Emerg. Sel. Top. Circuits Syst. 2012, 2, 380–391. [Google Scholar] [CrossRef]
  14. Wu, X.; Dong, W.; Zhang, X.; Shi, G. Model-assisted adaptive recovery of compressed sensing with imaging applications. IEEE Trans. Image Process. 2012, 21, 451–458. [Google Scholar] [PubMed]
  15. Dong, W.; Shi, G.; Li, X.; Ma, Y.; Huang, F. Compressive sensing via nonlocal low-rank regularization. IEEE Trans. Image Process. 2014, 23, 3618–3632. [Google Scholar] [CrossRef] [PubMed]
  16. Gan, L. Block compressed sensing of natural images. In Proceedings of the International Conference on Digital Signal Processing, Cardiff, UK, 1–4 July 2007; pp. 403–406. [Google Scholar]
  17. Stankovi, V.; Stankovi, L.; Cheng, S. Compressive image sampling with side information. In Proceedings of the IEEE International Conference on Image Processing, Cairo, Egypt, 7–10 November 2009; pp. 3001–3004. [Google Scholar]
  18. Zhang, J.; Xiang, Q.; Yin, Y.; Chen, C.; Luo, X. Adaptive compressed sensing for wireless image sensor networks. Multimed. Tools Appl. 2017, 76, 4227–4242. [Google Scholar] [CrossRef]
  19. Yu, Y.; Wang, B.; Zhang, L. Saliency-based compressive sampling for image signals. IEEE Signal Process. Lett. 2010, 17, 973–976. [Google Scholar]
  20. Zhang, L.; Li, X.; Zhang, D. Image denoising and zooming under the linear minimum mean square-error estimation framework. IET Image Process. 2012, 6, 273–283. [Google Scholar] [CrossRef]
  21. Marsaglia, G.; Tsang, W.W. The ziggurat method for generating random variables. J. Stat. Softw. 2000, 5, 1–7. [Google Scholar] [CrossRef]
  22. ITU Telecom. Advanced Video Coding for Generic Audio-Visual Services; ITU-T Recommendation H. 264 and ISO/IEC 14496-10 (AVC), ITU-T and ISO/IEC JTC 1; ITU Telecom: Geneva, Switzerland, 2003. [Google Scholar]
  23. Sullivan, G.J.; Ohm, J.-R.; Han, W.-J.; Wiegand, T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  24. Girod, B.; Aaron, A.M.; Rane, S.; Rebollo-Monedero, D. Distributed video coding. Proc. IEEE 2005, 93, 71–83. [Google Scholar] [CrossRef]
  25. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  26. Shen, Y.; Li, S. Sparse signals recovery from noisy measurements by orthogonal matching pursuit. Inverse Probl. Imaging 2015, 9, 231–238. [Google Scholar] [CrossRef]
  27. Becker, S.; Bobin, J.; Candes, E.J. NESTA: A fast and accurate first-order method for sparse recovery. SIAM J. Imaging Sci. 2011, 4, 1–39. [Google Scholar] [CrossRef]
  28. Tavli, B.; Bicakci, K.; Zilan, R.; Barcelo-Ordinas, J.M. A survey of visual sensor network platforms. Multimed. Tools Appl. 2012, 60, 689–726. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed compressive image coding.
Figure 1. Block diagram of the proposed compressive image coding.
Sensors 18 01231 g001
Figure 2. Illustration of visual contrasts among blocks on smooth, edge and texture regions of Lenna. Note that the size of block is 16 × 16.
Figure 2. Illustration of visual contrasts among blocks on smooth, edge and texture regions of Lenna. Note that the size of block is 16 × 16.
Sensors 18 01231 g002
Figure 3. Illustration of computing the block-based gradient of xi.
Figure 3. Illustration of computing the block-based gradient of xi.
Sensors 18 01231 g003
Figure 4. Comparison of sample and estimation of auto-correlation matrix.
Figure 4. Comparison of sample and estimation of auto-correlation matrix.
Sensors 18 01231 g004
Figure 5. Comparison of subjective visual qualities of Lenna reconstructed by different recovery algorithms at S = 0.1, 0.3 and 0.5, respectively. From left to right: OMP, NESTA and the proposed algorithm. (a) S = 0.1, (b) S = 0.3, and (c) S = 0.5. Note that S is the measurement rate.
Figure 5. Comparison of subjective visual qualities of Lenna reconstructed by different recovery algorithms at S = 0.1, 0.3 and 0.5, respectively. From left to right: OMP, NESTA and the proposed algorithm. (a) S = 0.1, (b) S = 0.3, and (c) S = 0.5. Note that S is the measurement rate.
Sensors 18 01231 g005
Figure 6. Time-distortion and rate-distortion curves of the proposed system and JEPG for Foreman and Container sequences. (a) Time-distortion curve and (b) Rate-distortion curve. Left is Foreman, and Right is Container.
Figure 6. Time-distortion and rate-distortion curves of the proposed system and JEPG for Foreman and Container sequences. (a) Time-distortion curve and (b) Rate-distortion curve. Left is Foreman, and Right is Container.
Sensors 18 01231 g006
Table 1. The time to encode 100 frames of various video coding systems.
Table 1. The time to encode 100 frames of various video coding systems.
SequenceTime/s
ProposedH.264/AVCHEVCDISCOVER
S = 0.1S = 0.3S = 0.5
Foreman3.916.549.47389.412306.6036.40
Mobile3.936.499.44296.322880.1260.08
Highway3.946.489.84372.601550.3131.21
Container3.976.519.79292.831301.9940.65
Table 2. Peak Signal-to-Noise Ratio (PSNR) (in dB) and reconstruction time (in s) of different recovery algorithms.
Table 2. Peak Signal-to-Noise Ratio (PSNR) (in dB) and reconstruction time (in s) of different recovery algorithms.
AlgorithmSLennaBarbaraPeppersGoldhillMandrill
PSNR/dBTime/sPSNR/dBTime/sPSNR/dBTime/sPSNR/dBTime/sPSNR/dBTime/s
OMP0.118.892.8016.642.7917.282.8020.492.8115.712.81
NESTA21.14198.4718.90174.1620.25197.2821.11115.3218.04159.45
Proposed27.410.8821.780.8826.790.8826.300.8919.760.88
OMP0.327.354.2924.024.2224.355.4023.864.3417.644.34
NESTA27.28134.3422.99170.1126.86126.0126.32125.4220.46117.25
Proposed32.671.7724.681.6631.361.7030.401.6822.911.65
OMP0.531.645.9728.355.9731.116.0829.196.0721.045.97
NESTA30.83120.8925.52100.8131.16117.1929.63110.1722.77196.12
Proposed36.042.9327.242.8234.112.9333.402.8825.622.75
Table 3. Structural SIMilarity (SSIM) comparison of different recovery algorithms.
Table 3. Structural SIMilarity (SSIM) comparison of different recovery algorithms.
AlgorithmSLennaBarbaraPeppersGoldhillMandrillAvg.
OMP0.10.59530.48230.55890.53140.33120.4998
NESTA0.72110.61040.73850.61490.43220.6234
Proposed0.82490.70480.83000.76380.58760.7422
OMP0.30.88500.84820.88260.82070.63340.8140
NESTA0.90360.82560.91390.94290.95960.9091
Proposed0.94090.85100.93180.91470.82500.8927
OMP0.50.95150.93670.94120.91640.79460.9081
NESTA0.95810.91470.95960.93520.85480.9245
Proposed0.97120.91850.96100.95950.91480.9450

Share and Cite

MDPI and ACS Style

Li, R.; Duan, X.; Li, X.; He, W.; Li, Y. An Energy-Efficient Compressive Image Coding for Green Internet of Things (IoT). Sensors 2018, 18, 1231. https://doi.org/10.3390/s18041231

AMA Style

Li R, Duan X, Li X, He W, Li Y. An Energy-Efficient Compressive Image Coding for Green Internet of Things (IoT). Sensors. 2018; 18(4):1231. https://doi.org/10.3390/s18041231

Chicago/Turabian Style

Li, Ran, Xiaomeng Duan, Xu Li, Wei He, and Yanling Li. 2018. "An Energy-Efficient Compressive Image Coding for Green Internet of Things (IoT)" Sensors 18, no. 4: 1231. https://doi.org/10.3390/s18041231

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop