Next Article in Journal
Introduction to the Special Issue on Evaluating the Security of Complex Systems
Next Article in Special Issue
Efficient Software HEVC to AVS2 Transcoding
Previous Article in Journal
IoT Privacy and Security Challenges for Smart Home Environments
Previous Article in Special Issue
Speech Compression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Visually Lossless JPEG 2000 for Remote Image Browsing

1
National Satellite Operation and Application Center, Korea Aerospace Research Institute (KARI), 169-84 Gwahak-ro, Yuseong-gu, Daejeon 34133, Korea
2
Department of Biomedical Engineering, The University of Arizona, 1127 E. James E. Rogers Way, Tucson, AZ 85721, USA
3
Department of Electrical and Computer Engineering, The University of Arizona, 1230 E. Speedway Blvd, Tucson, AZ 85721, USA
*
Author to whom correspondence should be addressed.
Information 2016, 7(3), 45; https://doi.org/10.3390/info7030045
Submission received: 17 May 2016 / Revised: 22 June 2016 / Accepted: 1 July 2016 / Published: 15 July 2016
(This article belongs to the Special Issue Multimedia Information Compression and Coding)

Abstract

:
Image sizes have increased exponentially in recent years. The resulting high-resolution images are often viewed via remote image browsing. Zooming and panning are desirable features in this context, which result in disparate spatial regions of an image being displayed at a variety of (spatial) resolutions. When an image is displayed at a reduced resolution, the quantization step sizes needed for visually lossless quality generally increase. This paper investigates the quantization step sizes needed for visually lossless display as a function of resolution, and proposes a method that effectively incorporates the resulting (multiple) quantization step sizes into a single JPEG 2000 codestream. This codestream is JPEG 2000 Part 1 compliant and allows for visually lossless decoding at all resolutions natively supported by the wavelet transform as well as arbitrary intermediate resolutions, using only a fraction of the full-resolution codestream. When images are browsed remotely using the JPEG 2000 Interactive Protocol (JPIP), the required bandwidth is significantly reduced, as demonstrated by extensive experimental results.

1. Introduction

With recent advances in computer networks and reduced prices of storage devices, image sizes have increased exponentially and user expectations for image quality have increased commensurately. Very large images, far exceeding the maximum dimensions of available display devices, are now commonplace. Accordingly, when pixel data are viewed at full resolution, only a small spatial region may be displayed at one time. To view data corresponding to a larger spatial region, the image data must be displayed at a reduced resolution, i.e., the image data must be downsampled. In modern image browsing systems, users can trade off spatial extent vs. resolution via zooming. Additionally, at a given resolution, different spatial regions can be selected via panning. In this regard, JPEG 2000 has several advantages in the way it represents images. Owing to its wavelet transform, JPEG 2000 supports inherent multi-resolution decoding from a single file. Due to its independent bit plane coding of “codeblocks” of wavelet coefficients, different spatial regions can be decoded independently, and at different quality levels [1]. Furthermore, using the JPEG 2000 Interactive Protocol (JPIP), a user can interactively browse an image, retrieving only a small portion of the codestream [2,3,4]. This has the potential to significantly reduce bandwidth requirements for interactive browsing of images.
In recent years, significant attention has been given to visually lossless compression techniques, which yield much higher compression ratios compared with numerically lossless compression, while holding distortion levels below those that can be detected by the human eye. Much of this work has employed the contrast sensitivity function (CSF) in the determination of quantization step sizes. The CSF represents the varying sensitivity of the human eye as a function of spatial frequency and orientation, and is obtained experimentally by measuring the visibility threshold (VT) of a stimulus, which can be a sinusoidal grating [5,6] or a patch generated via various transforms, such as the Gabor Transform [7], Cortex Transforms [8,9,10], Discrete Cosine Transform (DCT) [11], or Discrete Wavelet Transform (DWT) [12,13].
The visibility of quantization distortion is generally reduced when images are displayed at reduced resolution. The relationship between quantization distortion and display resolution was studied through subjective tests conducted by Bae et al. [14]. Using the quantization distortion of JPEG and JPEG 2000, they showed that users accept more compression artifacts when the display resolution is lower. This result suggests that quality assessments should take into account the display resolution at which the image is being displayed. Prior to this work, in [15], Li et al. proposed a vector quantizer that defines multiple distortion metrics for reduced resolutions and optimizes a codestream for multiple resolutions by switching the metric at particular bitrates. The resulting images show slight quality degradation at full resolution compared with images optimized only for full resolution, but exhibit considerable subjective quality improvement at reduced resolutions reconstructed at the same bitrate.
In [16], Hsiang and Woods proposed a compression scheme based on EZBC (Embedded image coding using ZeroBlocks of wavelet coefficients and Context modeling). Their system allows subband data to be selectively decoded to one of several visibility thresholds corresponding to different display resolutions. The visibility thresholds employed therein are derived from the model by Watson that uses the 9/7 DWT and assumes a uniform quantization distortion distribution [12].
This paper builds upon previous work to obtain a multi-resolution visually lossless coding method [17,18] that has the following distinct features:
  • The proposed algorithm is implemented within the framework of JPEG 2000, which is an international image compression standard. Despite the powerful scalability features of JPEG 2000, previous visually lossless algorithms using JPEG 2000 are optimized for only one resolution [19,20,21]. This implies that if the image is rendered at reduced resolution, there are significant amounts of unnecessary information in the reduced resolution codestream. In this paper, a method is proposed for applying multiple visibility thresholds in each subband corresponding to various display resolutions. This method enables visually lossless results with much lower bitrates for reduced display resolutions. Codestreams obtained with this method are decodable by any JPEG 2000 decoder.
  • Visibility thresholds measured using an accurate JPEG 2000 quantization distortion model are used. This quantization distortion model was proposed in [22,23] and was developed for the statistical characteristics of wavelet coefficients and the dead-zone quantizer of JPEG 2000. This model provides more accurate visibility thresholds than the commonly assumed uniform distortion model [16,19,20,21].
  • The proposed algorithm produces visually lossless images with minimum bitrates at the native resolutions inherently available in a JPEG 2000 codestream as well as at arbitrary intermediate resolutions.
  • The effectiveness of the proposed algorithm is demonstrated for remotely browsing images using JPIP, described in Part 9 of the JPEG 2000 standard [4]. Experimental results are presented for digital pathology images used for remote diagnosis [24] and for satellite images used for emergency relief. These high-resolution images with uncompressed file sizes of up to several gigabytes (GB) each are viewed at a variety of resolutions from remote locations, demonstrating significant savings in transmitted data compared to typical JPEG 2000/JPIP implementations.
This paper is organized as follows. Section 2 briefly reviews visually lossless encoding using visibility thresholds for JPEG 2000 as used in this work. Section 3 describes the change in visibility thresholds of a subband when the display resolution is changed. A method is then presented to apply several visibility thresholds in each subband using JPEG 2000. This method results in visually lossless rendering at minimal bitrates for each native JPEG 2000 resolution. Section 4 extends the proposed method to enable visually lossless encoding for arbitrary “intermediate” resolutions with minimum bitrates. In Section 5, the performance of the proposed algorithm is evaluated. Finally, Section 6 summarizes the work.

2. Visually Lossless JPEG 2000 Encoder

Distortion in JPEG 2000 results from differences between wavelet coefficient values at the encoder and the decoder that are generated by dead-zone quantization and mid-point reconstruction. This quantization distortion is then manifested as compression artifacts in the image, such as blurring or ringing artifacts, which are caused by applying the inverse wavelet transform. Compression artifacts have different magnitudes and patterns according to the subband in which the quantization distortion occurs. Thus, the visibility of quantization distortion varies from subband to subband. In [22,23], visibility thresholds (i.e., the maximum quantization step sizes at which quantization distortions remain invisible) were measured using an accurate model of the quantization distortion which occurs in JPEG 2000. A visually lossless JPEG 2000 encoder was proposed using these measured visibility thresholds. This section reviews this visually lossless JPEG 2000 encoder, which is the basis of the encoder proposed in subsequent sections of this paper.

2.1. Measurement of Visibility Thresholds

In this subsection, the measurement of the visibility thresholds employed in this work is summarized. Further details can be found in [22,23].
In what follows, we denote wavelet transform subbands using the notation LL, HL, LH, and HH, as is customary for separable transforms. In this notation, the first symbol denotes the filtering operation (low-pass or high-pass) carried out on rows, while the second symbol represents the filtering operation carried out on columns. For example, the LH subband is computed using low-pass filtering on the rows, and high-pass filtering on the columns.
Assuming that the wavelet coefficients in the HL, LH, and HH subbands have a Laplacian distribution and that the LL subband has wavelet coefficients of a uniform distribution [25], the distribution of quantization distortion for the HL, LH, and HH subbands can be modeled by the probability density function (PDF)
f ( d ) = 1 2 σ e - 2 | d | σ + 1 - p 1 Δ 0 | d | Δ 2 1 2 σ e - 2 | d | σ Δ 2 < | d | Δ 0 otherwise ,
where p 1 = - Δ Δ 1 1 2 σ 2 σ e - 2 | y | σ d y = 1 e - 2 Δ / σ . The parameters Δ and σ are the quantization step size and standard deviation of the wavelet coefficients, respectively. This model follows from the observation that wavelet coefficients in the dead-zone are quantized to 0 and coefficients outside the dead-zone yield quantization errors that are distributed approximately uniformly over ( Δ / 2 , Δ / 2 ). The distribution of quantization distortion for the LL subband can be modeled by the PDF
f ( d ) = 1 12 σ + 1 - p 2 Δ 0 | d | Δ 2 1 12 σ Δ 2 < | d | Δ 0 otherwise ,
where p 2 = Δ / 3 σ . These models are shown in Figure 1 for particular choices of quantization step size Δ and wavelet coefficient variance σ 2 .
The visibility thresholds for quantization distortion are obtained through psychophysical experiments with human subjects. A stimulus image is a red-green-blue (RGB) color image with a gray background ( Y = 128 , C b = 0 , and C r = 0 for 24-bit color images), obtained by applying the inverse wavelet transform and the inverse irreversible color transform (ICT) [1] to wavelet data containing quantization distortion. The quantization distortion is synthesized based on the JPEG 2000 quantization distortion model given above for a given coefficient variance σ 2 . The stimulus image is displayed together with a uniformly gray image (which does not contain a stimulus), and a human subject is asked to select the stimulus image. The quantization step size used to generate the quantization distortion is adaptively varied by the QUEST staircase procedure in the Psychophysics Toolbox v3.0.7 (available online at http://psychtoolbox.org). Through 32 iterations, the VT (the maximum quantization step size for which the stimulus remains invisible) is determined.
Unlike the conventional uniform quantization distortion model [12,13,16,19,20], indicated by the dashed line in Figure 1, the distribution of the quantization distortion is significantly affected by the variance of the wavelet coefficients. In general, an increase in coefficient variance leads to an increase in the visibility threshold. That is, larger distortions can go undetected when the coefficient variance is higher. For subband b = ( θ , k ) , where θ { L L , H L , L H , H H } is the orientation of the subband and k is the DWT level, the visibility threshold can be modeled as a function of coefficient variance σ b 2 by
t b ( σ b 2 ) = u b · σ b 2 + v b .
The parameters u b and v b were obtained from least-squares fits of thresholds measured via psycho-visual experiments for a variety of coefficient variances [23]. The resulting values for luminance are repeated here in Table 1. The chrominance thresholds were found to be insensitive to variance changes. That is, the corresponding values of u b were found to be significantly smaller than those for the luminance thresholds. Thus, constant thresholds (independent of coefficient variance) were reported in [23] and repeated here in Table 2 for ease of reference. Similarly, a fixed threshold value of t ( L L , 5 ) = 0 . 63 was reported for the LL subband of luminance components.

2.2. Visually Lossless JPEG 2000 Encoder

This section summarizes the visually lossless JPEG 2000 encoder of [22,23].
In JPEG 2000, the effective quantization step size of each codeblock is determined by the initial subband quantization step size and the number of coding passes included in the final codestream. To ensure that the effective quantization step sizes are less than the VTs, the following procedure is followed for all luminance subbands except (LL,5). First, the variance σ b , i 2 for the i-th codeblock B i in subband b is calculated. Then, t b , i ( σ b , i 2 ) for that codeblock is determined using (3). During bit-plane coding, the maximum absolute coefficient error in the codeblock is calculated after each coding pass z as
D ( z ) = max n B i | y [ n ] - y ˜ ( z ) [ n ] | ,
where y ˜ ( z ) [ n ] denotes the reconstructed value of y [ n ] using the quantization index q ˜ ( z ) [ n ] , which has been encoded only up to coding pass z. Coding is terminated when D ( z ) falls below the threshold t b , i ( σ b , i 2 ) . For the luminance (LL,5) and all chrominance subbands, the fixed VTs mentioned above are used as the initial subband quantization step size and all bit-planes are included in the codestream.
This JPEG 2000 Part 1 complaint visually lossless encoder can significantly reduce computational complexity since bit-plane coding is not carried out for coding passes that do not contribute to the final codestream, and provides visually lossless quality at competitive bitrates compared to numerically lossless or other visually lossless coding methods in the literature. To further reduce the bitrate, masking effects that take into account locally changing backgrounds can be applied to the threshold values, at the expense of increased computational complexity [23].
In the following sections, the main contribution of the present work is described. In particular, visually lossless encoding as described in [22,23] is extended to the multi-resolution case. For simplicity, masking effects are not considered.

3. Multi-Resolution Visually Lossless JPEG 2000

3.1. Multi-Resolution Visibility Thresholds

JPEG 2000, with K levels of dyadic tree-structured wavelet transform, inherently supports the synthesis of K + 1 different resolution images. As shown in Figure 2, the lowest resolution level R 0 corresponds to the lowest resolution image ( L L , K ) . The next lowest resolution level R 1 = { ( H L , K ) , ( L H , K ) , ( H H , K ) } together with ( L L , K ) can be used to render the next to lowest resolution image ( L L , K - 1 ) . Continuing in this fashion, resolution level R r together with the image ( L L , K - ( r - 1 ) ) can be used to synthesize the image ( L L , K - r ) , for 1 r K , with the full resolution image denoted by ( L L , 0 ) .
In what follows, it is assumed that images are always displayed so that each “image pixel” corresponds to one “monitor pixel.” Under this assumption, when displayed as an image, ( L L , K - r ) can be thought of as masquerading as a full resolution image. The subbands of resolution level R r can then be seen to play the role of the highest frequency subbands, normally played by R K . For example, in Figure 2, it can be seen that R 1 = { ( H L , 2 ) , ( L H , 2 ) , ( H H , 2 ) } contains the highest frequency subbands of the “image” ( L L , 1 ) . Similarly, the subbands of resolution level R r - 1 play the role normally played by those from resolution level R K - 1 , and so on. In general, when displaying image ( L L , K - r ) , resolution level R j behaves as resolution level R j + ( K - r ) , 0 < j r . The lowest resolution level R 0 = ( L L , K ) behaves as ( L L , r ) . Therefore, to have visually lossless quality of the displayed image ( L L , K - r ) , the visibility thresholds used for R j should be those normally used for R j + ( K - r ) . It then follows that when reduced resolution image ( L L , K - r ) is displayed, the visibility threshold for subband b = { θ , k } is given by
t ˜ b , r ( σ b 2 ) = t b ˜ ( σ b 2 ) for ( K - r ) < k K ,
where b ˜ = ( θ , k - ( K - r ) ) and t b ( σ b 2 ) is the threshold normally used for subband b to achieve visually lossless quality when the full resolution image ( L L , 0 ) is displayed. As usual, subbands with k ( K - r ) are discarded when forming ( L L , K - r ) . This can be considered as setting their thresholds to infinity.
For K = 5 , (3) and (5) together with Table 1 and Table 2 can then be used to determine appropriate visibility thresholds for all subbands except ( L L , 5 ) . As mentioned in the previous paragraph, and indicated by (5), ( L L , 5 ) plays the role of ( L L , r ) when the image ( L L , 5 - r ) is displayed. The work of [22,23] provided threshold values only for ( L L , 5 ) and not for ( L L , k ) , k < 5 . Therefore, in the work described herein, additional thresholds are provided for ( L L , k ) , 0 k 5 . For each such subband, thresholds were measured for a wide range of σ b 2 . Least squares fitting was then performed to obtain the parameters u ( L L , k ) and v ( L L , k ) for the model
t ( L L , k ) ( σ ( L L , k ) 2 ) = u ( L L , k ) log 10 σ ( L L , k ) 2 + v ( L L , k ) .
The resulting values for u ( L L , k ) and v ( L L , k ) are listed in Table 3. Substituting (6) into (5) yields the threshold value to be used for ( L L , K ) when displaying image ( L L , K - r ) . Specifically,
t ˜ ( L L , K ) , r ( σ ( L L , K ) 2 ) = u ( L L , r ) log 10 σ ( L L , K ) 2 + v ( L L , r ) .
Recall that a fixed threshold was employed for ( L L , 5 ) of the luminance component in [22,23]. In contrast, thresholds depending on codeblock variances are employed here for ( L L , k ) , 0 k 5 . This is due to the large number of codeblocks in these subbands exhibiting extreme variability in coefficient variances. Fixed thresholds still suffice for the chrominance components. Table 4 shows the chrominance threshold values t ( L L , k ) measured at an assumed typical variance of σ 2 = 150 .
Figure 3 illustrates the discussion above for K = 2 . When the full resolution image ( L L , 0 ) ( r = 2 ) is displayed, subband ( θ , k ) requires threshold t ( θ , k ) for visually lossless quality. However, when the one-level reduced resolution image ( L L , 1 ) ( r = 1 ) is displayed, the four subbands with k = 2 which previously needed thresholds t ( θ , 2 ) now require thresholds t ( θ , 1 ) . Similarly, when the lowest resolution image ( L L , 2 ) ( r = 0 ) is displayed, threshold t ( L L , 0 ) is applied.

3.2. Visually Lossless Quality Layers

From Table 1 through Table 4, it can be seen that visibility threshold values increase monotonically as the resolution level increases. This implies that the threshold for a given subband increases as the display resolution is decreased. Now, consider the case when an image is encoded with visibility thresholds designed for the full resolution image. Consider further forming a reduced resolution image in the usual manner by simply dropping the unneeded high frequency subbands. The (lower frequency) subbands still employed in image formation can be seen as being encoded using smaller thresholds than necessary, resulting in inefficiencies. Larger thresholds could be employed for these subbands resulting in smaller codestreams. In what follows, we describe how the (quality) layer functionality of JPEG 2000 can be used to apply multiple thresholds, each optimized for a different resolution.
JPEG 2000 layers are typically used to enable progressive transmission, which can increase the perceived responsiveness of remote image browsing. That is, when progressive transmission is used, the user often perceives that useful data are rendered faster for the same amount of data received. In JPEG 2000, each codeblock of each subband of each component contributes 0 or more consecutive coding passes to a layer. Beginning with the lowest (quality) layer Q 0 , image quality is progressively improved by the incremental contributions of subsequent layers. In typical JPEG 2000 encoder implementations, each layer is constructed to have minimum mean squared error (MSE) for a given bitrate, with the aid of post-compression rate-distortion optimization (PCRD-opt) [26]. More layers allow finer grained progressivity and thus more frequent rendering updates of displayed imagery, at the expense of a modest increase in codestream overhead. To promote spatial random access, wavelet data from each resolution level are partitioned into spatial regions known as precincts. Precinct sizes are arbitrary (user selectable) powers of 2 and can be made so large that no partitioning occurs, if desired. All coding passes from one layer that belong to codeblocks within one precinct of one resolution level (of one image component) are collected together in one JPEG 2000 packet. These packets are the fundamental units of a JPEG 2000 codestream.
In the work described here, layers are tied to resolutions so that layer Q 0 provides “just” visually lossless reconstruction of ( L L , K ) . The addition of layer Q 1 enables just visually lossless reconstruction of ( L L , K - 1 ) , and so on. More precisely, layer Q l is constructed so that when layers Q 0 through Q l are decoded, the maximum absolute quantization error, D ( z ) is just smaller than the visibility threshold t ˜ b , l ( σ b 2 ) for every codeblock in every resolution level R l , 0 l r . In this way, when image ( L L , K - r ) is decoded using only layers Q l , 0 l r , all relevant codeblocks are decoded at the quality corresponding to their appropriate visibility thresholds.
Figure 4 shows an example of quality layers generated for three display resolutions ( K = 2 ). The lowest resolution image ( L L , 2 ) needs only layer Q 0 for visually lossless reconstruction. At the next resolution, an additional layer Q 1 is decoded. That is, image ( L L , 1 ) is reconstructed using both Q 0 and Q 1 . At full resolution, the information from the final layer Q 2 is incorporated. It is worth reiterating that the JPEG 2000 codestream syntax requires that every codeblock contribute 0 or more coding passes to each layer. The fact that in the proposed scheme, each codeblock in R 1 and R 2 contribute 0 coding passes to Q 0 (and that R 2 contributes 0 coding passes to Q 1 ) is indicated in the figure. This results in a number of empty JPEG 2000 packets. The associated overhead is negligible, since each empty packet occupies only one byte in the codestream.
The advantages of the proposed scheme are clear from Figure 4. Specifically, when displaying ( L L , 2 ) , a straightforward treatment would discard ( H L , 2 ) through ( H H , 1 ) for considerable savings. However, it would retain unneeded portions ( Q 1 and Q 2 ) of ( L L , 2 ) . Similarly, when displaying ( L L , 1 ) , a straightforward treatment would discard ( H L , 1 ) through ( H H , 1 ) . However, it would still include unneeded data in the form of Q 2 for ( L L , 2 ) through ( H H , 2 ) . By discarding these unneeded data, the proposed scheme can achieve significant savings.

4. Visibility Thresholds for Downsampled Images

In the previous section, visibility thresholds and visually lossless encoding were discussed for the native resolutions inherently available in a JPEG 2000 codestream (all related by powers of 2). In this section, “intermediate” resolutions are considered. Such resolutions may be obtained via resampling of adjacent native resolution images. To obtain an image with a resolution between ( L L , K - r + 1 ) and ( L L , K - r ) , there are two possibilities: (1) upscaling from the lower resolution image ( L L , K - r + 1 ) ; or (2) downscaling from the higher resolution image ( L L , K - r ) . It is readily apparent that upscaling a decompressed version of ( L L , K - r + 1 ) will be no better than an upscaled (interpolated) version of an uncompressed version of ( L L , K - r + 1 ) . Visual inspection confirms that this approach does not achieve high quality rendering. Thus, in what follows, we consider only downsampling of the higher resolution image ( L L , K - r ) . A decompressed version of ( L L , K - r ) may be efficiently obtained by decoding Q l , 0 l r , as described in the previous section. However, even this method decodes more data than required for the rendering of imagery downsampled from ( L L , K - r ) . In what follows, the determination of the visibility thresholds for downscaled images is discussed.
Measurement of visibility thresholds for downscaled images is conducted in the fashion as described previously, but stimulus images are resampled by a rational factor of I / D before display. Resampling is performed in the following order: insertion of I zeros between each pair of consecutive samples, low-pass filtering, and decimation by D. In this experiment, visibility thresholds are measured for three intermediate resolutions below each native resolution. The resampling factors employed are 0 . 60 = 0 . 5 ( 1 . 2 ) , 0 . 72 = 0 . 5 ( 1 . 2 ) 2 , and 0 . 864 = 0 . 5 ( 1 . 2 ) 3 . These factors are applied as downscaling factors from the one-level higher resolution image ( L L , K - r ) for 0 r K . On the other hand, they can be thought of (conceptually) as resulting in successive 20% increases in resolution from ( L L , K - r + 1 ) . We begin by considering the subsampling of the full resolution image ( L L , 0 ) .
Table 5 lists measured visibility thresholds t ¯ b , n for K = 5 . The subscript n has been added to the notation to indicate the subsampling factor. Values, of n { 1 , 2 , 3 } correspond to subsampling factors of 0 . 5 ( 1 . 2 ) n , while n = 4 corresponds to a subsampling factor of 1.0 (no subsampling). As explained in Section II-A, quantization distortion varies with the variance of wavelet coefficients. However, the values in Table 5 were only measured for a fixed typical variance per subband. Psychovisual testing for all possible combinations of subbands, resolutions, and variances is prohibitive. Instead, the thresholds in Table 5 are adjusted to account for changes in variance as follows: in the case of chrominance components, the thresholds are not adjusted because, as before, chrominance threshold values are insensitive to coefficient variance (i.e., t b , n ( σ b 2 ) t ¯ b , n ). On the other hand, the luminance subbands are significantly affected by variance differences. For these subbands, the visibility threshold t b , 4 ( σ 2 ) corresponds to the non-subsampled case, and is given in (3) as before. Thresholds for n { 1 , 2 , 3 } are then obtained via
t b , n ( σ 2 ) = t ¯ b , n t b , 4 ( σ 2 ) t ¯ b , 4 .
From Equation (8) and Table 5, t b , n ( σ b 2 ) > t b , n + 1 ( σ b 2 ) for n { 1 , 2 , 3 } resulting in coarser quantization and smaller files.
Extension to downsampling starting from any native (power of 2) resolution results from updating Equation (5) to yield
t ˜ b , n , r ( σ b 2 ) = t b ˜ , n ( σ b 2 ) f o r ( K - r ) < k K .
Threshold t ˜ b , n , r ( σ b 2 ) is then the maximum quantization step size for subband b that provides visually lossless quality when image ( L L , K - r ) is downscaled by downsampling factor n, resulting in minimum file size for that downsampling. Table 6 contains the necessary values for t ¯ b , n for b ¯ = ( L L , k ) , 0 k < 5 .
The thresholds defined above are applied in JPEG 2000 by defining 4 ( K + 1 ) layers—one for each resolution to be rendered (each native resolution together with its three subsampled versions, n = 1 , 2 , 3 ). Layer Q l is then constructed such that codeblock i from subband b contains only all coding passes up to the first coding pass z that ensures the maximum quantization error D ( z ) falls just below the threshold t ˜ b , n ( l ) , r ( l ) ( σ b , i 2 ) , where r ( l ) = l / 4 and n ( l ) = l - 4 r ( l ) + 1 and σ b , i 2 is the variance of codeblock i from subband b . Decoding from Q 0 to Q l then ensures visually lossless reconstruction of image ( L L , K - r ( l ) ) when downscaled by downsampling factor n ( l ) .
A visually lossless image with completely arbitrary resolution scale p ( 0 , 1 ] with respect to the full resolution image ( L L , 0 ) can be obtained by calculating r = K - - log 2 p , n = ( log 2 p + - log 2 p + 1 ) / log 2 1 . 2 , and l = n + 4 r - 1 . The decoded version of image ( L L , K - r ) , using layers Q 0 through Q l , then provides enough quality so that appropriate resampling yields a visually lossless image with resolution scale p.

5. Experimental Results

5.1. Multi-Resolution Visually Lossless Coding

The proposed multi-resolution visually lossless coding scheme was implemented in Kakadu v6.4 (Kakadu Software, Sydney, Australia, http://www.kakadusoftware.com). Experimental results are presented for seven digital pathology images and eight satellite images. All of the images are 24-bit color high resolution images ranging in size from 527 MB ( 13165 × 14000 ) to 3.23 GB ( 39912 × 29032 ). Each image is identified with pathology or satellite together with an index, e.g., pathology 1 or satellite 3. Recent technological developments in digital pathology allow rapid processing of pathology slides using array microscopes [24]. The resulting high-resolution images (referred to as virtual slides) can then be reviewed by a pathologist either locally or remotely over a telecommunications network. Due to the high resolution of the imaging process, these images can easily occupy several GBytes. Thus, remote examination by the pathologist requires efficient methods for transmission and display of images at different resolutions and spatial extents at the reviewing workstation. The satellite images employed here show various locations on Earth before and after natural disasters. The images were captured by the GeoEye-1 satellite, at 0.5 meter resolution from 680 km in space and were provided for the use of relief organizations. These images are also so large that fast rendering and significant bandwidth savings are essential for efficient remote image browsing.
In this work, “reference images” corresponding to reduced native resolution images ( L L , 5 - r ) , r = 0 , 1 , . . . , 4 were created using the 9/7 DWT without quantization or coding. Reference images for intermediate resolutions were obtained by downscaling the next (higher) native resolution reference image. In what follows, the statement that a decompressed reduced resolution image is visually lossless means that it is visually indistinguishable from its corresponding reference image.
To evaluate the compression performance of the proposed method, each image was encoded using three different methods. The first method is referred to here as the six-layer method. As the name suggests, codestreams from this method employ six layers to apply the appropriate visually lossless thresholds for each of six native resolution images ( L L , 5 - r ) , r = 0 , 1 , . . . , 5 ( K = 5 ) . The second method, referred to as the 24-layer method, uses a total of 24 layers to provide visually lossless quality at each of the six native resolutions, plus three intermediate resolutions below each native resolution. The third method, used as a benchmark, employs the method from [22] to yield a visually lossless image optimized for display only at full resolution. The codestream for this method contains a single layer, so this benchmark is referred to as the single-layer method. To facilitate spatial random access, all images were encoded using the CPRL (component-precinct-resolution-layer) progression order with precincts of size 128 × 128 at each resolution level.
Figure 5 compares the number of bytes that must be decoded (transmitted) for each of the three coding methods to have visually lossless quality at various resolutions. Results are presented for one image of each type. Graphs for other images are similar. The number of bytes for the single-layer, six-layer, and 24-layer methods at each resolution are denoted by crosses, rectangles, and circles, respectively. As expected, the curves for the single-layer method generally lie above those of the six-layer method, which, in turn, generally lie above those of the 24-layer method. It is worth noting that the vertical axis employs a logarithmic scale, and that gains in compression ratio are significant for most resolutions.
Table 7 lists bitrates obtained (in bits-per-pixel with respect to the dimensions of the full-resolution images) averaged over all 15 test images. From this table, it can be seen that the six-layer method results in 39.3%, 50.0%, 48.1%, 42.1%, and 31.0% smaller bitrate compared to the single-layer method for reduced resolution images ( L L , 5 - r ) , r = 0 , 1 , 2 , 3 , and 4, respectively. In turn, for the downsampled images ( L L , 5 - r ) , n = 1 , r = 0 , 1 , 2 , 3 , 4 , and 5, the 24-layer method provides 25.0%, 30.3%, 35.5%, 39.1%, 39.1%, and 36.7% savings in bitrate, respectively compared to the six-layer method. These significant gains are achieved by discarding unneeded codestream data in the relevant subbands in a precise fashion, while maintaining visually lossless quality in all cases. Specifically, the six-layer case can discard data in increments of one layer out of six, while the 24-layer method can discard data in increments of one layer out of 24. In contrast, the single-layer method must read all data in the relevant subbands.
Figure 6 shows crops of the satellite 2 image reconstructed at resolution p = 0 . 09 for the three coding methods. Each of these images is downscaled from a version of ( L L , 3 ) ( p = 0 . 125 ) . Specifically, for the single-layer method, the image is downscaled from all ( L L , 3 ) data which amounts to 5.30 bpp, relative to the reduced resolution dimensions. For the proposed methods, the image is downscaled from three of six layers and 10 of 24 layers of ( L L , 3 ) data, which amounts to 2.98 bpp and 2.09 bpp, respectively. Although the proposed methods offer significantly lower bitrates, all three resulting images have the same visual quality. That is, they are all indistinguishable from the reference image.
Although this multi-layer method provides significant gains for most resolutions, there exist a few (negligible) losses for some resolutions. Specifically, the six-layer case is slightly worse than the single-layer case for ( L L , 0 ) as well as for the three intermediate resolutions immediately below ( L L , 0 ) . The average penalty in this case is 0.72%. Similarly, the 24-layer case is slightly worse than the six-layer case at each of the six native resolutions. The average penalty for ( L L , 5 - r ) , r = 0 , 1 , 2 , . . . , 5 is 0.19%, 0.48%, 0.88%, 1.30%, 1.89%, and 2.23%, respectively. These minor drops in performance are due to the codestream syntax overhead associated with including more layers.
As mentioned previously, the layer functionality of JPEG 2000 enables quality scalability. As detailed in the previous paragraph, for certain isolated resolutions, the single-layer method provides slightly higher compression efficiency as compared to the six-layer and 24-layer methods. However, it provides no quality scalability and therefore no progressive transmission capability. To circumvent this limitation, layers could be added to the so called single-layer method. To this end, codestreams from the single-layer method were partitioned into six layers. The first five layers were constructed via the arbitrary selection of five rate-distortion slope thresholds, as normally allowed by the Kakadu implementation. The six layers together yield exactly the same decompressed image as the single-layer method. In this way, the “progressivity” is roughly the same as the six-layer method, but visually lossless decoding of reduced resolution images is not guaranteed for anything short of decoding all layers for the relevant subbands. Table 8 compares the bitrates for decoding ( L L , 0 ) under the proposed six-layer visually lossless coding method vs. the one-layer method with added layers for the pathology 3 and satellite 6 images. As can be seen from the table, the results are nearly identical. Results for other images as well as for 24-layers are similar. Thus, the overhead associated with the proposed method is no more than that needed to facilitate quality scalability/progressivity.

5.2. Validation Experiments

To verify that the images encoded with the proposed scheme are visually lossless at each resolution, a three-alternative forced-choice (3AFC) method was used, as in [23]. Two reference images and one compressed image were displayed side by side for an unlimited amount of time, and the subject was asked to choose the image that looked different. The position of the compressed image was chosen randomly. In the validation experiments, all compressed images were reconstructed from codestreams encoded with the 24-layer method. Validating the 24-layer method suffices since this method uses the same or larger threshold values than those of the six-layer or single-layer method at each resolution. In other words, the 24-layer method never uses less aggressive quantization than the other two methods. All 15 images were used in the validation study. For each image, compressed images and reference images were generated at two native resolutions, ( L L , 5 ) ( p = 0 . 03125 ) and ( L L , 2 ) ( p = 0 . 25 ), as well as two intermediate resolutions corresponding to ( L L , 3 ) , n = 2 ( p = 0 . 09 ) and ( L L , 0 ) , n = 2 ( p = 0 . 72 ). Full resolution images ( p = 1 . 0 ) were validated in [23]. These five resolutions form a representative set spread over the range of possible values. After decompression, all images were cropped to 512 × 512 at a random image location so that three copies fit side by side on a Dell U2410 LCD monitor (Dell, Round Rock, TX, USA) with display size 1920 × 1200 . The validation experiment consisted of four sessions—one for each resolution tested. In each session, all 15 images were viewed five times at the same resolution (75 trials for each subject) in random order. Subjects were allowed to rest between sessions. Twenty subjects, who are familiar with image compression and have normal or corrected-to-normal vision, participated in the experiment. Each subject gave a total of 300 responses over the four sessions. The validation was performed under the same viewing conditions as in [23] (Dell U2410 LCD monitor, ambient light, a viewing distance of 60 cm). During the experiment, no feedback was provided on the correctness of choices.
If the compressed image is indistinguishable from the reference image, the correct response should be obtained with a frequency of 1/3. Table 9 shows the statistics obtained and t-test results with a test value of 1/3. It can be seen that the hypothesis that the responses were randomly chosen could not be rejected at the 5% significance level for each of the four sessions. Based on these results, it is claimed that the proposed coding method provides visually lossless quality at every tested resolution.
In addition to the testing performed by human observers, the structural similarity (SSIM) [27] was computed for each image, as shown in Table 10. The SSIM values ranged from a minimum of 0.9730 to a maximum of 0.9982, with an average value of 0.9890. These uniformly high values provide additional evidence in favor of the high quality of the images produced by the proposed encoder.
It is worth noting that for medical images, such as the pathology images employed as part of the test set herein, visually lossless compression may not be the most relevant goal. For instance, “diagnostically lossless” compression may be more interesting for this type of imagery. Indeed, our ongoing work is focused on the maximization of compression efficiency while maintaining diagnostic accuracy.

5.3. Performance Evaluation Using JPIP

JPIP is a connection-oriented network communication protocol that facilitates efficient transmission of images using the characteristics of scalable JPEG 2000 codestreams. A user can interactively browse spatial regions of interest, at desired resolutions, by retrieving only the corresponding minimum required portion of the codestream.
Figure 7 shows a block diagram of a JPIP remote image browsing system. First, the client requests the spatial region, resolution, number of layers, and image components of interest using a simple descriptive syntax to the server. In response to the client request, the server accesses the relevant packets from the JPEG 2000 codestream and sends them to the client. The client decodes the received packets and renders the image. Through a graphic user interface (GUI) on the client-side, a user can request different regions, resolutions, components, and number of layers at any time. To minimize the number of transmitted packets and maximize the responsiveness of interactive image browsing, the server assumes that the client employs a cache to hold data from previous requests, and maintains a cache model to keep track of the client cache. If a request is found to contain cached data, the server does not re-send that portion of the data.
The experimental results described below were obtained using the same codestreams employed in the experiments described in Section 5.1 above. As described there, these codestreams were created using five levels of 9/7 DWT, precincts of size 128 × 128 at each resolution level, and the CPRL progression order. All codestreams were JPEG 2000 Part 1 compliant. All JPIP experiments were conducted with an unmodified version of kdu_server (from Kakadu v6.4). The kdu_show client was adapted to specifically request only the number of quality layers commensurate with the codestream construction (single-layer, six-layer, 24-layer) and the resolution level currently being displayed by the client. All client/server communications were JPIP compliant.
Table 11 reports the number of bytes transmitted via JPIP for four different images and three different visually lossless coding methods. The dimensions (size) of each image are included in the table as well. The rows in the table correspond to different images, while the three rightmost columns correspond to the different coding methods. The same sequence of browsing operations was issued by the JPIP client for each image and each coding method. In particular, for each codestream under test, the ( L L , 5 ) image ( p = 0 . 03125 ) was first requested. This resulted in an overview of the entire image being displayed in a window of size 0.03125 times the dimensions of the image under test (e.g., 794 × 822 for pathology 3.) This initial window size was maintained throughout the browsing session of the image under test. Following the initial request, four different locations in the image were requested with progressively higher resolution scales ( p = 0 . 15 , 0 . 36 , 0 . 72 , 1 . 0 ). Three pan operations were included after each new location request. As mentioned above, only the appropriate layers were transmitted for the six-layer and 24-layer methods. As expected, in each case, the number of bytes required by the six-layer method was significantly less than that required for the single-layer method. In turn, the 24-layer method resulted in significant savings over the six-layer method. Averaged over the four images, the number of transmitted bytes for the single-layer method was 9180.9 KB, while the number of transmitted bytes for the six-layer and 24-layer methods was 6279.2 KB and 4679.2 KB, representing a decrease of 31.61% and 49.03% over the single-layer method, respectively. It is clear from these results that the codestreams encoded by the six-layer and 24-layer methods require considerably less bandwidth than a codestream optimized only for full resolution display.
As mentioned in Section 3, layers are essential to effective browsing of remote images. Specifically, no quality progressivity is possible for single-layer codestreams. Figure 8 demonstrates this via example images obtained via JPIP browsing of a single-layer file vs. a 24-layer file for roughly the same number of transmitted bytes. While neither image is (yet) visually lossless for the number of bytes transmitted up to the moment of rendering in the figure, the advantage of progressive transmission is readily apparent. Using the image from the 24-layer file, the user may be able to make their next request, thus preempting the current request, without waiting for the rest of the data to be transmitted. This can further reduce bandwidth, but was not used to obtain the values in Table 11.

6. Conclusions

This paper presents a multi-resolution visually lossless image coding method using JPEG 2000, which uses visibility threshold values measured for downsampled JPEG 2000 quantization distortion. This method is implemented via JPEG 2000 layers. Each layer is constructed such that the maximum quantization error in each codeblock is less than the appropriate visibility threshold at the relevant display resolution. The resulting JPEG 2000 Part 1 compliant coding method allows for visually lossless decoding at resolutions natively supported by the wavelet transform as well as arbitrary intermediate resolutions, using only a fraction of the full-resolution codestream. In particular, when decoding the full field of view of a very large image at various reduced resolutions, the proposed six-layer method reduces the amount of data that must be accessed and decompressed by 30 to 50 percent compared to that required by single-layer visually lossless compression. This is true despite the fact that the single-layer method does not access nor decode data from unnecessary high frequency subbands. The 24-layer method provides additional improvements over the six-layer method ranging from 25 to 40 percent. This, in turn, brings the gains of the 24-layer method over the single-layer method into the range of 55 to 65 percent. Stated another way, the gain in the amount of data access and decompressed (effective compression ratio) is improved by a factor of more than two. These gains are born out in a remote image browsing experiment using digital pathology images and satellite images. In this experiment, JPIP is used to browse limited fields of view at different resolutions while employing zoom and pan. In this scenario, the proposed method exhibits gains similar to those described above, with no adverse effects on visual image quality.

Acknowledgments

This work was supported in part by the National Institutes of Health (NIH)/National Cancer Institute (NCI) under grant 1U01CA198945.

Author Contributions

The authors contributed equally to this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Taubman, D.; Marcellin, M. JPEG2000: Image Compression Fundamentals, Standards, and Practice; Kluwer Academic Publishers: Boston, MA, USA, 2002. [Google Scholar]
  2. Taubman, D.; Marcellin, M. JPEG2000: Standard for interactive imaging. Proc. IEEE 2002, 90, 1336–1357. [Google Scholar] [CrossRef]
  3. Taubman, D.S.; Prandolini, R. Architecture, Philosophy and performance of JPIP: Internet protocol standard for JPEG 2000. Proc. SPIE 2003, 5150, 791–805. [Google Scholar]
  4. ISO/IEC15444. Information Technology—JPEG 2000 Image Coding System—Part 9: Interactive tools, APIs and Protocols; International Organization for Standardization (ISO): Geneva, Switzerland, 2003. [Google Scholar]
  5. Blakemore, C.; Campbell, F. On the existence of neurones in the human visual system selectively sensitive to the orientation and size of retinal images. J. Physiol. 1969, 203, 237–260. [Google Scholar] [CrossRef] [PubMed]
  6. Nadenau, M.; Reichel, J. Opponent color, human vision and wavelets for image compression. In Proceedings of the Seventh Color Imaging Conference, Scottsdale, AZ, USA, 16–19 November 1999; pp. 237–242.
  7. Taylor, C.; Pizlo, Z.; Allebach, J.P.; Bouman, C. Image quality assessment with a Gabor pyramid model of the Human Visual System. Proc. SPIE 1997, 3016, 8–69. [Google Scholar]
  8. Watson, A.B. The cortex transform: Rapid computation of simulated neural images. Comput. Vis. Graph. Image Process. 1987, 39, 311–327. [Google Scholar] [CrossRef]
  9. Lubin, J. The use of psychophysical data and models in the analysis of display system performance. In Digital Images and Human Vision; Watson, A.B., Ed.; The MIT Press: Cambridge, MA, USA, 1993; pp. 163–178. [Google Scholar]
  10. Daly, S. The visible differences predictor: An algorithm for the assessment of image fidelity. In Digital Images and Human Vision; Watson, A.B., Ed.; The MIT Press: Cambridge, MA, USA, 1993; pp. 179–206. [Google Scholar]
  11. Watson, A.B. DCT quantization matrices visually optimized for individual images. Proc. SPIE 1993, 1913, 202–216. [Google Scholar]
  12. Watson, A.B.; Yang, G.Y.; Solomon, J.A. Visibility of wavelet quantization noise. IEEE Trans. Image Process. 1997, 6, 1164–1175. [Google Scholar] [CrossRef] [PubMed]
  13. Chandler, D.M.; Hemami, S.S. Effects of natural images on the detectability of simple and compound wavelet subband quantization distortions. J. Opt. Soc. Am. A 2003, 20, 1164–1180. [Google Scholar] [CrossRef]
  14. Bae, S.H.; Pappas, T.N.; Juang, B.H. Subjective evaluation of spatial resolution and quantization noise tradeoffs. IEEE Trans. Image Process. 2009, 18, 495–508. [Google Scholar] [PubMed]
  15. Li, J.; Chaddha, N.; Gray, R.M. Multiresolution Tree Structured Vector Quantization. In Proceedings of the 1996 Conference Record of the Thirtieth Asilomar Conference on Date of Conference, Pacific Grove, CA, USA, 3–6 November 1996; Volume 2, pp. 922–925.
  16. Hsiang, S.; Woods, J.W. Highly Scalable and Perceptually Tuned Embedded Subband/Wavelet Image Coding. Proc. SPIE 2002, 4671, 1153–1164. [Google Scholar]
  17. Oh, H.; Bilgin, A.; Marcellin, M.W. Multi-resolution visually lossless image compression using JPEG2000. In Proceedings of the IEEE International Conference on Image Processing, Hong Kong, 26–29 September 2010; pp. 2581–2584.
  18. Oh, H.; Bilgin, A.; Marcellin, M.W. Visually lossless JPEG2000 at Fractional Resolutions. In Proceedings of the 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 309–312.
  19. Liu, Z.; Karam, L.J.; Watson, A.B. JPEG2000 encoding with perceptual distortion control. IEEE Trans. Image Process. 2006, 15, 1763–1778. [Google Scholar] [PubMed]
  20. Chandler, D.M.; Dykes, N.L.; Hemami, S.S. Visually lossless compression of digitized radiographs based on contrast sensitivity and visual masking. Proc. SPIE 2005, 5749, 359–372. [Google Scholar]
  21. Wu, D.; Tan, D.M.; Baird, M.; DeCampo, J.; White, C.; Wu, H.R. Perceptually lossless medical image coding. IEEE Trans. Image Process. 2006, 25, 335–344. [Google Scholar] [CrossRef] [PubMed]
  22. Oh, H.; Bilgin, A.; Marcellin, M.W. Visibility thresholds for quantization distortion in JPEG 2000. In Proceedings of the International Workshop on Quality of Multimedia Experience (QoMEx), San Diego, CA, USA, 29–31 July 2009; pp. 228–232.
  23. Oh, H.; Bilgin, A.; Marcellin, M.W. Visually lossless encoding for JPEG 2000. IEEE Trans. Image Process. 2013, 22, 189–201. [Google Scholar] [PubMed]
  24. Weinstein, R.S.; Descour, M.R.; Chen, L.; Barker, G.; Scott, K.M.; Richter, L.; Krupinski, E.A.; Bhattacharyya, A.K.; Davis, J.R.; Graham, A.R. An array microscope for ultrarapid virtual slide processing and telepathology. Design, fabrication, and validation study. Hum. Pathol. 2004, 35, 1303–1314. [Google Scholar] [CrossRef] [PubMed]
  25. Mallat, S.G. A Theory for Multiresolution Signal Decomposition—The Wavelet Representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef]
  26. Taubman, D.S. High performance scalable image compression with EBCOT. IEEE Trans. Image Process. 2000, 9, 1158–1170. [Google Scholar] [CrossRef] [PubMed]
  27. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Probability density functions of: (a) the quantization distortion in HL, LH, and HH subbands ( σ 2 = 50 , Δ = 5 ); and (b) the quantization distortion in the LL subband ( σ 2 = 2000 , Δ = 5 ). The dashed lines represent the commonly assumed uniform distribution.
Figure 1. Probability density functions of: (a) the quantization distortion in HL, LH, and HH subbands ( σ 2 = 50 , Δ = 5 ); and (b) the quantization distortion in the LL subband ( σ 2 = 2000 , Δ = 5 ). The dashed lines represent the commonly assumed uniform distribution.
Information 07 00045 g001
Figure 2. Resolution levels within a dyadic tree-structured subband decomposition with K = 2 levels.
Figure 2. Resolution levels within a dyadic tree-structured subband decomposition with K = 2 levels.
Information 07 00045 g002
Figure 3. Visibility thresholds at three display resolutions ( K = 2 ).
Figure 3. Visibility thresholds at three display resolutions ( K = 2 ).
Information 07 00045 g003
Figure 4. Quality layers for three display resolutions ( K = 2 ).
Figure 4. Quality layers for three display resolutions ( K = 2 ).
Information 07 00045 g004
Figure 5. Number of bytes required for decompression by the three described coding methods for the two images (a) pathology 1 and (b) satellite 1. The single-layer, six-layer, and 24-layer methods at each resolution are denoted by crosses, rectangles, and circles, respectively.
Figure 5. Number of bytes required for decompression by the three described coding methods for the two images (a) pathology 1 and (b) satellite 1. The single-layer, six-layer, and 24-layer methods at each resolution are denoted by crosses, rectangles, and circles, respectively.
Information 07 00045 g005
Figure 6. Satellite 2 image rendered at resolution p = 0 . 09 . (a) reference; (b) single-layer method; (c) six-layer method; and (d) 24-layer method. The images are cropped to 312 × 312 after rendering to avoid rescaling during display. The images should be viewed with PDF viewer scaling set to 100%. Satellite image courtesy of GeoEye (DigitalGlobe, Westminster, CO, USA).
Figure 6. Satellite 2 image rendered at resolution p = 0 . 09 . (a) reference; (b) single-layer method; (c) six-layer method; and (d) 24-layer method. The images are cropped to 312 × 312 after rendering to avoid rescaling during display. The images should be viewed with PDF viewer scaling set to 100%. Satellite image courtesy of GeoEye (DigitalGlobe, Westminster, CO, USA).
Information 07 00045 g006aInformation 07 00045 g006b
Figure 7. Client-server interaction in a JPEG 2000 Interactive Protocol (JPIP) remote image browsing system.
Figure 7. Client-server interaction in a JPEG 2000 Interactive Protocol (JPIP) remote image browsing system.
Information 07 00045 g007
Figure 8. Images rendered while the image data are being retrieved via a JPIP client. (a) single-layer method at 131.2 KB and (b) 24-layer method at 130.4 KB.
Figure 8. Images rendered while the image data are being retrieved via a JPIP client. (a) single-layer method at 131.2 KB and (b) 24-layer method at 130.4 KB.
Information 07 00045 g008
Table 1. Linear parameters u b and v b for luminance components.
Table 1. Linear parameters u b and v b for luminance components.
Subbandubvb
(HH,1) 105 . 67 × 10 - 4 4.85
(HL/LH,1) 46 . 03 × 10 - 4 1.98
(HH,2) 19 . 94 × 10 - 4 0.92
(HL/LH,2) 13 . 84 × 10 - 4 0.64
(HH,3) 11 . 04 × 10 - 4 0.51
(HL/LH,3) 10 . 83 × 10 - 4 0.50
(HH,4) 10 . 16 × 10 - 4 0.47
(HL/LH,4) 7 . 75 × 10 - 4 0.36
(HH,5) 7 . 91 × 10 - 4 0.36
(HL/LH,5) 7 . 16 × 10 - 4 0.33
u b and v b are the slope and y-intercept, respectively, for use in Equation (3).
Table 2. Visibility thresholds for the chrominance components Cb and Cr.
Table 2. Visibility thresholds for the chrominance components Cb and Cr.
SubbandCbCr
(HH,1)24.4015.60
(HL/LH,1)13.906.40
(HH,2)14.917.35
(HL/LH,2)6.392.55
(HH,3)10.892.65
(HL/LH,3)4.031.23
(HH,4)4.471.27
(HL/LH,4)2.970.72
(HH,5)1.100.65
(HL/LH,5)1.050.60
(LL, 5)1.190.66
Table 3. Parameters for t ( L L , k ) for luminance subband ( L L , k ) .
Table 3. Parameters for t ( L L , k ) for luminance subband ( L L , k ) .
ku(LL,K)v(LL,K)
0 0 . 2311 2.0170
1 0 . 3081 0.8095
2 0 . 0802 0.8270
3 0 . 1032 0.5893
4 0 . 0309 0.6848
5 0 . 0128 0.5923
Table 4. Thresholds t ( L L , k ) for chrominance subband ( L L , k ) .
Table 4. Thresholds t ( L L , k ) for chrominance subband ( L L , k ) .
kCbCr
04.734.50
13.783.40
22.452.12
32.311.85
41.601.00
51.190.66
Table 5. Visibility thresholds t ¯ b , n for downscaling of the full resolution image ( L L , 0 ) for typical subband variance values—Luminance (Y): 2000 for LL; 50 for HL/LH and HH. Chrominance (Cb and Cr): 150 for LL; 5 for HL/LH and HH.
Table 5. Visibility thresholds t ¯ b , n for downscaling of the full resolution image ( L L , 0 ) for typical subband variance values—Luminance (Y): 2000 for LL; 50 for HL/LH and HH. Chrominance (Cb and Cr): 150 for LL; 5 for HL/LH and HH.
ComponentSubband60% (n = 1)72% (n = 2)86.4% (n = 3)100% (n = 4)
Y ( L L , 5 ) 0.890.830.780.63
( H L / L H , 5 ) 0.700.590.570.37
( H L / L H , 4 ) 0.720.650.590.40
( H L / L H , 3 ) 1.000.870.790.55
( H L / L H , 2 ) 1.881.651.350.71
( H L / L H , 1 ) 5.394.903.152.21
( H H , 5 ) 0.70.630.600.40
( H H , 4 ) 0.750.680.650.52
( H H , 3 ) 1.130.980.900.57
( H H , 2 ) 4.202.952.051.02
( H H , 1 ) 17.4712.509.505.38
Cb ( L L , 5 ) 1.501.321.211.19
( H L / L H , 5 ) 1.321.201.121.05
( H L / L H , 4 ) 3.523.183.162.97
( H L / L H , 3 ) 6.245.004.354.03
( H L / L H , 2 ) 10.138.417.636.39
( H L / L H , 1 ) 55.6041.7027.8013.90
( H H , 5 ) 2.201.541.451.10
( H H , 4 ) 6.165.014.504.47
( H H , 3 ) 14.7014.3011.5910.89
( H H , 2 ) 19.4918.0515.1014.91
( H H , 1 ) 97.6073.2048.8024.40
Cr ( L L , 5 ) 1.050.980.950.66
( H L / L H , 5 ) 1.081.011.050.60
( H L / L H , 4 ) 1.271.231.100.72
( H L / L H , 3 ) 1.981.581.451.23
( H L / L H , 2 ) 4.783.502.712.55
( H L / L H , 1 ) 25.6019.2012.806.40
( H H , 5 ) 1.120.950.870.65
( H H , 4 ) 2.451.801.361.27
( H H , 3 ) 6.443.763.112.65
( H H , 2 ) 14.4712.5111.367.35
( H H , 1 ) 62.4046.8031.2015.60
Table 6. Visibility thresholds t ¯ b , n for the LL subband for typical subband variance values—Luminance (Y): 2000. Chrominance (Cb and Cr): 150.
Table 6. Visibility thresholds t ¯ b , n for the LL subband for typical subband variance values—Luminance (Y): 2000. Chrominance (Cb and Cr): 150.
ComponentSubband60% (n = 1)72% (n = 2)86.4% (n = 3)100% (n = 4)
Y ( L L , 4 ) 0.920.890.850.79
( L L , 3 ) 1.060.980.970.93
( L L , 2 ) 1.351.221.171.09
( L L , 1 ) 2.502.251.881.83
( L L , 0 ) 7.334.653.852.78
Cb ( L L , 4 ) 2.151.891.721.60
( L L , 3 ) 2.602.502.452.31
( L L , 2 ) 3.603.132.752.45
( L L , 1 ) 4.554.104.053.78
( L L , 0 ) 8.807.506.304.73
Cr ( L L , 4 ) 1.511.401.381.00
( L L , 3 ) 2.082.001.951.85
( L L , 2 ) 3.272.702.452.12
( L L , 1 ) 4.43.903.603.40
( L L , 0 ) 5.255.104.904.50
Table 7. Average bits-per-pixel (bpp) decoded with respect to the full-resolution dimensions.
Table 7. Average bits-per-pixel (bpp) decoded with respect to the full-resolution dimensions.
ResolutionSingle-Layer6-Layer24-Layer
( L L , 5 ) , n = 1 0.009830.005970.00448
( L L , 5 ) , n = 2 0.009830.005970.00529
( L L , 5 ) , n = 3 0.009830.005970.00529
( L L , 5 ) , n = 4 0.009830.005970.00598
( L L , 4 ) , n = 1 0.027370.013700.00955
( L L , 4 ) , n = 2 0.027370.013700.01078
( L L , 4 ) , n = 3 0.027370.013700.01156
( L L , 4 ) , n = 4 0.027370.013700.01376
( L L , 3 ) , n = 1 0.084360.043770.02823
( L L , 3 ) , n = 2 0.084360.043770.03205
( L L , 3 ) , n = 3 0.084360.043770.03535
( L L , 3 ) , n = 4 0.084360.043770.04416
( L L , 2 ) , n = 1 0.268340.153650.09468
( L L , 2 ) , n = 2 0.268340.153650.10177
( L L , 2 ) , n = 3 0.268340.153650.13000
( L L , 2 ) , n = 4 0.268340.153650.15566
( L L , 1 ) , n = 1 0.847070.584760.35592
( L L , 1 ) , n = 2 0.847070.584760.40256
( L L , 1 ) , n = 3 0.847070.584760.43872
( L L , 1 ) , n = 4 0.847070.584760.59581
( L L , 0 ) , n = 1 1.684681.696891.07471
( L L , 0 ) , n = 2 1.684681.696891.12744
( L L , 0 ) , n = 3 1.684681.696891.41057
( L L , 0 ) , n = 4 1.684681.696891.73476
Table 8. Bitrates (bpp) of the proposed six-layer method vs. the single-layer method with added layers.
Table 8. Bitrates (bpp) of the proposed six-layer method vs. the single-layer method with added layers.
ImageProposed 6-LayerSingle-Layer
pathology 31.14291.1441
satellite 62.14032.1425
Table 9. t-test results (test value = 1 / 3 , N = 15 ).
Table 9. t-test results (test value = 1 / 3 , N = 15 ).
Session (p)MeanStandard DeviationStandard Error Meant-ScoreSignificance (2-tailed)95% Confidence Interval
LowerUpper
0.031250.33070.056880.01469−0.1790.8600.29920.3622
0.090.33470.047790.012340.1110.9130.30820.3611
0.250.35400.039960.010322.0060.0650.33190.3761
0.720.34530.030210.007801.5430.1450.32860.3621
Table 10. SSIM values for the test images.
Table 10. SSIM values for the test images.
Imagep = 0.03125 p = 0.09 p = 0.25 p = 0.72
Pathology 10.99750.98970.99370.9752
Pathology 20.99640.98860.99430.9912
Pathology 30.99750.99330.99510.9907
Satellite 10.99630.99070.99140.9907
Satellite 20.99730.97370.98950.9730
Satellite 30.99820.97800.99110.9743
Satellite 40.99800.97890.99130.9756
Table 11. Transmitted bytes for the three visually lossless coding methods while remotely browsing compressed images.
Table 11. Transmitted bytes for the three visually lossless coding methods while remotely browsing compressed images.
ImageDimensionsSingle-Layer6-Layer24-Layer
pathology 1 22 , 040 × 21 , 320 7966.6 KB5691.4 KB4222.3 KB
pathology 3 25 , 408 × 26 , 288 7917.0 KB5133.1 KB3855.9 KB
pathology 5 39 , 912 × 29 , 032 10,699.8 KB7199.5 KB5190.9 KB
satellite 2 20 , 480 × 31 , 744 10,140.3 KB7092.6 KB5447.5 KB

Share and Cite

MDPI and ACS Style

Oh, H.; Bilgin, A.; Marcellin, M. Visually Lossless JPEG 2000 for Remote Image Browsing. Information 2016, 7, 45. https://doi.org/10.3390/info7030045

AMA Style

Oh H, Bilgin A, Marcellin M. Visually Lossless JPEG 2000 for Remote Image Browsing. Information. 2016; 7(3):45. https://doi.org/10.3390/info7030045

Chicago/Turabian Style

Oh, Han, Ali Bilgin, and Michael Marcellin. 2016. "Visually Lossless JPEG 2000 for Remote Image Browsing" Information 7, no. 3: 45. https://doi.org/10.3390/info7030045

APA Style

Oh, H., Bilgin, A., & Marcellin, M. (2016). Visually Lossless JPEG 2000 for Remote Image Browsing. Information, 7(3), 45. https://doi.org/10.3390/info7030045

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop