Next Article in Journal
Hierarchical Fusion Using Subsets of Multi-Features for Historical Arabic Manuscript Dating
Next Article in Special Issue
Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content
Previous Article in Journal
Scanning Hyperspectral Imaging for In Situ Biogeochemical Analysis of Lake Sediment Cores: Review of Recent Developments
Previous Article in Special Issue
Texture Management for Glossy Objects Using Tone Mapping
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Glossiness Index of Objects in Halftone Color Images Based on Structure and Appearance Distortion

1
Graduate School of Science and Engineering, Chiba University, Yayoi-cho 1-33, Inage-ku, Chiba 263-8522, Japan
2
Graduate School of Global and Transdisciplinary Studies, Chiba University, Yayoi-cho 1-33, Inage-ku, Chiba 263-8522, Japan
*
Author to whom correspondence should be addressed.
J. Imaging 2022, 8(3), 59; https://doi.org/10.3390/jimaging8030059
Submission received: 27 December 2021 / Revised: 21 February 2022 / Accepted: 22 February 2022 / Published: 27 February 2022
(This article belongs to the Special Issue Intelligent Media Processing)

Abstract

:
This paper proposes an objective glossiness index for objects in halftone color images. In the proposed index, we consider the characteristics of the human visual system (HVS) and associate the image’s structure distortion and statistical information. According to the difference in the number of strategies adopted by the HVS in judging the difference between images, it is divided into single and multi-strategy modeling. In this study, we advocate multiple strategies to determine glossy or non-glossy quality. We assumed that HVS used different visual mechanisms to evaluate glossy and non-glossy objects. For non-glossy images, the image structure dominated, so the HVS tried to use structural information to judge distortion (a strategy based on structural distortion detection). For glossy images, the glossy appearance dominated; thus, the HVS tried to search for the glossiness difference (an appearance-based strategy). Herein, we present an index for glossiness assessment that attempts to explicitly model the structural dissimilarity and appearance distortion. We used the contrast sensitivity function to account for the mechanism of halftone images when viewed by the human eye. We estimated the structure distortion for the first strategy by using local luminance and contrast masking; meanwhile, local statistics changing in the spatial frequency components for skewness and standard deviation were used to estimate the appearance distortion for the second strategy. Experimental results showed that these two mixed-distortion measurement strategies performed well in consistency with the subjective ratings of glossiness in halftone color images.

1. Introduction

The halftoning technique is commonly used in the printing industry to reproduce the tone of an image with limited colors, e.g., black and white. The earliest error diffusion (ED) was proposed by Floyd and Steinberg [1,2]. However, early ED algorithms had problems with distortion, reduced visibility, worms and false textures, as well as additional noise. Later, references [3,4] improved Floyd’s ED. Pang et al. preserved the similarities of tone and structure in ED [5]. Akarun et al. improved the dithering halftone by using variable thresholds [6]. To overcome these shortcomings, Xia et al. developed two convolutional neural networks (CNNs) to learn the halftone scheme by using a nontrivial self-supervision formulation [7]. Thus, improvements to the ED algorithm were implemented with the goal of improving “image quality”. With the maturity of halftone technology, how to evaluate halftone accurately and objectively is still an unsolved problem. Most image glossiness assessment studies [8] are all device-based, i.e., based on photometer measurement. In the study of gloss perception in an image, it is currently debated whether the perception of gloss is linked to the statistical parameters of the retinal image [9,10,11,12]. Wiebel et al. analyzed many images of natural surfaces to search for potential statistical correlations of perceived gloss [13]. Pont and Koenderink found that skewness correlated with gloss when using rendered stimuli; however, the standard deviation, a measure of contrast, correlated better with perceived gloss when using photographs of natural surfaces [14]. Finally, the study verified the key role of contrast by manipulating the skewness and contrast within images. The structural similarity (SSIM) index [15] has become a standard in the image processing field. Studies in Refs. [16,17,18] used multiple strategies for image quality assessment (IQA). However, although the proposed methods simulated the human visual system (HVS) as much as possible, they could not be applied to the assessment of halftone images directly.
To evaluate the glossiness of objects in halftone images, the isolated dots in halftone images should be processed first. The dot process is a proposed method for material printing [19,20]. However, the method only searches for the inter-dot relationship measurement, spectral characteristics between dots, and the image spatial characteristics. Lee et al. proposed an innovative halftone IQA approach for color images [21]. HVS characteristics were applied to the proposed method, which used color spatial HVS filters for luminance, red-green, and blue-yellow components. This model effectively measured color distortion consistent with that of a human observer. However, the index evaluated “image quality” and could not be used to evaluate the gloss of an object.
Here, we propose a fully referenced image glossiness index for halftone color images. Instead of treating image distortions in the same way, we assume that they are decoupled into two groups: structural dissimilarity (SDSIM) and appearance perception, which would be correlated in diverse ways with visual non-glossiness and glossiness perception.

2. Related Work

Given a distorted image, humans can easily rate image quality. To eliminate the shortcomings of evaluation due to human observers, many researchers have focused on the assessment of computational models based on HVS [22,23,24]. The image quality metrics developed for traditional images are usually meant for multi-level (8-bit) images; hence, they cannot be applied directly to 1- or 2-bit discrete images. To overcome this issue, it is necessary to devise a process (e.g., a visual filter) to increase gradation.
To the best of our knowledge, the earliest halftone assessment index was proposed by Näsänen [25]. Näsänen’s method used an exponential function for the luminance component to evaluate a dithering halftone image. Later, Lee et al. proposed a new exponential function for evaluating halftone color images [21]. To consider the visual color characteristics, they performed a new function for each luminance, red-green, and blue-yellow components. Then, the color SSIM was applied to the reference and distorted images for the purpose of localized structural difference assessment. The index in [21] is useful for non-gloss images because their approach is based on SSIM, so it is effective for non-gloss objects with rich structural information. If there is an object such as a mirror, for a single-strategy index, they work well on all parts of the glossy object that reflect the environment surrounding it. However, for gloss objects, the proposed methods would fail because the glossy part contains little structural information. Therefore, to realize the correct assessment of an image, which contains glossy and non-glossy parts objectively, a single assessment strategy could not meet this demand.
In contrast to single-strategy image assessment methods, studies in Refs. [16,17,18] used a variety of strategies to simulate HVS closely for image assessment. In [26], instead of treating the image distortions equally, they proposed treating distortions as linear frequency distortions and additive noise degradations. However, this algorithm only focused on the halftone artifacts. Furthermore, the methods proposed in [17] did not solve the problem of combining the separated distortion measurement. Based on [27], they proposed two simple quality measures, i.e., the detail loss measure and the additive impairment measure, and developed a method of adaptively merging the two strategies. In [16], it was assumed that the HVS determined the image quality by performing different strategies and modeled them together using trained parameters. To solve the texture resampling low tolerance problem for the image assessment index, Ding et al. designed a monotonous and differentiable function using a CNN. Through this function, the image was transformed into a multi-scale representation [18]. Then, IQA metric that mixed correlations of texture similarity and correlations of structure similarity was developed.
Inspired by these strategies, this study develops an image glossiness index for halftone color images based on structure and appearance distortion.

3. Proposed Method

The most apparent distortion (MAD) method was proposed by Larson [16]. In this paper, the author uses two strategies for distortion detection: the first strategy is based on the high-quality image detection strategy, which uses the mean squared error (MSE) to calculate the distortion of visibility. The MSE is the simplest and most widely used full-reference quality metric, which is computed by averaging the squared intensity differences of distorted and reference image pixels. The second detection strategy is based on appearance distortion. The strategy uses standard deviation, skewness, and kurtosis to calculate the distortion of low-quality images. Based on the idea of the MAD algorithm, the proposed algorithm (The execute file is available at https://github.com/donghuilee2022/IQA-for-halftone-image/tree/master, accessed on 12 February 2022) also evaluates the gloss of the image through two strategies. In the first strategy, we remove the HVS processing in the first strategy of MAD and use SDSIM replace MSE to calculate the distortion of visibility. In the second strategy, we only use the standard deviation and skewness related to gloss perception to calculate the appearance distortion. In addition, we also use a new HVS filter to preprocess the image. Finally, the adaptive method proposed in MAD is used to merge two strategies. The details are presented in the following subsections. First, we explain the near-threshold distortions using a method for quantifying perceived distortion, which was used to model the HVS detection. Second, we explain suprathreshold distortion, which was used to model HVS statistical characteristics. Third, we used a parametric method to combine the two modeled perceptual distortions, thereby achieving a single perception of overall distortion. Figure 1 presents a flowchart of the proposed algorithm.

3.1. HVS Filter

A visual perceptual model was used in our algorithm to consider the color characteristics of the HVS, which is described in Section 3.1.1 and Section 3.1.2.

3.1.1. Color Space

The RGB images were transformed to CIEXYZ and then to CIELAB color space. In our experiments in Section 4, we assume color space is sRGB. XYZ is a special set of tristimulus values used for transforming between Lab and RGB. We use L , a , and b to represent the components of CIELAB, respectively. X n , Y n , and Z n are tristimulus values with the D65 white point.
L = 116 · ( Y Y n ) 1 3 16 f o r Y / Y n > 0.008856 903.3 · ( Y Y n ) o t h e r w i s e
a = 500 · f X X n f Y Y n
b = 200 · f Y Y n f Z Z n
w h e r e f ( t ) = t 1 3 f o r t > 0.008856 7.787 · t + 16 116 o t h e r w i s e

3.1.2. Contrast Sensitivity Function

To further improve the accuracy of the model to simulate the HVS, we chose the following exponential function for the luminance contrast sensitivity function.
W ( L ) ( ρ ˜ ) = K ( L ) e α ( L ) ρ ˜
K ( L ) = a L b
α ( L ) = 1 c ln ( L ) + d
The luminance for an image was represented by L [ c d / m 2 ] , ρ (cycles/degree) was the spatial frequency, and a = 131.6, b = 0.3188, c = 0.525, d = 3.91. In contrast to Näsänen’s model, ρ ˜ was defined as the weighted magnitude of ρ = ( u , v ) :
ρ = u 2 + v 2
ϕ = arctan v u
s ( ϕ ) = 1 ω 2 cos 4 ϕ + 1 + ω 2
ρ ˜ = ρ s ( ϕ )
The value of ω set to 0.7. s ( ϕ ) is a weight function. At odd multiples of 45 , this function reduced the contrast sensitivity to the components of the spatial frequency.
Comparing the sensitivity of human observers to spatial variations in luminance and to spatial variations in chromaticity, it was found that the latter decreased faster as the spatial frequency increased. The HVS chrominance model used here was based on Mullen’s [28]. The chromaticity CSF
W ( a , b ) ( ρ ) = A e α ρ
The parameters α and A were set to 0.419 and 100 for the chrominance component.
Using a chromaticity response model that was different from the above luminance resulted in low-frequency chromaticity errors, which were difficult to perceive by HVS. Figure 2 shows two different frequency response models, which represent the luminance and chromaticity response models.
In the luminance model, the weighting function effectively reduces the contrast sensitivity to spatial frequency components at odd multiples of 45 . In the chrominance model, the contrast sensitivity of the human observer to spatial variations in chrominance falls off faster as a function of increasing spatial frequency than the response to spatial variations in luminance.

3.2. Sdsim Distortion Detection Strategy

We used the visible distortion location calculation method proposed in MAD, but not to conduct the CSF filter. We only used perceived luminance and contrast masking methods in the high-quality assessment of MAD.

3.2.1. Calculation of the Locations of Visible Distortion

Let I o r g and I d s t represent the reference and halftone images, respectively, both of which were processed by the HVS filter in Section 3.1.
Perceived luminance: The reference and distorted images were transformed to luminance images via:
L = ( b + k I ) γ
where L represents the luminance image and the parameters b = 0, k = 0.02874, and γ = 2.2. L o r g and L d s t were calculated using the above equation. To consider the HVS nonlinear response to luminance, L o r g and L d s t were transformed to luminance perception images L ^ o r g and L ^ d s t via:
L ^ = L 3
L ^ e r r = L ^ o r g L ^ d s t is defined as the error image.
Contrast masking: This masking explained the fact that image presence reduced the distortion detectability. First, the original image was divided into several 16 × 16 blocks; the change in the block size changed the average contrast of each block. The local contrast map corresponding to each block was calculated. Second, we calculated the root mean square (RMS) contrast for each block. The RMS contrast for block p of I o r g was calculated as:
C o r g ( p ) = σ ˜ o r g ( p ) / μ o r g ( p )
where the mean value of block p is μ o r g ( p ) , and σ ˜ o r g ( p ) is calculated from the standard deviations of the four sub-blocks of p. C o r g ( p ) represented the local RMS contrast measurement of the original image, which was separated from the distortion of the image. Then, a local contrast map was calculated for the error image, which explained the distorted spatial distribution of the distorted image. In addition, I e r r was divided into 16 × 16 blocks, corresponding to the original image I o r g . Each block’s RMS contrast C e r r ( p ) was calculated via:
C e r r ( p ) = σ e r r ( p ) / μ o r g ( p ) i f μ o r g ( p ) > 0.5 0 o t h e r w i s e
where σ o r g ( p ) represents the standard of block p in I e r r . The lightness threshold of 0.5, which explained why the HVS is insensitive to variations in dark regions. Finally, C o r g ( p ) and C e r r ( p ) are used to calculate the local distortion visibility map ξ ( p ) :
ξ ( p ) = ln C e r r ( p ) ln C o r g ( p ) i f ln C e r r ( p ) > ln C o r g ( p ) > δ ln C e r r ( p ) δ i f ln C e r r ( p ) > δ > ln C o r g ( p ) 0 o t h e r w i s e
specifically, ξ ( p ) reflected the amount by which the contrast of the error was larger than the contrast of the original image, if their contrast was greater than the threshold ( δ = 5 ) .

3.2.2. The Combination of Local Structure Errors and Visibility Map

After calculating the visible location map, we use visibility-weighted local SDSIM which is used in the lightness domain to obtain the distortion of the perceived structure. The M S S I M M a p is the M S S I M distribution map of one image, and S D ( p ) is calculated from the M S S I M mapping as follows:
S D ( p ) = 1 16 2 i , j N p M S S I M M a p 2
where quantity S D ( p ) represents the local SDSIM calculated for each 16 × 16 block p. N p is the set of pixels inside block p, i, and j is the position of the pixel ( i , j ) in block p. M S S I M is the mean of S S I M , which was proposed in [18]:
M S S I M ( m , n ) = 1 M j = 1 M S S I M ( x j , y j )
where m and n represent the original and halftone images. x j and y j are the image contents at the local window of jth, and M is the total number of local windows. Therefore, the perceived distortion d s d s i m was computed by:
d s d s i m = 1 P p [ ξ ( p ) × S D ( p ) ] 2 1 2
Equation (20) represents a single value calculated from the visibility-weighted local SDSIM by using the L2 norm, which represented the sum of the visual structure of image distortion. d s d s i m = 0 meant that the distortion in the distorted image would not be perceived by the visual system, i.e., it was not visible. The larger the value of d s d s i m , the greater the distortion perceived. Figure 3 shows the images of the maps involved in the d s d s i m computation for a halftone image. Figure 3a,b show the original and halftone images, respectively.
These figures showed the calculated visibility map, local SDSIM map, and visibility-weighted local SDSIM map. In Figure 3c, the visibility map captured visible artifacts. In Figure 3d, the local SDSIM indicated that the greatest distortions appeared in the regions of greatest energy, so the distortions in these regions were invisible. As Figure 3e shows, the visibility-weighted local SDSIM map had a better performance in predicting the locations and perceived visible intensities of distortions.

3.3. Appearance Distortion Detection Strategy

When an image is of low quality, visual masking is less important for image quality judgment; on the contrary, when the image distortion exceeds the threshold, the degree of quantification of the distortion to reduce the appearance of the image subject can better simulate the visual system perceived distortion. Therefore, in this type of distortion, the HVS’s judgment on the image was switched to a judgment based on the appearance of the image.
To model this appearance perception mechanism of the visual system, a method based on local statistics was developed, which used a multiscale log-Gabor filter response for statistical calculation. The use of this type of statistical model to capture the appearance of texture has been used in various image processing studies. In addition, existing research showed that log-Gabors better simulated and modeled simple cells in the primary visual cortex. In the processing of texture, the change in pixel-based statistics was less obvious than that in the log-Gabor filter response-based statistics.

3.3.1. Log-Gabor Decomposition

Both original and halftone images are first transformed into a set of sub-bands using a log-Gabor filter bank. By calculating the inverse Discrete Fourier Transform (DFT) of the product of the image DFT and the following two-dimensional frequency response, filtering for obtaining sub-bands are performed in the frequency domain.
The original image and the halftone image are decomposed by log-Gabor by multiplication in the frequency domain. { c ´ s , o } represents the set of log-Gabor sub-bands calculated for either the original or halftone image; here, each sub-band c ´ s , o R M × N has the size of the images. The log-Gabor decomposition is calculated using five scales s = 1 , , 5 , and 4 orientations o = 1 , , 4 ; thus, each image has 20 sub-bands. This decomposition is applied to both the filtered original image I o r g and the halftone image I d s t to obtain the sub-band sets { c ´ s , o o r g } and { c ´ s , o d s t } .

3.3.2. Compare Sub-Band Statistics

By comparing the local sub-band statistics of the original image with the corresponding local sub-band statistics of the distorted image, the local statistical difference map η ( p ) was computed. For each 16 × 16 block, the difference in standard deviation and skewness of the corresponding sub-band coefficients of the block were calculated as:
η ( p ) = s = 1 5 o = 1 4 w s [ | σ s , o o r g ( p ) σ s , o d s t ( p ) | + | ς s , o o r g ( p ) ς s , o d s t ( p ) | ]
where σ s , o ( p ) and ς s , o ( p ) represent the standard deviation and skewness of the 16 × 16 sub-band coefficients corresponding to scale s and orientation o, respectively, and corresponded to block p at a certain location. The fixed scale weights w s = 0.5 , 0.75 , 1 , 5 , and 6 are used to explain that the HVS-preferred coarse scales rather than fine scales. The final scalar value of the perceived distortion is given as:
d a p p e a r a n c e = 1 P p η ( p ) 2 1 2
where the summation is for all blocks, and P represents the total number of blocks. d s d s i m = 0 meant no distortion perceived, and an increase in the d s d s i m value indicates an increase in perceived distortion, which reduces visual quality.

3.4. Adaptive Combination of Two Strategies

The adaptive combination method is based on the study in [19] that the observer intends to interactively judge low-quality and high-quality images. Here, our proposed index uses a weighted geometric mean of d s d s i m and d a p p e a r a n c e , given by:
I n d e x = ( d s d s i m ) α ( d a p p e a r a n c e ) 1 α
Here, α is computed via:
α = 1 1 + β 1 ( d s d s i m ) β 2
The parameters β 1 and β 2 were provided in [16]. For the database of A57, the optimization values of these parameters are β 1 = 0.467 and β 2 = 0.130 .

4. Experiment

4.1. Subjective Image Database and Processing

The proposed algorithms were implemented using MATLAB (R2021a) running on the Mac OS. The experiments were conducted using 100 images from the Flickr Material Database [29]. The images we selected included not only pure glossy and non-glossy images but also images that contained both glossy and non-glossy parts, such as images containing textured leaves and water drops. Figure 4 shows the partial glossy and non-glossy images. Based on the above images, we conducted subjective and objective experiments, respectively. Then, we calculated the correlation between them. To verify the correlation between them, we ranked the observers’ scores from high to low, and the corresponding objective scores were automatically ranked. Then, we specified the 50 data pairs with the highest subjective score as the score of the non-glossy image and the other 50 data pairs as the score of gloss images.
For comparison, we selected several halftone algorithms to process the selected 100 images. These algorithms included three different halftone algorithms: dithering [27], Floyd [1], and direct binary search (DBS) [30]. These different processing data were treated as three diverse types of distorted images. Different halftone technologies often produced different halftone effects. We process the color image by using color separation where an RGB image is divided into separate R, G, and B components with sRGB. After being processed by the halftone algorithm, the components are finally synthesized into an RGB image. In this process, the RGB color space is used. It was believed that the ED halftone reproduced more details of the original image than the dithering algorithm. Therefore, we believed that the distortion effect of the ED algorithm was less than that of the dithering halftone algorithm. Figure 5 shows an example of the application of each algorithm to the enlarged images. The size of the images is 384 × 384 pixels. The error diffusion and dithering halftone dots occupy 1 pixel and 4 × 4 pixels, respectively. In dither halftoning, the dot size determines the image quality of halftone and affects the printed image. Too large a dot size will cause the image to lose too many details and also affect the output effect of printer. Too small a dot size will appear as regular fence phenomenon, reducing the image quality.
In the actual printing process, various effects (e.g., dot gain) occur, depending on the characteristics of the paper. They can be perceived differently depending on lighting effects. Therefore, it is difficult to isolate how these effects impact the subjective evaluation of the printed material. Therefore, we evaluated the digital data presented on the display device. The original image and the halftone image were displayed simultaneously, and the observer scored them. Observers were asked to assign a score from 0 to 100 for each test image pair. The viewing distance was designed to be equivalent to the retinal image of a 600 dpi print at 24 cm. Five observers participated in a subjective observer experiment with normal vision.
For comparison, we chose several image quality evaluation algorithms for the experiments. We calculated the peak signal-to-noise ratio (PSNR), color structural similarity (CSIM) [21], MAD [16], and the proposed index to compare the correlation between the results of each index and the subjective observer score. We showed normalized index values for MAD and the proposed method because the indexes belonged to [ 0 , ] .

4.2. Experimental Procedure

There are five students (three males and two females; age 27.2 ± 6.65) who participated in the experiment. One student (the author) had significant experience with the subjective assessment of printing image quality. The other four students were all non-experts in printing and image quality assessment. All students were screened prior to participation for normal, or corrected to normal, visual acuity and normal color vision. Image pairs were presented on the display for a duration of 10s. A reference image and a halftone image are placed on the left and right, respectively. Observers were given instructions to judge the reproduction of objects’ glossiness in images and to provide corresponding scores. After observing an image pair, observers were asked to provide corresponding scores to the image pair using a continues horizontal scale, as depicted in Figure 6. The scores corresponding to the above interval are 100, 80, 60, 40, 20, and 0, respectively. Scores ranged from 100 to 0 using a continuous horizontal scale, with 0 representing poor reproduction and 100 representing objects that were most similar to the original image.

4.3. Results and Discussion

To assess the consistency by five observers measuring the same quality, we calculated the interclass correlation coefficient (ICC) for the inter-observer consistency. ICC is a descriptive statistic that can be used when quantitative measurements are made on units that are organized into groups. It describes how strongly units in the same group resemble each other. Before we calculated the consistency with the subjective score, we conducted a reliability study to evaluate the inter-observer test-retest reliability. Based on 100 images of each halftone, we repeated the subjective experiment for each image pair with five observers, we analyzed the data using a single-measurement, absolute-agreement, two-way mixed-effects model. For the different halftones, our ICC report is summarized in Table 1.
As shown in Table 1, we concluded that the test-retest reliability of our subjective experiment is “moderate” and “good”.
The Pearson correlation coefficient (PCC) was the most common measure of predictive performance. Here, we used PCC to calculate the correlation between the observer scores and objective index values. To distinguish the effects of metrics on glossy and non-glossy images, we sorted the subjective data in descending order, and the objective data were sorted accordingly. The first 50 images of the arranged data were classified as non-glossy images, and the rest were classified as glossy images. The former was represented by red points, the latter by blue points. Figure 7 shows the scatter plots of PSNR, CSIM, MAD, and the proposed image glossiness metric for 100 images based on Floyd. In all graphs, each point represented a test image. Figure 8 shows the error bar of the subjective score based on Floyd from the five observers. In Figure 7, the vertical axis represents the subjective ratings of the perceived distortions, and the horizontal axis represents the metrics. As Figure 7 shows, the proposed index evaluated the glossiness correctly compared to other indices. Comparable results were obtained for the other halftone methods.
Table 2, Table 3 and Table 4 show the PCC between four kinds of different metrics and subjective observer scores based on three different halftoning algorithms (dither, Floyd, and DBS). We separately calculated the correlation coefficients of glossy, non-glossy, and 100 images. Based on different distortion types, we separately calculated the PCC of glossy images, non-glossy images, and 100 images. The proposed index was consistent with the subjective evaluation regardless of the glossiness of the object.

5. Conclusions

We proposed an effective index that explicitly separated structure detection and appearance for the glossiness of objects in halftone images. For structure detection, the SDISM algorithm was developed, which worked effectively for structural distortion. Two important HVS characteristics, i.e., CSF and contrast masking, were incorporated into the metric to better simulate HVS responses to visual inputs. We proposed two simple quality measures, SDSIM and appearance perception, which were responsible for correlating structure, skewness, and standard deviation. Through experiments using 300 test images halftoned from 100 images in FMD, we demonstrated the effectiveness of the proposed index, which was consistent with the subjective evaluation scores regardless of the glossiness of the object.
We will continue to investigate effective indices for other appearances of objects, such as perceptual transparency. Our future research will consider more advanced color spaces, such as S-CIELAB and iCAM. New material perception models will also provide potential research possibilities, such as the new material perception model proposed in paper [31]. Furthermore, we will further verify the effectiveness of our index in the actual printing with different papers and lighting environments.

Author Contributions

Conceptualization, T.H. and M.T.; methodology, T.H. and D.L.; software, D.L.; investigation, D.L., M.T. and T.H.; resources, T.H.; data curation, D.L.; writing—original draft preparation, D.L. and T.H.; writing—review and editing, M.T. and T.H.; supervision, T.H. and M.T.; project administration, T.H.; funding acquisition, T.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by JSPS KAKENHI (Grant 20H05957 to T.H.).

Institutional Review Board Statement

Ethics approval was not required for this study. All necessary participant consent has been obtained, and the appropriate institutional forms have been archived.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Floyd, R. An Adaptive Algorithm for Spatial Grey Scale. Soc. Inf. Display 1975, 17, 36–37. [Google Scholar]
  2. Floyd, R. An adaptive algorithm for spatial grey scale. Proc. Soc. Inf. Display 1976, 17, 75–77. [Google Scholar]
  3. Damera-Venkata, N.; Evans, B.L. Adaptive threshold modulation for error diffusion halftoning. IEEE Trans. Image Process. 2001, 10, 104–116. [Google Scholar] [CrossRef] [PubMed]
  4. Monga, V.; Evans, B.L. Tone dependent color error diffusion. In Proceedings of the 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing, Montreal, QC, Canada, 17–21 May 2004; Volume 3, pp. 3–101. [Google Scholar]
  5. Pang, W.M.; Qu, Y.; Wong, T.T.; Cohen-Or, D.; Heng, P.A. Structure-aware halftoning. In Proceedings of the ACM SIGGRAPH 2008 Special Interest Group on Computer Graphics and Interactive Techniques Conference, Los Angeles, CA, USA, 11–15 August 2008; pp. 1–8. [Google Scholar]
  6. Akarun, L.; Yardunci, Y.; Cetin, A.E. Adaptive methods for dithering color images. IEEE Trans. Image Process. 1997, 6, 950–955. [Google Scholar] [CrossRef] [PubMed]
  7. Xia, M.; Hu, W.; Liu, X.; Wong, T.T. Deep halftoning with reversible binary pattern. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montrel, QC, Canada, 11–17 October 2021; pp. 14000–14009. [Google Scholar]
  8. Leloup, F.B.; Audenaert, J.; Hanselaer, P. Development of an image-based gloss measurement instrument. J. Coatings Technol. Res. 2019, 16, 913–921. [Google Scholar] [CrossRef] [Green Version]
  9. Anderson, B.L.; Kim, J. Image statistics do not explain the perception of gloss and lightness. J. Vis. 2009, 9, 10–17. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Ferwerda, J.A.; Pellacini, F.; Greenberg, D.P. Psychophysically based model of surface gloss perception. In Proceedings of the Human Vision and Electronic Imaging VI. International Society for Optics and Photonics, San Jose, CA, USA, 20–26 January 2001; Volume 4299, pp. 291–301. [Google Scholar]
  11. Thompson, W.; Fleming, R.; Creem-Regehr, S.; Stefanucci, J.K. Visual Perception from a Computer Graphics Perspective; CRC Press: Boca Raton, FL, USA, 2011. [Google Scholar]
  12. Toscani, M.; Valsecchi, M.; Gegenfurtner, K.R. Optimal sampling of visual information for lightness judgments. Proc. Natl. Acad. Sci. USA 2013, 110, 11163–11168. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Wiebel, C.B.; Toscani, M.; Gegenfurtner, K.R. Statistical correlates of perceived gloss in natural images. Vis. Res. 2015, 115, 175–187. [Google Scholar] [CrossRef] [PubMed]
  14. Pont, S.C.; Koenderink, J.J. Reflectance from locally glossy thoroughly pitted surfaces. Comput. Vis. Image Underst. 2005, 98, 211–222. [Google Scholar] [CrossRef]
  15. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Larson, E.C.; Chandler, D.M. Most apparent distortion: Full-reference image quality assessment and the role of strategy. J. Electron. Imaging 2010, 19, 011006. [Google Scholar]
  17. Li, S.; Zhang, F.; Ma, L.; Ngan, K.N. Image quality assessment by separately evaluating detail losses and additive impairments. IEEE Trans. Multimed. 2011, 13, 935–949. [Google Scholar] [CrossRef]
  18. Ding, K.; Ma, K.; Wang, S.; Simoncelli, E.P. Image quality assessment: Unifying structure and texture similarity. arXiv 2020, arXiv:2004.07728. [Google Scholar] [CrossRef] [PubMed]
  19. Ulichney, R. Digital Halftoning; MIT Press: Cambridge, MA, USA, 1987. [Google Scholar]
  20. Lau, D.L.; Arce, G.R. Modern Digital Halftoning; CRC Press: Boca Raton, FL, USA, 2018; Volume 1. [Google Scholar]
  21. Lee, J.; Horiuchi, T. Image quality assessment for color halftone images based on color structural similarity. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2008, 91, 1392–1399. [Google Scholar] [CrossRef]
  22. Mannos, J.; Sakrison, D. The effects of a visual fidelity criterion of the encoding of images. IEEE Trans. Inf. Theory 1974, 20, 525–536. [Google Scholar] [CrossRef]
  23. Winkler, S. Visual quality assessment using a contrast gain control model. In Proceedings of the 1999 IEEE Third Workshop on Multimedia Signal Processing (Cat. No. 99TH8451), Copenhagen, Denmark, 13–15 September 1999; pp. 527–532. [Google Scholar]
  24. Osberger, W.; Bergmann, N.; Maeder, A. An automatic image quality assessment technique incorporating higher level perceptual factors. In Proceedings of the 1998 International Conference on Image Processing ICIP98 (Cat. No. 98CB36269), Chicago, IL, USA, 7 October 1998; pp. 414–418. [Google Scholar]
  25. Násánen, R. Visibility of halftone dot textures. IEEE Trans. Syst. Man Cybern. 1984, 14, 920–924. [Google Scholar] [CrossRef]
  26. Damera-Venkata, N.; Kite, T.D.; Geisler, W.S.; Evans, B.L.; Bovik, A.C. Image quality assessment based on a degradation model. IEEE Trans. Image Process. 2000, 9, 636–650. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Verevka, O.; Buchanan, J.W. Halftoning with image-based dither screens. GI 1999, 99, 167–174. [Google Scholar]
  28. Mullen, K.T. The contrast sensitivity of human colour vision to red-green and blue-yellow chromatic gratings. J. Physiol. 1985, 359, 381–400. [Google Scholar] [CrossRef] [PubMed]
  29. Sharan, L.; Liu, C.; Rosenholtz, R.; Adelson, E.H. Flickr Material Database. Available online: https://people.csail.mit.edu/lavanya/fmd.html (accessed on 12 February 2022).
  30. Analoui, M.; Allebach, J.P. Model-based halftoning using direct binary search. In Proceedings of the SPIE/IS&T 1992 Symposium on Electronic Imaging: Science and Technology, San Jose, CA, USA, 9–14 February 1992; Volume 1666, pp. 96–108. [Google Scholar]
  31. Serrano, A.; Chen, B.; Wang, C.; Piovarči, M.; Seidel, H.P.; Didyk, P.; Myszkowski, K. The effect of shape and illumination on material perception: Model and applications. ACM Trans. Graph. (TOG) 2021, 40, 1–16. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Jimaging 08 00059 g001
Figure 2. (a) Luminance frequency response model. (b) Chrominance frequency response model.
Figure 2. (a) Luminance frequency response model. (b) Chrominance frequency response model.
Jimaging 08 00059 g002
Figure 3. (a) Original image, (b) Floyd, (c) visible map, (d) SDSIM, and (e) perceived distortion.
Figure 3. (a) Original image, (b) Floyd, (c) visible map, (d) SDSIM, and (e) perceived distortion.
Jimaging 08 00059 g003
Figure 4. Test samples: (a) glossy images: glass, metal, plastic, and water droplets; (b) non-glossy images: fabric, plants, paper, stone, and wood.
Figure 4. Test samples: (a) glossy images: glass, metal, plastic, and water droplets; (b) non-glossy images: fabric, plants, paper, stone, and wood.
Jimaging 08 00059 g004
Figure 5. Examples of samples. Left and right images show gloss and non-gloss images, respectively: (a) original images; (b) halftone images by dithering; (c) halftone images by Floyd; (d) halftone images by DBS.
Figure 5. Examples of samples. Left and right images show gloss and non-gloss images, respectively: (a) original images; (b) halftone images by dithering; (c) halftone images by Floyd; (d) halftone images by DBS.
Jimaging 08 00059 g005
Figure 6. Rating scale used in subjective experiment.
Figure 6. Rating scale used in subjective experiment.
Jimaging 08 00059 g006
Figure 7. Scatter plots of the PSNR, CSIM, MAD, and proposed metric for test samples halftoned by Floyd: (a) PSNR, (b) CSIM, (c) MAD, and (d) Ours.
Figure 7. Scatter plots of the PSNR, CSIM, MAD, and proposed metric for test samples halftoned by Floyd: (a) PSNR, (b) CSIM, (c) MAD, and (d) Ours.
Jimaging 08 00059 g007
Figure 8. Subjective scores with error bars based on Floyd from 5 observers.
Figure 8. Subjective scores with error bars based on Floyd from 5 observers.
Jimaging 08 00059 g008
Table 1. ICC and 95% confident interval for different halftone.
Table 1. ICC and 95% confident interval for different halftone.
HalftoneICC95% Confident Interval
Dithering0.780.72–0.84
Floyd0.820.74–0.85
DBS0.850.76–0.87
Table 2. Pearson correlation coefficient for PSNR, CSIM, MAD, and Ours based on dithering.
Table 2. Pearson correlation coefficient for PSNR, CSIM, MAD, and Ours based on dithering.
IndexGlossiness ImagesNon-Glossiness ImagesAll 100 Images
PSNR0.22750.23750.102
CSIM0.1944−0.14920.0038
MAD0.2566−0.02440.0355
Ours0.75630.77070.9066
Table 3. Pearson correlation coefficient for PSNR, CSIM, MAD, and Ours based on Floyd.
Table 3. Pearson correlation coefficient for PSNR, CSIM, MAD, and Ours based on Floyd.
IndexGlossiness ImagesNon-Glossiness ImagesAll 100 Images
PSNR0.15950.06260.0029
CSIM0.06680.50670.4053
MAD0.08960.65420.5788
Ours0.76620.87160.9131
Table 4. Pearson correlation coefficient for PSNR, CSIM, MAD, and Ours based on DBS.
Table 4. Pearson correlation coefficient for PSNR, CSIM, MAD, and Ours based on DBS.
IndexGlossiness ImagesNon-Glossiness ImagesAll 100 Images
PSNR−0.0954−0.2506−0.2509
CSIM0.62940.54530.7132
MAD0.57350.61540.7045
Ours0.73110.88810.9101
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, D.; Tanaka, M.; Horiuchi, T. Glossiness Index of Objects in Halftone Color Images Based on Structure and Appearance Distortion. J. Imaging 2022, 8, 59. https://doi.org/10.3390/jimaging8030059

AMA Style

Li D, Tanaka M, Horiuchi T. Glossiness Index of Objects in Halftone Color Images Based on Structure and Appearance Distortion. Journal of Imaging. 2022; 8(3):59. https://doi.org/10.3390/jimaging8030059

Chicago/Turabian Style

Li, Donghui, Midori Tanaka, and Takahiko Horiuchi. 2022. "Glossiness Index of Objects in Halftone Color Images Based on Structure and Appearance Distortion" Journal of Imaging 8, no. 3: 59. https://doi.org/10.3390/jimaging8030059

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop