Next Article in Journal
Adaptive Machine Learning Based Distributed Denial-of-Services Attacks Detection and Mitigation System for SDN-Enabled IoT
Next Article in Special Issue
Assessment for Different Neural Networks with FeatureSelection in Classification Issue
Previous Article in Journal
Advanced Fiber Sensors Based on the Vernier Effect
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Iris Image Compression Using Deep Convolutional Neural Networks

Department of Computer Science, University of Salzburg, Jakob-Haringer-Straße 2, 5020 Salzburg, Austria
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(7), 2698; https://doi.org/10.3390/s22072698
Submission received: 11 March 2022 / Revised: 25 March 2022 / Accepted: 29 March 2022 / Published: 31 March 2022
(This article belongs to the Special Issue Feature Extraction and Forensic Image Analysis)

Abstract

:
Compression is a way of encoding digital data so that it takes up less storage and requires less network bandwidth to be transmitted, which is currently an imperative need for iris recognition systems due to the large amounts of data involved, while deep neural networks trained as image auto-encoders have recently emerged a promising direction for advancing the state-of-the-art in image compression, yet the generalizability of these schemes to preserve the unique biometric traits has been questioned when utilized in the corresponding recognition systems. For the first time, we thoroughly investigate the compression effectiveness of DSSLIC, a deep-learning-based image compression model specifically well suited for iris data compression, along with an additional deep-learning based lossy image compression technique. In particular, we relate Full-Reference image quality as measured in terms of Multi-scale Structural Similarity Index (MS-SSIM) and Local Feature Based Visual Security (LFBVS), as well as No-Reference images quality as measured in terms of the Blind Reference-less Image Spatial Quality Evaluator (BRISQUE), to the recognition scores as obtained by a set of concrete recognition systems. We further compare the DSSLIC model performance against several state-of-the-art (non-learning-based) lossy image compression techniques including: the ISO standard JPEG2000, JPEG, H.265 derivate BPG, HEVC, VCC, and AV1 to figure out the most suited compression algorithm which can be used for this purpose. The experimental results show superior compression and promising recognition performance of the model over all other techniques on different iris databases.

1. Introduction

Biometric human recognition systems are extensively used, and yet are increasing in demand for various applications in recent years. Among all biological traits used, iris is well suited for the most accurate and secure personal identification/verification because of the distinctive patterns present in the iris textures of individuals. To maintain high level of accuracy, the iris images presented to the iris recognition systems need to possess a relentless virtual quality and perception. Nonetheless, considering interoperability and vendor neutrality, often authorities, regulatory bodies, and international standards organizations specify that biometric data must be stored and preserved in image form, rather than in (or in addition to) extracted templates that may depend on proprietary algorithms. Recording raw (high quality) image data in such systems also allows them to benefit from inevitable future improvements in recognition algorithms. However, storage, processing and transmission of such high quality data comes at high cost. To this extent, efficient storage and rapid transmission of iris biometric records is a driving implementation factor in iris recognition system development (especially on low-powered mobile sensors and for portable devices) currently. Image compression techniques reduce the amount of memory used by reducing the number of bits without losing the important data. Image compression will also reduce the transmission time as the transmission time of any image is directly proportional to size of the image. There are two types of compression algorithms, namely “lossless compression” and “lossy compression”. A lossless compression is a reversible process in which no information is lost, but the compression ratio is lower. It is mainly used in the domain where reliability is important: for example executable files and medical data. Lossy compression is a non-reversible process and some information may be lost, but compression ratio is very high and is mainly used in applications where loss of data is acceptable to a certain degree.
Recent machine learning techniques proposed for lossy image compression have generated considerable interest in both the machine learning and image processing communities. Like all lossy compression methods, such models operate on a simple principle: an image, typically proposed as a vector of pixel intensities, is quantized, reducing the amount of information required to record or transmit it, which incorporates introducing some error at the same time. Typically, it is not the pixel intensities that are quantized directly. Rather, an alternative (latent) representation of the image is found, and quantization takes place in this representation, yielding a discrete-valued vector. To be more precise, deep learning models learn (extract) the key features in the input image through back propagation training and enable the compression of image information, preserving and restoring such features without too much prior knowledge.
In this work, we extend our previous study [1] on evaluating the expediency of a Deep Semantic Segmentation-based Layered image Compression (DSSLIC) model [2] for iris compression within a biometric recognition framework. As a distinction to the previous work: We utilize a more powerful deep-learning-based model (End-to-end Optimized Image Compression model (EOIC)) to address the lower performance issue of the Conditional Probability Models for Deep Image Compression (CPDIC), as used as the comparing model in the previous experiments. We further consider three new state-of-the-art (non-learning-based, lossy) image compression models along with a couple of other commonly used image compression algorithms to compress iris images in four well-known iris datasets. To evaluate the compression performance in each case, in addition to the Multi-scale Structural Similarity Index (which was used in our previous work), another “Full-Reference” image quality assessment measure, namely Local Feature Based Visual Security (LFBVS), and a “No-Reference” images quality assessment measure, namely Blind Reference-less Image Spatial Quality Evaluator (BRISQUE) are used. The biometric recognition performance then is evaluated, in terms of Equal Error Rate (EER), by using the compressed iris images in some predefined iris biometric systems. At the end, the compression and the corresponding recognition results are compared and carefully analyzed to figure out a well suited compression algorithm to be employed in iris recognition systems. The rest of the paper is structured as follows: Section 2 will review related works and state of the art, and Section 3 will describe the deep-learning based model (DSSLIC). Section 4 will give the details of the experimental framework, and the experiments and analysis will be presented in Section 5. Finally, Section 6 will conclude the paper.

2. Related Work

Many studies have been conducted on iris image compression and the subsequent recognition performance during the past decades). For example: Matschitsch et al. [3] studied the impact of applying different lossy compression algorithms on the matching accuracy of iris recognition systems, relating rate-distortion performance to the matching scores. The authors in this work concluded that JPEG2000, SPIHT, and PRVQ are almost equally well suited for iris compression. Daugman and Downing [4] analyzed the effect of severe image compression on iris recognition performance and introduced a schemes that combine region-of-interest isolation with JPEG and JPEG2000 compression at severe levels, and tested them using a publicly available database of iris images. Grother [5] investigated existing image compression approaches and compared JPEG and JPEG2000 to provide a quantitative support to the revision of the ISO/IEC IS 19794-6, considering: cropped format (IREX K3), masked and cropped format (IREX K7), and an unsegmented polar format (IREX K16) in his experiments. Ives et al. [6] investigated the effects of image compression on recognition system performance using a commercial version of the Daugman “iris2pi” algorithm along with JPEG-2000 compression, and linked that to image quality. Korvath et al. [7] evaluated the impact of dedicated lossless image codecs (lossless JPEG, JPEG-LS, PNG, and GIF), lossless variants of lossy codecs (JPEG2000, JPEG XR, and SPIHT), and some general purpose file compression schemes on the iris images. Bergmüller et al. [8] studied the impact of using pre-compressed data in iris segmentation and evaluated the relation between iris segmentation performance and general image quality metrics. Rathgeb et al. [9] investigated the impact of image compression on iris segmentation algorithms. In their work, they examined the impact of severe image compression using in particular, JPEG, JPEG 2000, and JPEG-XR algorithms on the performance of different iris segmentation approaches.
With recent advancements in deep learning techniques, researchers have proposed few learning based image compression methods as well. Toderici et al. [10] used a recurrent neural networks, based on convolution and deconvolution long short-term memory (LSTM), to extract binary representations which are later compressed with entropy coding. Ballé et al. [11] proposed a compression framework that included a generalized divisive normalization (GDN)-based nonlinear analysis transform, a uniform quantizer, and a nonlinear synthesis transform. Theis et al. [12] introduced a deep-learning-based auto-encoder in which they used smooth approximation instead of quantization to obtain different rates. Agustsson et al. [13] used a soft-to-hard vector quantization model along with a unified formulation for both the compression of deep learning models and image compression. Jiang et al. [14] utilized a compact convolutional neural network (ComCNN) and a reconstruction convolutional neural network (RecCNN) to encode and decode the original image, respectively. Johnston et al. [15] utilized the structural similarity (SSIM) quality measure and spatially adaptive bit allocation to further improve the performance. Li et al. [16] introduced a model which was based on image content weighting. They used the edge feature maps, extracted by a convolution neural network, as the importance map of the original image. In this work, an innovative algorithm was introduced to solve the non-differentiated calculation in the quantization rounding function to achieve a backward propagation gradient in the standard image algorithm. Luo et al. [17] used the benefit of image compression and classification to reconstruct the images and to generate corresponding semantic representations simultaneously. Mantzer et al. [18] proposed a conditional probability model for deep image compression (CPDIC), with concentration on improving the entropy rate of the latent image representation using a context model (a 3D-CNN which learns a conditional probability model of the latent distribution). During training the auto-encoder makes use of the context model to estimate the entropy of its representation, and the context model is concurrently updated to learn the dependencies between the symbols in the latent representation. Wang et al. [19] proposed a compression bit allocation algorithm, which allows a recurrent neural network (RNN)-based compression network to hierarchically compress the images according to semantic importance maps.
Some other works utilized Generative Adversarial networks (GAN) in their architecture to learn the image compression. Santurkar et al. [20] utilized a discriminator to help training of a compression decoder. They calculated a perceptual loss, which was based on the feature map obtained from pretrained ImageNet and AlexNet. Only low-resolution image coding results were reported in this work. Ripple et al. [21] embedded an auto-encoder in a GAN architecture, where the feature extraction adopted a pyramid of inter-scale alignments. They considered the target and its reconstruction jointly as a single entity and, instead of producing an output for classification at the last layers of the network, accumulated scalar outputs along branches constructed at different roots. The average of these scalars was used as the final value passed to a sigmoid function. Augustesson et al. [22] proposed an image compression (synthesis) model based on a GAN architecture. They consider two modes of operations namely “generative compression” and “selective generative compression”. The later approach generated parts of image from a semantic label map. Their model was concentrated on obtaining extremely low bit-rates rather than on preserving key features in image, and to this extent a clear trade off between these two factors remained unclear. Balle et al. [23] proposed an end-to-end trainable encoder–decoder framework for image compression. The encoder part includes a variational generative model, whose out put is quantized, compressed, and transmitted as side information. The decoder recovers the quantized vector from the compressed signal. It then obtains the correct probability estimates to successfully recover and reconstruct the image. Minnen et al. [24] introduced a model which jointly optimizes an autoregressive component that predicts latents from their causal context (Context Model) along with a hyperprior and the underlying autoencoder. Real-valued latent representations then are quantized to create integer-valued latents and hyperlatents, which are compressed into a bit-stream using an arithmetic encoder and decompressed by an arithmetic decoder. Cheng et al. [25] used discretized Gaussian Mixture Likelihoods to parameterize the distributions of latent codes, which removes redundancy to achieve an accurate entropy model, and leads to fewer encoding bits. They take advantage of attention modules to make the learned models pay more attention to complex regions. In any case, the main drawback of applying GAN networks (at least on their own and without further information fusion) to reconstruct the images is their inability to preserve key fine features in the images reconstructed. This phenomenon is observable in the corresponding results obtained using the above models too. In fact, this issue can directly affect the iris key traits and their subsequent recognition performance in the reconstructed images. Figure 1 shows a sample iris image (Figure 1a) (from the Notredame iris dataset, as used in this work), and its compressed version (Figure 1b), using the last GAN based model, along with their difference image (Figure 1c), and the overlaid ground-truth mask (Figure 1d). Gray regions in the difference image show where the two images have the same intensities, and magenta and green regions show where the intensities are different. Furthermore, the overlaid ground-truth mask shows how the actual iris outer and inner boundaries (as specified by the mask) are distorted in the compressed (reconstructed) image. Nevertheless, while the recent architectures (e.g., et al. [23,24,25]) contributed to improve the compressive performance in one way or another, yet their experimental trail (carried out in this research work) revealed that such improvement has come at the cost of loosing some level of scalability. In particular, during our experiments we noticed that on certain data (e.g., Casia5a as used in our experiments), these models are only able to preserve and reconstruct key image features when higher bits rates are used, and simply do not converge to lower bit rates even when highest compression parameters are used. To this extent, we excluded them from our comparison experiments.

3. Deep-Learning-Based Image Compression Model (DSSLIC)

Figure 2 demonstrates the overall architecture of the model used in our work, which is derived from the DSSLIC model already introduced in [2]. As a key distinction to the original model, here we do not use a segmentation network in our model, and instead we provide the manually segmented labels directly to the model. Doing so, we try to provide a better evaluation of the actual performance of the model by introducing more accurate labeling data. In fact, such a configuration is practically justifiable as the recent advancements in the deep-learning-based segmentation techniques has made highly accurate segmentations (that are comparable to the manually segmented labels) available in a timely manner (i.e., [26]).
The overall architecture of the model is composed of an encoding part and a decoding part. The encoding part includes two deep learning networks, namely “CompNet” and “FiNet” (a GAN-based network). The ComNet network takes the iris image as input, while the corresponding segmentation map is encoded to serve as side information to this network for generating a low-dimensional version of the original image. Both the segmentation map and the compact version are losslessly encoded using the FLIF codec [27], which is a state-of-the-art lossless image compression algorithm. Having the segmentation map and up-sampled compact image, the FiNet works to obtain a high-quality reconstruction of the input image. Unlike the similar architectures in which the GAN networks operate directly only on the input images, here in this model, the GAN network takes the segmentation map as the input and tries to learn the missing detail information of the up-sampled version of the compacted input image to minimize the distortion of the synthesized images. It should be noted that although GAN-based synthesized images generated with the help of the segmentation maps are visually appealing, their details can be quite different from the original images. To minimize the errors of the synthesized images, the up-sampled version of the compact image, as an additional input, is used. By doing so, the FiNet learns the missing detail information of the up-sampled version of a compact image with respect to the input image, which in turn controls the output of the GAN network. After adding the up-sampled version of the compact image and the FiNet’s output, we obtain a better estimate of the input. The residual difference between the input and the estimate is then obtained and encoded by a lossy codec (H.265/HEVC intra coding-based BPG). In order to deal with negative values, the residual image is re-scaled to [0, 255] with a min-max normalization before encoding. The min and max values are also sent to the decoder for inverse scaling. In this scheme the segmentation map serves as the base layer and the compact image and the residual are the first and second enhancement layers, respectively. At the decoder side the segmentation map and the compact representation are decoded to be used by the FiNet to obtain an estimate of the input image. The output of FiNet is then added to the decoded residual image to obtain the reconstructed image as output.

4. Experimental Framework

In this section, we describe the methodological details of our experiments including: the databases used, the comparison compression algorithms, metrics and measures, and the recognition pipeline.

4.1. Datasets

For our experiments, we used four different iris datasets: The Notredame dataset [28] includes 835 iris images of 30 different subjects. The images in this dataset are taken in near-infrared spectrum in an indoor environment with the LG 2200 iris biometric system. The Casia4i dataset [29] includes 2640 iris images of 249 subjects. Images in this dataset are acquired under near-infrared illumination, with a close-up iris camera. The IITD dataset [30] includes 2240 iris images of 224 subjects. The images are acquired in the indoor environment, with the Jiris, Jpc1000 digital CMOS camera in near-infrared spectrum. The Casia5a dataset [29] includes 1880 images of both eyes of 94 users. The dataset comprises images captured from a video sequences taken in 2009 and 2013. Special attention should be paid to the fact that: the selection of the datasets and the number of samples included in each case were subject to the availability of the ground-truth masks which were required for the training and the segmentation process. The ground-truth masks were provided by the Multimedia Signal Processing and Security Lab (WaveLab) group at the University of Salzburg.

4.2. Comparing Compression Algorithms

To evaluate the model expediency, primarily we compare its performance against a well-known and popular deep-learning based (lossy) compression algorithm. End-to-end Optimized Image Compression model (EOIC) [11] consists a nonlinear analysis transform, a nonlinear synthesis transform and a uniform quantizer. Each transform begins with an affine convolution on the ith input channel u of the kth stage at spatial location ( m , n ) :
v i ( k ) ( m , n ) = j h k , i j u j ( k ) ( m , n ) + c k , i ,
followed by a down-sampling:
w i ( k ) ( m , n ) = v i ( k ) s k m , s k n ,
where s k is the down-sampling factor for stage k. Each stage concludes with a Generalized Divisive Normalization (GDN) transform:
u i ( k + 1 ) ( m , n ) = w i ( k ) ( m , n ) β k , i + j γ k , i j w j ( k ) ( m , n ) 2 1 2 .
The full set of h, c, β , and γ parameters (across all the three stages) constitute the optimization parameters. This representation then is quantized, yielding a discrete-valued vector which is then compressed. The rate of this discrete code is lower-bounded by the entropy of the discrete probability distribution of the quantized vector. To reconstruct the compressed image, the discrete elements of the vector are reinterpreted as a continuous-valued vector which are transformed back to the data space using a parametric synthesis transform. Analogously, the synthesis transform consists of three stages, with the order of operations reversed within each stage, down-sampling replaced by up-sampling, and GDN replaced by an approximate inverse called IGDN:
w ^ i ( k ) ( m , n ) = u ^ i ( k ) ( m , n ) · β ^ k , i + j γ ^ k , i j u ^ j ( k ) ( m , n ) 2 1 2 .
which is followed by up-sampling:
v ^ i ( k ) ( m , n ) = w ^ i ( k ) m / s ^ k , n / s ^ k ,
where s ^ k is the up-sampling factor for stage k. Finally, this is followed by an affine convolution:
u ^ i ( k + 1 ) ( m , n ) = j h ^ k , i j v ^ j ( k ) ( m , n ) + c ^ k , i .
Analogously, the set of h ^ , c ^ , β ^ , and γ ^ make up the parameter vector θ . We further compare the model performance against some of the most popular and state-of-the-art non-learning-based (lossy) compression methods including: JPEG [31], the current ISO standard JPEG2000 (J2K) [31], the H.265 derived BPG [32], HEVC [33] (HM version 16.20) [34], VVC [35] (VTM version [36] and AV1 [37] (version 2.0) [38] algorithms.

4.3. Metrics and Measures

To evaluate the compression performance we considered three different measures, including two Full-Reference (FR) image quality assessment measures and one objective blind or No-Reference (NR) image quality assessment measure. Full-Reference (FR) image quality assessment requires as input not only the compressed image, but also a “clean”, pristine reference image with respect to which the quality of the compressed image is assessed. Objective blind or No-Reference image quality assessment requires as input only the compressed image, and assesses the quality of the compressed image objectively. Multi-Scale Structural Similarity Index Measure (MS-SSIM) is the primary Full-reference (FR) image quality assessment measure we considered in our experiments. Unlike in Structural Similarity Index Measure (SSIM), where variation in luminance, contrast and structure of “single-scale” input images are compared, MS-SSIM repeatedly down-samples the input images up to M scales. At each scale, the contrast comparison and the structure comparison are calculated. The luminescence comparison is computed only at scale M, and the final MS-SSIM evaluation is obtained by combining the measurements at different scales [39]. The value of the MS-SSIM measure is bounded between 0 and 1, where 0 represents the worst quality and 1 the best quality. Local Feature Based Visual Security (LFBVS) [40] is the second Full-reference (FR) image quality assessment measure we considered in our experiments. The algorithm utilizes localized edge and luminance features which are combined and weighted according to error magnitude, i.e., error pooling. The value of the LFBVS measure is bounded between 0 and 1, where 0 represents the best quality and 1 the worst quality. We further used Blind/Reference-less Image Spatial Quality Evaluator (BRISQUE) [41] as our No-Reference image quality assessment measure. BRISQUE does not compute distortion-specific features, such as ringing, blur, or blocking, but instead uses scene statistics of locally normalized luminance coefficients to quantify possible losses of “naturalness” in the image due to the presence of distortions, thereby leading to a holistic measure of quality. The underlying features used derive from the empirical distribution of locally normalized luminances and products of locally normalized luminances under a spatial natural scene statistic model. The value of the BRISQUE measure is bounded between 0 and 100, where 0 represents the best quality and 100 the worst quality. In the recognition experiments, the Equal Error Rate (EER) was chosen as an overall measure of biometric recognition performance. EER is the operation point on the receiver operating characteristic curve where the false non-match rate and the false match rate are equal.

4.4. Recognition Pipeline

In our recognition pipeline, we used the contrast adjusted Hough transform (CAHT) [42], and Osiris [43] for iris segmentation, local Gabor filters (LG) for feature extraction, and the Hamming distance with rotation correction for matching. Apart from the Osiris, we used the implementations provided in the USIT toolkit of the University of Salzburg [44].

5. Experiments and Analysis

Addressing the input size requirement of the deep-learning-based models and also a fair evaluation policy, we rescaled all the images in our datasets to the size: 256 × 512 during the compression evaluation experiments. Since the networks are trained on RGB format we cloned each image two times to generate 3 channel (RGB) images ( 256 × 512 × 3 ). We applied a cross-fold scheme to train the (deep-learning-based) models. For this, first we partitioned each dataset into two equal parts and trained the models on one partition and tested it on the other partition. We switched the partitions role next, and doing so we tested the networks on all samples in each dataset without overlapping the training and testing sets. For the DSSLIC model, we set the down-scaling factor α = 8 to obtain the compact representation of the inputs. Table 1 summarizes the training parameters used for each model. Likewise, we applied the other comparison (non-learning-based) compression algorithms to our datasets. To address the preset bandwidth/storage compression limit requirement we defined two bandwidth limits of 0.30 (A) and 0.60 (B), corresponding to the higher and the lower compression levels, respectively, for each dataset in terms of bit-per-pixel (bpp). Obviously, not all algorithms allow to set the exact output file size. Thus, we set the compression parameter for each algorithm in way that the achieved bpp of the resulting compressed images are equal to or less than the predefined bandwidth/storage limit. It is also important to note that the resulting file sizes using the DSSLIC model are among the smallest in the majority of cases (i.e., Casia4i, Casia5a-A, IITD-B, and Notredame). Table 2 shows the selected compression parameters (par) and the resulting bpps per algorithm and dataset. Furthermore, samples of the output (compressed) images in each dataset using the compression methods used are presented in Figure 3 per column and row, respectively.
Table 3 presents the quality compression results in terms of MS-SSIM measure for each dataset (averaged over all images) using the different compression algorithms. As it can be seen, the DSSLIC model shows superior performance over all other algorithms for both compression levels considered. This is a quite interesting result given the fact that the files produced by the DSSLIC are among the smallest files produced by the methods applied. Visual inspection of the obtained output iris images as presented in Figure 3 (the first row) also shows that the model is able to preserve the key iris biometric traits and the structural features very well. Across all datasets, and both compression settings, BPG is (almost) always the second-best. The performance of the other six algorithms vary depending on the datasets: The VCC algorithm shows relatively better performance (compared to the other five algorithms) on the Casia4i, Notredame, and IITD datasets (only when the higher compression rate (A) is considered), while the AV1 algorithm demonstrates overall a better performance than VCC on the Casia5a dataset. EOIC performs better (than VCC and AV1) on Notredame datasets (only when the lower compression rate (B) is considered). The other algorithms (J2K, HEVC, and JPEG) come thereafter while their performance in the majority of cases is lower than the other algorithms across different datasets. For the sake of better interpretation, we visualized the corresponding performance in the form of bar-graphs (for each compression level) in Figure 4.
Table 4 shows the corresponding results for each dataset (averaged over all images) in terms of LFBVS. The superior performance of DSSLIC over the other algorithms is visible in these results too. However, when it comes to the other algorithms’ performance, the MS-SSIM experiment rankings do not apply any more. All together HEVC, VCC, and AV1, rank the second best, showing very close performance to each other. Among the other four algorithms: BPG and J2K come thereafter showing better performance than JPEG in the majority of the cases. Yet, their performance seems to vary depending on the compression level applied and across different datasets. For instance, BPG shows better performance (than J2K) when considering IITD dataset or the lower compression rate (B) on the Casia4i dataset, while J2K shows better or at least equal performance in all other cases. At the end comes EOIC which shows the worst performance almost in all cases. Figure 5 demonstrates the algorithms’ performance for each compression level in the form of bar-graphs.
In order to provide an objective insight into the quality of images generated, we calculated the BRISQUE measure for the obtained compressed images. Table 5 shows the corresponding results for each dataset (averaged over all images) in terms of BRISQUE. The dominating performance of DSSLIC over the other algorithms is visible in these results also. When it comes to the other algorithms’ performance, the LFBVS experiment performance rankings rather applies: EOIC again shows the worst performance almost in all the cases, and all together HEVC, VCC, and AV1, rank the second best, showing very close performance to each other. Among the other three algorithms: BPG comes next showing better performance than the other two algorithms (J2K and JPEG) in the majority of the cases (excluding the Lower compression level (B) for the Casia5a, Notredame, and IITD datasets), and J2K and JPEG together come at the end, showing rather similar performance. Figure 6 demonstrates the algorithms’ performance for each compression level in the form of bar-graphs as well. Unsurprisingly, in all the compression evaluation experiments (MS-SSIM, LFBVS, and BRISQUE), the higher compression rate (reflected in the left graphs in the corresponding figures) decreased the compression performance over all the datasets and algorithms (including DSSLIC).
In the next stage of our experiments, we applied the biometric recognition pipeline (as described in Section 4) to all the compressed images obtained, and evaluated the biometric comparison accuracy, in terms of EER, for the two levels of compression. In addition to the two segmentation algorithms (Osiris and CAHT) used, we considered an optimal segmentation configuration as well, utilizing the manually annotated segmentation drop masks. The optimal segmentation configuration was used to disentangle the distorting effect of the compression on the iris unique biometric traits (which directly affect their subsequent recognition performance) from the iris structural feature distortions which may cause the segmentation failures. Table 6 and Table 7 show the recognition results using the CAHT segmentation and manual segmentation, respectively. As it can be seen in Table 6, recognition does not work at all for the Casia5a and Notredame datasets when using the CAHT segmentation algorithm. The DSSLIC compression shows the best performance only on the IITD and Casia4i datasets (when higher compression rate (A) considered). However, when we use the manual segmentations (Table 7): while the recognition still does not work for Notredame data, yet for the remaining datasets, DSSLIC results are never surpassed by any other comparison compression algorithm. Given the fact that the DSSLIC also produces the smallest actual files, these results imply that DSSLIC compression is able to preserve the iris unique biometric traits bests-certainly better than the other comparison algorithms, as the segmentation defects were ruled out using the manual segmentations.
Table 8 demonstrates the results when applying the OSIRIS algorithm for segmentation. As it can be observed, recognition on the Notredame data does not work either, but otherwise the ranking of the algorithms is fairly different. DSSLIC no longer performs best in any case. If we compare the results obtained using CAHT and OSIRIS segmentation algorithms, we can easily interpret that the segmentation methods and the logic behind them can react quite differently to the artifacts introduced to the image during the compression process, and thus deliver very different results having identically compressed iris images as the input. Overall, the clearly higher compression performance of the DSSLIC algorithm is not directly translated into best recognition accuracy, except in the configuration where the manual segmentation is used. It should be noted that by using the manual drop masks, the measures are only impacted by the possible artifacts introduced by the compression to the iris texture, whereas when using CAHT and OSIRIS as the segmentation modules, the measures are impacted by both the eventual segmentation failures introduced by the segmentation modules, as well as the artifacts introduced by the compression.
Concerning the other deep-learning-based algorithm (EOIC) performance: The algorithm proved (based on the results obtained) to perform better than the CPDIC model which we used in our previous work [1]. The analysis of the output images generated by the CPDIC model (see Figure 7) in our previous work revealed some artifacts which were distributed uniformly over all the images in a block-shaped pattern. These artifacts were more severe and intense in the high texture areas, specifically the iris texture areas. The persistence of these artifacts over all images clearly undermined their recognition performance. When inspecting the iris images obtained using the EOIC model (see Figure 8), we can clearly observe that such artifacts do not exist, and thus the perceptual quality of the images generated are much better than those generated by CPDIC model. Yet, we should note that the iris unique biometric traits seem not to be well preserved and reconstructed in the generated images compared to the images generated by, e.g., DSSLIC. This in fact directly affects the the corresponding recognition results as reflected in Table 7.
We further analyzed the distribution of the genuine and impostor scores obtained during the recognition experiments, to provide a better understanding of how the quality of biometric traits in the compressed images can affect their actual recognition performance. Figure 9 demonstrates the genuine and impostor distributions for the different compression methods for each dataset when considering the optimal segmentation configuration (to exclude the influence of probable segmentation errors). Each pair of curves (genuine and impostor) are indicated by color while line-type distinguishes between impostor (dashed) and genuine (solid). As it can be seen, the impostor curves remain virtually unchanged, while the genuine curves fluctuate almost in all cases. This leads us to the observation that the compression process affects the genuine scores, by introducing artifacts into the iris images which subsequently alter the distinct patterns that are present in the genuine samples, making the compressed images more dissimilar, and thus deteriorating the overall EER scores at the end.

6. Conclusions

We investigated the performance of a deep-learning-based image compression model (DSSLIC) along with some other well-known models in terms of rate-distortion and the subsequent recognition accuracy. The model showed superior compression performance over all other algorithms using different datasets and compression rates. Unlike the other algorithms, the DSSLIC model was able to cope with iris images with complex feature characteristics, and possessed stable performance on all different types of iris data. Visual inspection of the iris images obtained using the other comparison deep-learning-based model (EOIC) showed that: while DSSLIC model possesses better preforming profile than the CPDIC (as used in our previous work), yet it is not able to well preserve and reconstruct the iris unique biometric traits in the images, resulting in lower recognition performance (compared to the DSSLIC model). Furthermore, the experiments with different segmentation algorithms revealed that the segmentation technique and the logic used in it could react quite differently to the compressed images, which in the case of, e.g., the Osiris algorithm caused considerable degradation of the segmentation performance. The results obtained using the manual drop masks supported this argument too. To this extent, as the recognition experiments results showed, the higher compression performance of the DSSLIC algorithm was directly translated into better recognition rates only when the affecting role of the segmentation failures were precluded with the help of the manual drop masks. The experiments also showed that an increase in compression results in reduction of recognition performance in the majority of cases. Analysis of the genuine and the impostor scores indicated that compression introduces artifacts into the iris images which alter the iris unique biometric traits that are present in the genuine samples, making the compressed images more dissimilar. Overall, the results showed that the presented deep-learning-based model is capable of efficient iris image compression, and can be used in an iris biometric recognition system efficiently.

Author Contributions

Conceptualization, E.J. and A.U.; Formal analysis, E.J. and H.H.; Funding acquisition, A.U.; Investigation, H.H.; Methodology, E.J.; Software, E.J.; Supervision, A.U.; Validation, H.H.; Visualization, E.J.; Writing—original draft, E.J.; Writing—review & editing, E.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly funded from the FWF project “Tools for the Generation of Synthetic Biometric Sample Data” grant number I4272.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jalilian, E.; Hofbauer, H.; Uhl, A. Deep Iris Compression. In Pattern Recognition. ICPR International Workshops and Challenges, Proceedings, Part V; Springer: Milan, Italy, 2021; Volume 12665, pp. 1–15. [Google Scholar]
  2. Akbari, M.; Liang, J.; Han, J. Dsslic: Deep semantic segmentation-based layered image compression. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Brighton, UK, 12–17 May 2019; pp. 2042–2046. [Google Scholar]
  3. Matschitsch, S.; Tschinder, M.; Uhl, A. Comparison of compression algorithms’ impact on iris recognition accuracy. In International Conference on Biometrics; Springer: Berlin/Heidelberg, Germany, 2007; pp. 232–241. [Google Scholar]
  4. Daugman, J.; Downing, C. Effect of severe image compression on iris recognition performance. IEEE Trans. Inf. Forensics Secur. 2008, 3, 52–61. [Google Scholar] [CrossRef] [Green Version]
  5. Grother, P. Quantitative standardization of iris image formats. In Biometrics and Electronic Signatures (BIOSIG); Gesellschaft für Informatik e.V.: Bonn, Germany, 2009. [Google Scholar]
  6. Ives, R.; Bishop, D.; Du, Y.; Belcher, C. Iris recognition: The consequences of image compression. EURASIP J. Adv. Signal Process. 2010, 2010, 680845. [Google Scholar] [CrossRef] [Green Version]
  7. Horvath, K.; Uhl, H.S.A.; Weinhandel, G. Lossless compression of polar iris image data. In Iberian Conference on Pattern Recognition and Image Analysis; Springer: Berlin/Heidelberg, Germany, 2011; pp. 329–337. [Google Scholar]
  8. Bergmüller, T.; Christopoulos, E.; Fehrenbach, K.; Schnöll, M.; Uhl, A. Recompression effects in iris recognition. Image Vis. Comput. 2017, 58, 142–157. [Google Scholar] [CrossRef]
  9. Rathgeb, C.; Uhl, A.; Wild, P. Effects of severe image compression on iris segmentation performance. In Proceedings of the IEEE International Joint Conference on Biometrics, Clearwater, FL, USA, 29 September–2 October 2014; pp. 1–6. [Google Scholar] [CrossRef]
  10. Toderici, G.; Malley, S.M.; Hwang, S.J.; Vincent, D.; Minnen, D.; Baluja, S.; Covell, M.; Sukthankar, R. Variable Rate Image Compression with Recurrent Neural Networks. arXiv 2016, arXiv:1511.06085. [Google Scholar]
  11. Ballé, J.; Laparra, V.; Simoncelli, E.P. End-to-end optimized image compression. arXiv 2017, arXiv:1611.01704. [Google Scholar]
  12. Theis, L.; Shi, W.; Cunningham, A.; Huszár, F. Lossy Image Compression with Compressive Autoencoders. arXiv 2017, arXiv:1703.00395. [Google Scholar]
  13. Agustsson, E.; Mentzer, F.; Tschannen, M.; Cavigelli, L.; Timofte, R.; Benini, L.; Van-Gool, L. Soft-to-Hard Vector Quantization for End-to-End Learned Compression of Images and Neural Networks. arXiv 2017, arXiv:1704.00648. [Google Scholar]
  14. Jiang, F.; Tao, W.; Liu, S.; Ren, J.; Guo, X.; Zhao, D. An end-to-end compression framework based on convolutional neural networks. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 3007–3018. [Google Scholar] [CrossRef] [Green Version]
  15. Johnston, N.; Vincent, D.; Minnen, D.; Covell, M.; Singh, S.; Chinen, T.; Jin, S.H.; Shor, J.; Toderici, G. Improved lossy image compression with priming and spatially adaptive bit rates for recurrent networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4385–4393. [Google Scholar]
  16. Li, M.; Zuo, W.; Gu, S.; Zhao, D.; Zhang, D. Learning convolutional networks for content-weighted image compression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3214–3223. [Google Scholar]
  17. Luo, S.; Yang, Y.; Yin, Y.; Shen, C.; Zhao, Y.; Song, M. DeepSIC: Deep semantic image compression. In International Conference on Neural Information Processing; Springer: Cham, Switzerland, 2018; pp. 96–106. [Google Scholar]
  18. Mentzer, F.; Agustsson, E.; Tschannen, M.; Timofte, R.; Gool, L.V. Conditional Probability Models for Deep Image Compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4394–4402. [Google Scholar]
  19. Wang, C.; Han, Y.; Wang, W. An End-to-End Deep Learning Image Compression Framework Based on Semantic Analysis. Appl. Sci. 2019, 9, 3580. [Google Scholar] [CrossRef] [Green Version]
  20. Santurkar, S.; Budden, D.; Shavit, N. Generative Compression. In Proceedings of the 2018 Picture Coding Symposium (PCS), San Francisco, CA, USA, 24–27 June 2018; pp. 258–262. [Google Scholar]
  21. Rippel, O.; Bourdev, L. Real-Time Adaptive Image Compression. In Proceedings of the 34th International Conference on Machine Learning; Proceedings of Machine Learning Research; Precup, D., Teh, Y.W., Eds.; International Convention Centre: Sydney, Australia, 2017; Volume 70, pp. 2922–2930. [Google Scholar]
  22. Agustsson, E.; Tschannen, M.; Mentzer, F.; Timofte, R.; Gool, L.V. Generative adversarial networks for extreme learned image compression. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 221–231. [Google Scholar]
  23. Ballé, J.; Minnen, D.; Singh, S.; Hwang, S.J.; Johnston, N. Variational image compression with a scale hyperprior. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  24. Minnen, D.; Ballé, J.; Toderici, G. Joint Autoregressive and Hierarchical Priors for Learned Image Compression. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
  25. Cheng, Z.; Sun, H.; Takeuchi, M.; Katto, J. Learned Image Compression with Discretized Gaussian Mixture Likelihoods and Attention Modules. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–18 June 2020. [Google Scholar]
  26. Jalilian, E.; Uhl, A. Iris Segmentation Using Fully Convolutional Encoder–Decoder Networks. In Deep Learning for Biometrics; Springer: Cham, Switzerland, 2017; pp. 133–155. [Google Scholar]
  27. Sneyers, J.; Wuille, P. FLIF: Free lossless image format based on MANIAC compression. In Proceedings of the IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016; pp. 66–70. [Google Scholar]
  28. University of Notre Dame. Available online: https://cvrl.nd.edu/projects/data (accessed on 10 March 2022).
  29. BIT. Available online: http://biometrics.idealtest.org (accessed on 10 March 2022).
  30. Publicly and Freely Available Biometrics Databases. Available online: http://www4.comp.polyu.edu.hk/~csajaykr/database.php (accessed on 10 March 2022).
  31. The MathWorks, Inc. Available online: https://www.mathworks.com/help/matlab/ref/imwrite.html (accessed on 10 March 2022).
  32. BPG. Available online: https://bellard.org/bpg (accessed on 10 March 2022).
  33. Sullivan, G.J.; Ohm, J.R.; Han, W.J.; Wiegand, T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  34. High Efficiency Video Coding (HEVC). Available online: https://hevc.hhi.fraunhofer.de (accessed on 10 March 2022).
  35. Bross, B.; Liu, S.; Chen, J.; Wang, Y.K. Versatile Video Coding (VVC); Joint Video Exploration Team (JVET). 2021. Available online: https://jvet.hhi.fraunhofer.de (accessed on 10 March 2022).
  36. Video Coding Standardization GitLab. Available online: https://vcgit.hhi.fraunhofer.de/jvet/VVCSoftwareVTM (accessed on 10 March 2022).
  37. Chen, Y.; Murherjee, D.; Han, J.; Grange, A.; Xu, Y.; Liu, Z.; Parker, S.; Chen, C.; Su, H.; Joshi, U. An overview of core coding tools in the AV1 video codec. In Proceedings of the Picture Coding Symposium (PCS), San Francisco, CA, USA, 24–27 June 2018; pp. 41–45. [Google Scholar]
  38. aom. Available online: https://aomedia.googlesource.com/aom/ (accessed on 10 March 2022).
  39. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Monterey, CA, USA, 9–12 November 2003; Volume 2, pp. 1398–1402. [Google Scholar]
  40. Tong, L.; Dai, F.; Zhang, Y.; Li, J. Visual Security Evaluation for Video Encryption; Association for Computing Machinery: New York, NY, USA, 2010. [Google Scholar] [CrossRef]
  41. Mittal, A.; Moorthy-Krishna, A.; Conrad-Bovik, A. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef] [PubMed]
  42. Rathgeb, C.; Uhl, A.; Wild, P. Iris Recognition: From Segmentation to Template Security. In Advances in Information Security; Springer: Berlin/Heidelberg, Germany, 2013; Volume 59. [Google Scholar]
  43. Othman, N.; Dorizzi, B.; Salicetti, S.G. OSIRIS: An open source iris recognition software. Pattern Recognit. Lett. 2016, 82, 124–131. [Google Scholar] [CrossRef]
  44. Rathgeb, C.; Uhl, A.; Wild, P.; Hofbauer, H. Design Decisions for an Iris Recognition SDK. In Handbook of Iris Recognition, 2nd ed.; Bowyer, K., Burge, M.J., Eds.; Advances in Computer Vision and Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar] [CrossRef]
Figure 1. An iris image (a) and its corresponding output using a GAN based model (b), along with an image visualizing their difference (c), and the overlaid ground-truth (d).
Figure 1. An iris image (a) and its corresponding output using a GAN based model (b), along with an image visualizing their difference (c), and the overlaid ground-truth (d).
Sensors 22 02698 g001
Figure 2. The deep-learning-based iris compression model.
Figure 2. The deep-learning-based iris compression model.
Sensors 22 02698 g002
Figure 3. Samples of highly (A) compressed iris images from the Casia4i, Casia5a, IITD, and Notredame datasets, per column, respectively, using DSSLIC, EOIC, BPG, J2K, JPEG, HEVC, VVC, and AV1 algorithms per row, respectively.
Figure 3. Samples of highly (A) compressed iris images from the Casia4i, Casia5a, IITD, and Notredame datasets, per column, respectively, using DSSLIC, EOIC, BPG, J2K, JPEG, HEVC, VVC, and AV1 algorithms per row, respectively.
Sensors 22 02698 g003
Figure 4. Compression performance in terms of MS-SSIM for the high (left graph) and low (right graph) compression levels.
Figure 4. Compression performance in terms of MS-SSIM for the high (left graph) and low (right graph) compression levels.
Sensors 22 02698 g004
Figure 5. Compression performance in terms of LFBVS for the high (left graph) and low (right graph) compression levels.
Figure 5. Compression performance in terms of LFBVS for the high (left graph) and low (right graph) compression levels.
Sensors 22 02698 g005
Figure 6. Compression performance in terms of BRISQUE for the high (left graph) and low (right graph) compression levels.
Figure 6. Compression performance in terms of BRISQUE for the high (left graph) and low (right graph) compression levels.
Sensors 22 02698 g006
Figure 7. Sample compressed images in the Casia5a dataset (a) and Notredame datasets (b) and their corresponding normalized iris images (c,d), respectively, both generated by CPDIC algorithm.
Figure 7. Sample compressed images in the Casia5a dataset (a) and Notredame datasets (b) and their corresponding normalized iris images (c,d), respectively, both generated by CPDIC algorithm.
Sensors 22 02698 g007
Figure 8. Sample images in the Casia5a (a) and Notredame (b) datasets, and their corresponding normalized iris images (c,d), along with their compressed images (e,f) and their corresponding normalized versions (g,h), respectively, using EOIC algorithm.
Figure 8. Sample images in the Casia5a (a) and Notredame (b) datasets, and their corresponding normalized iris images (c,d), along with their compressed images (e,f) and their corresponding normalized versions (g,h), respectively, using EOIC algorithm.
Sensors 22 02698 g008
Figure 9. Genuine and impostor distributions of the different compression methods.
Figure 9. Genuine and impostor distributions of the different compression methods.
Sensors 22 02698 g009
Table 1. Networks’ training parameters.
Table 1. Networks’ training parameters.
NetworkDSSLICEOIC
OptimizerAdamAdam
Learning rate0.00020.0001
Momentum--
Weight decay0.00050.1
Iteration2501,000,000
Table 2. Selected compression parameters (par) and their corresponding compression performance in bits per pixel (bpp) for each algorithm.
Table 2. Selected compression parameters (par) and their corresponding compression performance in bits per pixel (bpp) for each algorithm.
DatasetCasia4iCasia5aIITDNotredame
Methodpar bppparbppparbppparbppparbppparbppparbppparbpp
 DSSLIC230.20160.44230.16140.45270.30190.53230.16140.51
 EOIC10240.2140960.4715360.1676800.432560.2817920.535120.1651200.53
 BPG370.21300.54300.19240.42330.29260.60330.18240.55
 J2K350.23210.55450.18140.55280.27140.55450.18140.55
 JPEG230.20570.50120.19570.51170.30570.53090.18570.58
 HEVC320.24270.44280.18200.44360.30200.56320.21250.51
 VCC310.23250.48300.16220.51340.31280.57320.17250.51
 AV1310.22250.44300.20130.55420.30300.58350.20200.51
Bpp (max)A (0.30)B (0.60)A (0.30)B (0.60)A (0.30)B (0.60)A (0.30)B (0.60)
Table 3. Average MS-SSIM scores using high (A) and low (B) compression levels.
Table 3. Average MS-SSIM scores using high (A) and low (B) compression levels.
DatasetCasia4iCasia5aIITDNotredame
LevelBABABABA
DSSLIC0.9980.9940.9950.9900.9980.9980.9970.998
EOIC0.9930.9820.9940.9820.9850.9610.9940.978
BPG0.9960.9880.9940.9850.9970.9920.9960.988
J2K0.9910.9660.9920.9700.9870.9450.9880.964
JPEG0.9930.9500.9880.9310.9940.9570.9910.949
HEVC0.9900.9790.9930.9830.9900.9770.9910.980
VCC0.9940.9820.9950.9840.9930.9820.9920.980
AV10.9910.9780.9950.9850.9920.9780.9920.980
Table 4. Average LFBVS scores using high (A) and low (B) compression levels.
Table 4. Average LFBVS scores using high (A) and low (B) compression levels.
DatasetCasia4iCasia5aIITDNotredame
LevelBABABABA
DSSLIC0.120.150.110.160.130.150.140.19
EOIC0.320.410.370.450.320.380.360.44
BPG0.180.310.250.340.180.260.240.35
J2K0.190.290.180.340.200.270.230.35
JPEG0.190.410.210.450.200.310.210.44
HEVC0.130.260.210.330.150.210.220.31
VCC0.130.250.170.330.130.200.220.32
AV10.160.280.170.320.140.210.210.32
Table 5. Average BRISQUE scores using high (A) and low (B) compression levels.
Table 5. Average BRISQUE scores using high (A) and low (B) compression levels.
DatasetCasia4iCasia5aIITDNotredame
LevelBABABABA
DSSLIC29.030.426.429.435.836.119.723.4
EOIC44.850.852.060.944.844.447.656.0
BPG33.639.840.246.243.344.834.443.4
J2K37.248.039.654.944.449.137.253.7
JPEG36.058.135.461.939.748.531.257.6
HEVC26.736.636.245.737.141.331.140.4
VCC29.136.133.747.537.543.130.142.2
AV129.839.032.947.038.444.830.042.4
Table 6. EERs for the different datasets using the CAHT algorithm.
Table 6. EERs for the different datasets using the CAHT algorithm.
DatasetCasia4iCasia5aIITDNotredame
LevelBABABABA
DSSLIC1.21.021.121.21.41.829.929.9
EOIC1.21.423.625.02.82.330.938.8
BPG1.01.221.621.31.62.429.630.3
J2K1.11.320.622.32.02.630.030.1
JPEG1.22.820.626.11.92.529.932.4
HEVC0.91.017.319.11.91.830.530.5
VCC0.71.017.018.61.42.330.129.9
AV10.91.117.418.51.62.029.730.0
Table 7. EERs for the different datasets using manual masks.
Table 7. EERs for the different datasets using manual masks.
DatasetCasia4iCasia5aIITDNotredame
LevelBABABAAB
DSSLIC0.40.42.52.90.40.523.823.9
EOIC0.50.73.56.70.50.624.339.8
BPG0.40.62.93.90.40.523.823.9
J2K0.40.62.75.10.40.523.824.0
JPEG0.51.73.014.00.40.523.825.7
HEVC0.40.52.93.90.40.524.223.9
VCC0.50.52.84.10.40.523.924.0
AV10.50.52.83.80.50.524.024.2
Table 8. EERs for different the datasets using the Osiris algorithm.
Table 8. EERs for different the datasets using the Osiris algorithm.
DatasetCasia4iCasia5aIITDNotredame
LevelBABABAAB
DSSLIC1.11.02.02.20.70.825.225.5
EOIC0.90.92.85.20.50.425.326.1
BPG0.91.02.02.50.30.326.926.4
J2K0.80.92.03.10.40.725.725.1
JPEG0.81.82.49.70.50.624.724.7
HEVC0.70.91.92.60.80.825.725.8
VCC0.90.92.02.40.91.225.725.5
AV10.90.91.92.21.21.225.825.9
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jalilian, E.; Hofbauer, H.; Uhl, A. Iris Image Compression Using Deep Convolutional Neural Networks. Sensors 2022, 22, 2698. https://doi.org/10.3390/s22072698

AMA Style

Jalilian E, Hofbauer H, Uhl A. Iris Image Compression Using Deep Convolutional Neural Networks. Sensors. 2022; 22(7):2698. https://doi.org/10.3390/s22072698

Chicago/Turabian Style

Jalilian, Ehsaneddin, Heinz Hofbauer, and Andreas Uhl. 2022. "Iris Image Compression Using Deep Convolutional Neural Networks" Sensors 22, no. 7: 2698. https://doi.org/10.3390/s22072698

APA Style

Jalilian, E., Hofbauer, H., & Uhl, A. (2022). Iris Image Compression Using Deep Convolutional Neural Networks. Sensors, 22(7), 2698. https://doi.org/10.3390/s22072698

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop