Next Article in Journal
Research on IEEE 802.11 OFDM Packet Detection Algorithms for Household Wireless Sensor Communication
Next Article in Special Issue
Digital Watermarking System for Copyright Protection and Authentication of Images Using Cryptographic Techniques
Previous Article in Journal
Connectivity Alterations in Vascular Parkinsonism: A Structural Covariance Study
Previous Article in Special Issue
Blockchain and Cloud to Overcome the Problems of Buyer and Seller Watermarking Protocols
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluation of a Framework for Robust Image Reversible Watermarking

by
Jose Juan Garcia-Hernandez
1,*,
Claudia Feregrino-Uribe
2,
Alejandra Menendez-Ortiz
2 and
Dan Williams Robledo-Cruz
1
1
Cinvestav Unidad Tamaulipas, Parque Cientifico y Tecnologico TECNOTAM, Km. 5.5 Carr. a Soto la Marina, Ciudad Victoria CP 87130, Mexico
2
Coordinacion de Ciencias Computacionales, INAOE Luis Enrique Erro #1, Sta. Ma. Tonantzintla, Puebla CP 72840, Mexico
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(14), 7242; https://doi.org/10.3390/app12147242
Submission received: 14 June 2022 / Revised: 6 July 2022 / Accepted: 8 July 2022 / Published: 18 July 2022

Abstract

:
In the literature, robust reversible watermarking schemes (RWSs) allow the extraction of watermarks after the images have suffered attacks; however, the modified images are compromised. On the other hand, self-recovery schemes will restore the compromised regions of the images, but no secret messages are inserted in these schemes. A framework for robust reversible watermarking with signal restoration capabilities was previously proposed in the literature. This study selects four fragile reversible watermarking techniques and two self-recovery schemes to design different framework configurations. These configurations are evaluated to test the framework’s performance and determine the structure that yields better results in terms of perceptual transparency using a well-known image database as the signal input. It was found that fragile reversible watermarking schemes hold low perceptual distortion, while self-recovery schemes produce high perceptual distortion levels. The inherent characteristics of each algorithm determine, a priori, the behavior of the framework, which is approximated by a proposed equation.

1. Introduction

Due to available fast Internet connections, massive access to multimedia material is common currently. However, this ease of connection produces some risks in the reliability of the accessed material. Although different strategies have been proposed, in particular, digital watermarking has been shown to be an appropriate option for the protection of multimedia content. Despite its protection capabilities, conventional digital watermarking produces distortion in the host signal. These distortions are not acceptable for critical applications such as medical or military imaging [1]. Reversible watermarking schemes (RWSs) emerge as an option to overcome the distortion induced by classical marking schemes. RWSs allow the host signal to be reconstructed to its original state after removing the inserted watermark.
On the other hand, most of the RWSs are not very robust to attacks. If a tagged signal is attacked, the recovery of the information and the reconstruction of the host signal will be very difficult. Recently, some robust reversible watermarking schemes have been proposed in the literature. Most of these schemes are deployed in the spatial domain and are robust to JPEG compression and the addition of Gaussian noise. Some examples of this kind of scheme are the ones presented by [2,3]. Furthermore, there are schemes that use the frequency domain for the insertion of the watermark. As is well known, the frequency domain is usually more robust to attacks than the spatial domain. There are some examples of such schemes in the literature, such as those proposed in [4,5,6]. Typically, these schemes are robust to attacks such as clipping, noise addition, scaling, and histogram modification. Furthermore, RWSs for medical images have been recently proposed under a hybrid approach such as in [7].
An interesting watermarking application is known as signal authentication, where a fragile watermark is used to identify and localize tampering in the attacked signals. A representative fragile watermarking scheme for medical images is reported in [8]. Due to the need to restore the attacked signals, beyond identifying the attack, self-recovery schemes have emerged. In [9], the idea was proposed of embedding an image into itself to restore the tampered regions; this is the first self-recovering scheme proposed in the literature. The steps typically followed by self-recovering schemes are: obtain a compressed version of the host signal and insert this version into the host signal itself. Upon detection, using a fragile watermark, of an attacked region, it is possible to restore it using the compressed version originally inserted into the host signal. It has been observed in the literature that there are two types of results delivered by self-recovery schemes: approximate recovery and perfect recovery. The former refers to the ability to recover the attacked signal to levels of high similarity with the original one, while perfect recovery refers to the fact that the recovered signal will be exactly the same as the original signal.
Self-recovery schemes for audio, image, or video signals have been proposed in the literature in recent years. In general, the strategies deal with modifying the insertion domain or improving the compression process required for watermark insertion. Some relevant proposals that insert the watermark in the spatial domain are [10,11,12,13]; similarly, the use of the frequency domain as the insertion space was reported in [14,15,16]. The use of the two domains, spatial and temporal, has also been explored by utilizing video signal as the host in [17,18]. Reference [19] proposed a scheme for video signals in the DCT domain, and it resists MPEG attacks. However, self-recovering schemes are limited in their additional payload to the compressed version of the host signal that is used as a watermark. If additional information is inserted, the distortion caused can be prohibitive for real-world applications. In [20], a framework was first introduced; the idea behind this framework is the combination of a fragile reversible watermarking scheme that allows the insertion of a useful payload to the image and a self-recovery scheme to insert information that can aid in the recovery of tampered regions of the images. By combining these schemes, the framework achieves adequate transparency for the insertion of both control information and useful payload. At the same time, it maintains the capacity of restoring the areas tampered with content replacement. This approach is different from the framework proposed by Coltuc in [21], where the first stage is a robust watermarking algorithm followed by a reversible watermarking algorithm. Coltuc’s algorithm has inspired other works recently, such as [22]; however, the perceptual distortion induced by that framework has been studied as a whole; thus, the contribution of each stage is little known.
In this manuscript, an evaluation of the framework proposed in [20] is presented. The main goal is to observe the perceptual distortion due to the framework and how it is related to distortion due to both reversible and self-recovery stages. The main contribution of this study is to determine the perceptual transparency of the framework from the knowledge of each of the two algorithms selected for a given configuration. No more experimentation of the whole framework will be needed for each new algorithm used to know its perceptual transparency.
The rest of the document is organized in the following way. In Section 2, the framework proposed in [20] among several reversible watermarking and self-recovery schemes is described. Numerical results and discussions are given in Section 3. Finally, Section 4 concludes this paper.

2. Evaluation of the Robust Framework for Robust Reversible Image Watermarking

The framework proposed by [20] was taken as a general outline to be used with different combinations of fragile reversible and self-recovery schemes to determine which combination obtains better robustness results and transparency. The idea behind this framework is detailed below.

2.1. Description of the Framework

In [20], a framework is proposed to construct a robust reversible watermarking scheme with signal restoration. Figure 1 shows the framework, which consists of two processes: encoding and decoding. Each process consists of a fragile stage and a self-recovery stage. Figure 2 shows a general fragile reversible watermarking scheme. Secret message M is hidden in a host signal X, resulting in a watermarked signal Y transmitted through a noise-free channel. The extraction process requires Y and control data to reconstruct the host signal X, and it also recovers a message M .
On the other hand, Figure 3 shows a general self-recovery scheme that can detect tampered regions and reconstruct them with the embedded control data.
In the encoding process, the fragile embedding stage inserts the secret message m into the host signal x, producing a watermarked signal y. Then, the self-recovery stage embeds control information, which allows the recovery of the signal after a content replacement attack. The result is the protected signal y .
In the presence of a content replacement, attack signal y becomes an attacked signal y ^ . The decoding process receives y ^ , then the restoration stage tries to revert the attack. The self-recovery restoration stage extracts the control information embedded in the encoding process and utilizes it along with the remaining non-attacked regions of the signal to restore the regions altered by the attack. If the attack is limited to a threshold, the watermarked signal y produced by the encoding process is obtained. From the watermarked signal y, the fragile RWS extraction and recovery stage can extract the secret message and recover the original samples from the host signal.

2.2. Selection of Schemes

The framework consists of two stages in each process that use the encoding and decoding algorithms of a fragile reversible watermarking and a self-recovery scheme. There are several fragile reversible watermarking schemes in the literature, but only three were selected to test the framework; the criteria used to choose them were their transparency and their embedding capacity; schemes with transparency over 40 dB and a payload over one bpp were selected. On the other hand, self-recovery schemes are not as typical in the literature; however, self-recovery schemes with perfect restoration capabilities were the ones used for this methodology since that is a requirement for the framework to properly extract the watermarks in the fragile stage.

2.2.1. Fragile Reversible Watermarking Schemes

The scheme by [23] uses a histogram shifting technique, using the high correlation of neighboring pixels in each block of the image; the authors use a threshold K to insert the watermark bits; based on the difference values calculated for each block, the embedding strategy is divided into two categories: positives and negatives. The embedding strategy first divides the image in blocks of size 2 × 2 , 4 × 4 or 8 × 8 pixels. Then, the difference values are calculated using the following equation:
α = 1 n i = 1 n ( a i b i ) ,
where n is the number of pairs of pixels, a i are all the pixels marked with (+), and b i are all the pixels marked with (−). After the difference values have been calculated, a threshold K + is selected using Equations (2) and (3) and a threshold K using Equations (4) and (5):
K p h = α maxP α zero + 1 Partition level ,
K + = K p h + mod ( K p h , 2 ) ,
K N h = α zero α minN + 1 Partition level ,
K = 1 × ( K N h + mod ( K N h , 2 ) ) ,
where α maxP is the maximum difference value of the positive part, α zero is 0 for the center of the coordinate, Partition level is the number of the partition, α minN is the minimum difference value of the negative part, and . is the ceiling function. The watermark bits are inserted based on the values of α , according to the following conditions:
  • If α 0 , there can be three cases:
    Case 1 
    0 α k ; if 1 is inserted, the histogram from that block is shifted a distance k to the right; if a 0 is inserted, the block is left intact.
    Case 2 
    k α 2 k ; if a 1 is inserted, the histogram from that block is shifted a distance 2 k to the right; if a 0 is inserted, the block is shifted a distance k to the right.
    Case 3 
    2 k α 3 k ; if a 1 is inserted, the histogram from that block is shifted a distance 3 k to the right; if a 0 is inserted, the block is shifted a distance 2 k to the right.
  • If α < 0 , there can be three cases:
    Case 1 
    k α 0 ; if 1 is inserted, the histogram from that block is shifted a distance k to the left; if a 0 is inserted, the block is left intact.
    Case 2 
    2 k α k ; if a 1 is inserted, the histogram from that block is shifted a distance 2 k to the left; if a 0 is inserted, the block is shifted a distance k to the left.
    Case 3 
    3 k α 2 k ; if a 1 is inserted, the histogram from that block is shifted a distance 3 k to the left; if a 0 is inserted, the block is shifted a distance 2 k to the left.
To extract the watermark bits, the image is divided into blocks in the same way as for embedding. Then, the difference values are calculated using Equation (1). After the difference values are obtained, the threshold K is obtained in the following way:
K = K p H + mod ( K p h , 2 ) ,
where α maxP is the maximum difference value of the positive part, α z e r o is 0 for the center of the coordinate, Partition level is the number of the partition, and . is the ceiling function. Finally, the watermark bit is extracted using the threshold K and the original image is reconstructed.
The scheme by [24] uses a histogram shifting technique as well; it divides the intensity range into non-overlapped segments and seeks the peak pixel, i.e., the pixel with a more significant occurrence in the histogram of each segment. Bit insertion is performed only in the peak pixels, except for each segment’s first peak pixel, which is a reference for extraction. These schemes use a location map where all the peak pixels are marked. The embedding process first divides the intensity levels of the image into non-overlapped segments of equal size. Then, it generates a histogram for each block and identifies the intensity value of the peak pixel in each segment. The pixels from the original image are read, and then, k bits from the secret message M are embedded using a pixel substitution strategy on the peak pixels. Afterward, a location map is created, where a 1 is inserted if the pixel is a peak pixel or 0 otherwise. For the extraction process, the intensity values of the image are divided into non-overlapped segments of equal size in the same way as in the embedding process. The watermarked image and the location map are read. The peak pixels of each reference segment are obtained; then the watermark bits are extracted, and the current pixel is replaced with the reference peak pixel.
The scheme proposed in [25] expands the image size using an interpolation method that uses trigonometric functions; watermark bits are inserted in the interpolated pixels; this algorithm uses a location map. The extraction process recovers the watermark and eliminates the pixels generated by interpolation during embedding; therefore, the original image can be recovered. The embedding process first divides the host image into blocks of 2 × 2 pixels. Each block is transformed into blocks of 3 × 3 pixels through interpolation, using the following equations:
B i ( 1 , 1 ) = B o ( 1 , 1 ) B i ( 3 , 3 ) = B o ( 2 , 2 ) B i ( 2 , 2 ) = ( B i ( 1 , 2 ) 2 ) + ( B i ( 3 , 2 ) ) 2 2 B i ( 1 , 3 ) = B o ( 1 , 2 ) B i ( 1 , 2 ) = ( B i ( 1 , 1 ) ) 2 + ( B i ( 1 , 3 ) ) 2 2 B i ( 2 , 3 ) = ( B i ( 1 , 3 ) ) 2 + ( B i ( 3 , 3 ) ) 2 2 B i ( 3 , 1 ) = B o ( 2 , 1 ) B i ( 2 , 1 ) = ( B i ( 1 , 1 ) ) 2 + ( B i ( 3 , 1 ) ) 2 2 B i ( 2 , 3 ) = ( B i ( 3 , 1 ) ) 2 + ( B i ( 3 , 3 ) ) 2 2 ,
where B o is the original image block and B i is the interpolated block. The new values are modified using the following trigonometric functions:
B f ( 1 , 2 ) = B i ( 1 , 2 ) cos ( B i ( 1 , 1 ) + B i ( 1 , 3 ) 2 ) B f ( 2 , 2 ) = B i ( 2 , 2 ) cos ( B i ( 1 , 2 ) + B i ( 3 , 2 ) 2 ) B f ( 2 , 1 ) = B i ( 2 , 1 ) cos ( B i ( 1 , 1 ) + B i ( 3 , 1 ) 2 ) B f ( 2 , 3 ) = B i ( 2 , 3 ) cos ( B i ( 1 , 3 ) + B i ( 3 , 3 ) 2 ) B f ( 3 , 2 ) = B i ( 3 , 2 ) cos ( B i ( 3 , 1 ) + B i ( 3 , 3 ) 2 ) ,
where B f is the block modified by the trigonometric functions. Obtain 5 bits from the watermark M, and insert them in the interpolated pixels. The extraction process divides the image into blocks of size 3 × 3 pixels, called C i . Then, calculate a new block using trigonometric functions, called C i . The watermark bits are extracted by subtracting the two blocks C i C i , and finally, the original image is obtained by eliminating the interpolated pixels.

2.2.2. Self-Recovery Schemes

The scheme by [26] is a self-recovery one, where reference bits and check bits are inserted into the images themselves to restore the original samples and identify the tampered regions. The embedding process uses a difference expansion (DE) strategy to insert both references and check bits in all the blocks of the host image; if the watermark inserted in one region of the image is altered, the rest of the watermark bits are not altered. The image can be recovered using the reference bits extracted from non-tampered regions. The embedding process divides the image into blocks of size 8 × 8 , and then, each block is divided into 16 sub-blocks, where changeable and unchangeable pixels are assigned. For each changeable pixel, the inequality g m ( i , j ) g u is verified:
g u + [ g m ( i , j ) g u ] · 2 + 1 255 .
If g m ( i , j ) < g u , then:
g u + [ g m ( i , j ) g u ] · 2 255 ,
where g m represents the grayscale values of the changeable pixels for each block and g u are the pixels designated as unchangeable of each block. When Equations (7) and (8) are fulfilled, the pixel g m ( i , j ) is deemed as unusable, otherwise as usable. Afterwards, the reference bits are generated, taking as the base the pixels from the original image. Then, the check bits are calculated in order to identify the tampered regions, using a 64 bit hash function. The embedding strategy used is DE using the following equation:
g ˜ m = g u + [ g m ( i , j ) g u ] · 2 + w ,
where w is constructed with the reference and check bits to be embedded. For the extraction process, the image is divided into N / 64 blocks and N / 4 sub-blocks, in the same way as in the embedding process, where N is the total number of pixels in the image. The check and reference bits are extracted to identify the tampered and reserved blocks. Finally, the original grayscale values from all the pixels in the blocks identified as tampered are restored.
The scheme by [27] proposes a secure block mechanism, resilient to content replacement attacks. To locate the tampered blocks, it uses the unaltered pixels and the reference bits to estimate the original five MSB of the altered pixels, using an iterative and exhaustive restoration mechanism. The embedding process divides the image in a pseudo-random manner using a secret key k 1 in a subset of m pixels. The reference bits b r i are calculated, using Equations (10) and (11), and the b r i bits are embedded into the third LSB of the pixels in each subset.
b r i = H ( x ^ i , 1 , , x ^ i , m )
x ^ i , j = x i , j 8 mod 16 , j = 1 , , m ,
where H ( . ) is a hash function, . is the floor function, and x ^ [ 0 , 15 ] is the decimal value of the bits b 4 , , b 7 of each pixel x i , j . A second set of reference bits b r 2 is calculated using Equations (10) and (12). The image is divided into subsets of m pixels, each using a second secret key k 2 . A hash function with the five MSBs of each pixels for every subset is calculated, then b r 2 is embedded into the pixels of each subset.
x ^ i , j = x i , j 8 , j = 1 , , m .
Then, the image is divided into non-overlapped blocks of size 8 × 8 . Afterward, the authentication code c a for each block is calculated using Equation (13), which is inserted in the LSB of every pixel in each block.
c a i = I n 1 n 2 p i ,
where I is an exclusive index associated with each image, p i is the index of the block, and n 1 and n 2 is the bit concatenation of the image size. The extraction process divides the image in non-overlapped blocks of 8 × 8 , as in the embedding process. Then, the authentication code is extracted from the LSB of each block, and the correct authentication code is assigned; the correct authentication code is obtained by majority voting. Afterward, the image is divided pseudo-randomly into subsets of m pixels, using a secret key k 1 . The reference bits b r 1 are restored from bit b 3 , then the four MSB of every altered pixel of every block are calculated, by calculating the test codes c p i = H ( y ^ i , 1 , , y ^ i , m ) , where y ^ i , j = x i , j 8 mod 16 . If y i , j is an altered pixel, then all the possible values of the four bits are exhaustively assigned to y ^ i , j . The test codes and the reference bits are compared to identify the four MSB matches. Then, the four MSB are extended, adding to the beginning a bit 1 and a bit 0, generating two new values of five bits each, called restoration candidates. The image is divided into subsets of m pixels each, but now using the second secret key k 2 , and only one restoration candidate is associated with the altered pixel. Finally, the original image is restored.

2.3. Perceptual Distortion Evaluation

The peak-signal-to-noise ratio (PSNR) measures the similarity of two signals, a reference signal, and a processed version of it, and this ratio is given in decibels [28,29] as some signals hold a very wide dynamic range. The PSNR is a metric complementary to distortion. The higher the PSNR, the lower the distortion is.
In image processing, a typically PSNR is measured in an 8 bit grayscale version. For an image f and its processed version g, both of size M × N , the PSNR between f and g is computed by:
PSNR ( f , g ) = 10 · log 10 255 2 M S E ( f , g )
M S E ( f , g ) = 1 ( M · N ) i = 1 M j = 1 N ( f i j g i j ) 2
As for the mean-squared error (MSE), the difference between the pixels f i j and g i j is considered an error that generates image quality loss. The lower the MSE, the higher the PSNR; therefore, the higher the PSNR(f, g) values, the higher the image quality is. For digital images, sometimes, a PSNR > 35 dB is considered as good quality [30].
The Watson metric, which quantifies the distortion of an image based on just noticeable differences [31], is another tool to evaluate the distortion of a processed image. This metric measures the errors for each DCT coefficient in each block by its corresponding sensitivity threshold. That threshold considers contrast sensitivity, luminance masking, and contrast masking.
In the Watson model, for each i j DCT coefficient, the relation between the luminance and frequency is considered, as follows:
t i j k = t i j ( c 00 k c ¯ 00 ) a t
where t i j is the threshold for the smallest frequency coefficient that yields a visible signal, c 00 k is the DC coefficient of block k, c ¯ 00 is the DC coefficient corresponding to the mean luminance of the display (1024 for an 8 bit image), and a t determines the degree of masking (set to 0.65) [32].
When the visibility of a pattern is reduced by the presence of another pattern in the image, this phenomenon is known as texture masking. Watson extends the results of luminance and frequency masking to include texture masking as follows:
m i j k = max [ t i j k , c i j k w i j t i j k 1 w i j ]
where m i j k is the masked threshold and w i j determines the degree of texture masking. Typically, w 00 = 0 and w i j = 0.7 for all other coefficients. The perceptual error in each frequency of each block is given by
d i j k = e i j k m i j k
where e i j k is the quantization error.
To associate the errors in the model, the Minkowski metric is used as follows:
p i j = ( k d i j k β s ) 1 β s
where different values of the exponent β s implement different types or degrees of pooling. In practice, β s = 100 is commonly used.
A typical threshold for the Watson metric is 0.4, since measures below this point guarantee visual imperceptibility [33].
The following section presents the experimental results obtained after implementing the framework using the fragile reversible watermarking schemes, and self-recovery schemes were selected.

3. Results and Discussions

In this section, framework evaluation results are presented. The evaluation is divided into three stages: first, the fragile reversible watermarking schemes are evaluated in terms of perceptual distortion versus payload; then, the maximum robustness of self-recovery schemes is measured; finally, the framework is evaluated for each possible configuration of reversible watermarking and self-recovery stages.
The six algorithms used in this study were implemented in the MatLab R2014 language, using a computer with an Intel Core i7-1255U processor, 16 GB RAM, 128 GB solid-state disk, and an Nvidia GeForce MX550 2 GB GDDR6 graphics card. Five algorithms were implemented in their sequential form following their description reported in the literature. The algorithm proposed by Bravo-Solorio et al. [27] was optimized using the MatLab parallel computing toolbox due to its very high time complexity. For each experiment, 2000 gray-level images from the BOWS 2 [34] data set were used; each image is 512 × 512 pixels in size.

3.1. Fragile Reversible Watermarking Schemes

Each fragile reversible watermarking scheme was applied to the data set using watermarks of 1000 bits to 10,000 bits with 1000 bits steps. The peak-signal-to-noise ratio (PSNR) and Watson metrics evaluated the algorithm’s distortion for different watermark lengths. Figure 4 shows the performance of the fragile reversible watermarking schemes using the PNSR metric as an evaluation tool. A comparison between fragile reversible watermarking techniques using the Watson metric as an evaluation tool is shown in Figure 5.
From Figure 4 and Figure 5, it is possible to observe that, although performance decreases when a higher payload is hidden, all the selected fragile reversible watermarking schemes hold high transparency for high payloads. The PSNR and Watson values for the highest payload applied are better than the minimum suggested in the literature.

3.2. Self-Recovery Schemes

Each self-recovery scheme was tested using the images from the data set to evaluate the perceptual distortion and robustness of each scheme; the perceptual distortion is measured using the PSNR and Watson metrics. To determine the robustness of the schemes, content replacement attacks were applied to the watermarked images using different percentages to corroborate their robustness to a maximum percentage of attack, which is the same maximum as the ones reported by the authors. A content replacement attack takes a region of the image to attack and replaces it with another image with the same dimensions as the region being replaced. Although other common attacks exist, we focus on content replacement in this work since it is the attack that the two self-recovery schemes resist. Figure 6 shows an example of content replacement attacks using different percentages of substitution. Table 1 shows the mean value for the PSNR, and Watson’s values were obtained from measuring the perceptual quality of the images watermarked with the self-recovery schemes; the maximum percentage of attack supported by each scheme is included in the last column. The maximum percentage of attack refers to the highest level of substitution where each scheme is capable of producing an image with perfect restoration, i.e., the scheme produces, as a result, an image containing the exact pixel values as the host image. Furthermore, the time complexity for processing one image is shown in Table 1.
From Table 1, it is possible to observe that self-recovery strategies generate high distortion levels on the watermarked images; this is caused because a big amount of information from the image itself is embedded into the carrier image to guarantee self-recovery capacities. In terms of robustness, it can be observed that the scheme by Zhang and Wang [26] achieves robustness for a maximum attack of 3.2%, while the scheme by Bravo-Solorio et al. [27] achieves robustness for a maximum attack of 20%. It is worth mentioning that the scheme [27] can achieve robustness to higher percentages of attack at higher computational time cost; in this work, an optimized implementation of the scheme is used, which needs about 50 min to reconstruct an image tampered at 20%. Then, because of our limited computational resources, it was prohibitive to test [27] for percentages of tampering higher than 20%.
The workflow for the restoration of the self-recovery scheme is as follows: during encoding, the self-recovery stage receives an image watermarked with the fragile reversible scheme and produces a second watermarked image that contains information for self-recovery, then that second watermarked image is attacked with a percentage of content replacement. In the decoding phase, the self-recovery stage receives the attacked image and extracts self-recovery information that allows perfect restoration, given that the attacked region is smaller than the maximum percentage supported; after perfect restoration, the image is passed to the fragile reversible stage to extract the secret message and recover the host image. Figure 7 depicts an example of perfect restoration after a 10% attack.

3.3. Framework

As described in Section 2, the framework consists of two main stages, a fragile reversible watermarking stage and a self-recovery stage. This section evaluates the framework through each possible combination of fragile reversible watermarking and self-recovery schemes. Table 2 shows the perceptual distortion performance for each scheme in detail.
The framework was evaluated using six different configurations by combining the three selected fragile reversible watermarking and the two self-recovery schemes. Table 3 shows the results in terms of visual transparency for each configuration; the mean, standard deviation, minimum, and maximum values for PNSR and Watson metrics are presented. The robustness of the framework depends only on the self-recovery stage; therefore, no experimentation was carried out to observe the robustness of each configuration.
As it is possible to observe from Table 2, fragile reversible watermarking schemes hold low perceptual distortion; thus, self-recovery schemes produce high perceptual distortion levels. This behavior is due to self-recovery schemes that need to embed considerable image information to have reconstruction capacities. Table 3 shows that the perceptual distortion due to the framework is very close to the distortion due to the self-recovery scheme used in the configuration; in this table, the best configuration in terms of average PSNR and Watson values is highlighted in bold, which is the combination of the fragile reversible scheme [24] and the self-recovery [27]. However, additional experimentation was carried out to corroborate the framework’s performance in terms of perceptual distortion. A fragile reversible watermarking scheme with high perceptual distortion was incorporated into the framework, and the perceptual distortion of the framework was measured. The fragile reversible watermarking scheme used is proposed by Coltuc and Tudoroiu [35]. Table 4 shows the performance of [35] in perceptual distortion terms when applied to the image data set. In Table 4, n is a payload controller; thus, log 2 n is the payload achieved per pixel. In [35], n [ 2 , 16 ] ; in this study, the minimum and maximum values are considered.
As it can be observed in Table 4, the scheme proposed in [35] holds high perceptual distortion; in fact, distortion due to this fragile reversible watermarking scheme is more elevated than self-recovery schemes.
Table 5 shows the performance of the framework for each possible configuration using the scheme in [35].
From Table 5, it is possible to observe that the framework performance is determined by the fragile reversible watermarking scheme instead of the self-recovery scheme, as in Table 3. Then, it is possible to claim that the expected perceptual distortion of the framework depends on the performance at each stage, regardless of its nature. Thus, the framework distortion could be modeled as follows:
D f r a m e max { D f w , D s r }
where D f r a m e is the distortion due to the framework, D f w is due to the fragile reversible watermarking scheme, and D s r is due to the self-recovery scheme. Alternatively, the performance of the framework can be expressed in terms of the PSNR as follows:
P S N R f r a m e min { P S N R f w , P S N R s r }
where P S N R f r a m e is the PSNR value for the framework, P S N R f w is the PSNR value for the fragile reversible watermarking scheme, and P S N R s r is the PSNR value for the self-recovery scheme.

4. Conclusions

In this study, the framework proposed in [20] was evaluated using ten different configurations to determine its perceptual distortion and robustness performance behavior. Using the appropriate reversible and self-recovery schemes within the framework makes it possible to achieve high robustness to content replacement attacks. At the same time, perceptual distortion is kept at practical levels. Moreover, an expression that approximated the framework performance in terms of perceptual distortion was proposed. The findings of this study will help the fast evaluation of reversible watermarking and self-recovery schemes to provide reversible watermarking robust to content replacement attack as the approximate perceptual distortion can be estimated by (20) and robustness is only due to the self-recovery stage.

Author Contributions

Conceptualization, C.F.-U. and A.M.-O.; methodology, J.J.G.-H.; software, D.W.R.-C. and J.J.G.-H.; validation, J.J.G.-H., C.F.-U. and A.M.-O.; formal analysis, J.J.G.-H. and A.M.-O.; investigation, J.J.G.-H. and D.W.R.-C.; resources, C.F.-U.; data curation, J.J.G.-H.; writing—original draft preparation, A.M.-O.; writing—review and editing, J.J.G.-H. and C.F.-U.; visualization, D.W.R.-C.; supervision, C.F.-U.; project administration, J.J.G.-H. and C.F.-U.; funding acquisition, J.J.G.-H. and C.F.-U. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the project 2017-01-7092 from CONACyT, Mexico.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

In this study, 2000 gray-level images from the BOWS 2 [34] data set were used. The BOWS 2 data set is freely available in http://bows2.ec-lille.fr (last access: 13 June 2022).

Conflicts of Interest

The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Menendez-Ortiz, A.; Feregrino-Uribe, C.; Hasimoto-Beltran, R.; Garcia-Hernandez, J.J. A Survey on Reversible Watermarking for Multimedia Content: A Robustness Overview. IEEE Access 2019, 7, 132662–132681. [Google Scholar] [CrossRef]
  2. An, L.; Gao, X.; Yuan, Y.; Tao, D.; Deng, C.; Ji, F. Content-adaptive reliable robust lossless data embedding. Neurocomputing 2012, 79, 1–11. [Google Scholar] [CrossRef]
  3. An, L.; Gao, X.; Yuan, Y.; Tao, D. Robust lossless data hiding using clustering and statistical quantity histogram. Neurocomputing 2012, 77, 1–11. [Google Scholar] [CrossRef]
  4. Yang, C.Y.; Lin, C.H.; Hu, W.C. Reversible Watermarking by Coefficient Adjustment Method. In Proceedings of the 2010 Sixth International Conference on Intelligent Information Hiding and Multimedia Signal Processing (IIH-MSP), Darmstadt, Germany, 15–17 October 2010; pp. 39–42. [Google Scholar] [CrossRef]
  5. Tsai, H.H.; Tseng, H.C.; Lai, Y.S. Robust lossless image watermarking based on α-trimmed mean algorithm and support vector machine. J. Syst. Softw. 2010, 83, 1015–1028. [Google Scholar] [CrossRef]
  6. An, L.; Gao, X.; Li, X.; Tao, D.; Deng, C.; Li, J. Robust reversible watermarking via clustering and enhanced pixel-wise masking. IEEE Trans. Image Process. 2012, 21, 3598–3611. [Google Scholar] [CrossRef] [PubMed]
  7. Dai, Z.; Lian, C.; He, Z.; Jiang, H.; Wang, Y. A Novel Hybrid Reversible-Zero Watermarking Scheme to Protect Medical Image. IEEE Access 2022, 10, 58005–58016. [Google Scholar] [CrossRef]
  8. Soualmi, A.; Alti, A.; Laouamer, L.; Benyoucef, M. A Blind Fragile Based Medical Image Authentication Using Schur Decomposition. In The International Conference on Advanced Machine Learning Technologies and Applications (AMLTA2019); Hassanien, A.E., Azar, A.T., Gaber, T., Bhatnagar, R., Tolba, M.F., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 623–632. [Google Scholar]
  9. Fridrich, J.; Goljan, M. Protection of digital images using self embedding. In Symposium on Content Security and Data Hiding in Digital Media; New Jersey Institute of Technology: Newark, NJ, USA, 1999. [Google Scholar]
  10. He, H.; Zhang, J.; Chen, F. A self-recovery fragile watermarking scheme for image authentication with superior localization. Sci. China Ser. F Inf. Sci. 2008, 51, 1487–1507. [Google Scholar] [CrossRef]
  11. He, H.J.; Zhang, J.S.; Tai, H.M. Self-recovery Fragile Watermarking Using Block-Neighborhood Tampering Characterization. In Information Hiding; Katzenbeisser, S., Sadeghi, A.R., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2009; Volume 5806, pp. 132–145. [Google Scholar] [CrossRef]
  12. Bravo-Solorio, S.; Li, C.T.; Nandi, A. Watermarking with low embedding distortion and self-propagating restoration capabilities. In Proceedings of the 19th IEEE International Conference on Image Processing (ICIP), Orlando, FL, USA, 30 September–3 October 2012; pp. 2197–2200. [Google Scholar] [CrossRef]
  13. Aminuddin, A.; Ernawan, F. AuSR1: Authentication and self-recovery using a new image inpainting technique with LSB shifting in fragile image watermarking. J. King Saud Univ.-Comput. Inf. Sci. 2022; in press. [Google Scholar] [CrossRef]
  14. Wu, H.C.; Chang, C.C. Detection and restoration of tampered JPEG compressed images. J. Syst. Softw. 2002, 64, 151–161. [Google Scholar] [CrossRef]
  15. Zhang, X.; Qian, Z.; Ren, Y.; Feng, G. Watermarking With Flexible Self-Recovery Quality Based on Compressive Sensing and Compositive Reconstruction. IEEE Trans. Inf. Forensics Secur. 2011, 6, 1223–1232. [Google Scholar] [CrossRef]
  16. Li, C.; Wang, Y.; Ma, B.; Zhang, Z. A novel self-recovery fragile watermarking scheme based on dual-redundant-ring structure. Comput. Electr. Eng. 2011, 37, 927–940. [Google Scholar] [CrossRef]
  17. Hassan, A.M.; Al-Hamadi, A.; Hasan, Y.M.Y.; Wahab, M.A.A.; Michaelis, B. Secure Block-Based Video Authentication with Localization and Self-Recovery. World Acad. Sci. Eng. Technol. 2009, 2009, 69–74. [Google Scholar]
  18. Shi, Y.; Qi, M.; Lu, Y.; Kong, J.; Li, D. Object based self-embedding watermarking for video authentication. In Proceedings of the International Conference on Transportation, Mechanical, and Electrical Engineering (TMEE), Changchun, China, 16–18 December 2011; pp. 519–522. [Google Scholar] [CrossRef]
  19. Mobasseri, B. A spatial digital video watermark that survives MPEG. In Proceedings of the International Conference on Information Technology: Coding and Computing, Las Vegas, NV, USA, 27–29 March 2000; pp. 68–73. [Google Scholar] [CrossRef]
  20. Menendez-Ortiz, A.; Feregrino-Uribe, C.; Garcia-Hernandez, J.J. Reversible image watermarking scheme with perfect watermark and host restoration after a content replacement attack. In Proceedings of the The 2014 International Conference on Security and Management (SAM’14), Las Vegas, NV, USA, 7–9 April 2014; Volume 13, pp. 385–391. [Google Scholar]
  21. Coltuc, D. Towards distortion-free robust image authentication. J. Physics: Conf. Ser. 2007, 77, 012005. [Google Scholar] [CrossRef]
  22. Hu, R.; Xiang, S. Cover-Lossless Robust Image Watermarking Against Geometric Deformations. IEEE Trans. Image Process. 2021, 30, 318–331. [Google Scholar] [CrossRef] [PubMed]
  23. Huang, L.C.; Tseng, L.Y.; Hwang, M.S. A Reversible Data Hiding Method by Histogram Shifting in High Quality Medical Images. J. Syst. Softw. 2013, 86, 716–727. [Google Scholar] [CrossRef]
  24. Wang, Z.H.; Lee, C.F.; Chang, C.Y. Histogram-Shifting-Imitated Reversible Data Hiding. J. Syst. Softw. 2013, 86, 315–323. [Google Scholar] [CrossRef]
  25. Chakraborty, S.; Maji, P.; Pal, A.; Biswas, D.; Dey, N. Reversible Color Image Watermarking Using Trigonometric Functions. In Proceedings of the International Conference on Electronic Systems, Signal Processing and Computing Technologies (ICESC), Nagpur, India, 9–11 January 2014; pp. 105–110. [Google Scholar]
  26. Zhang, X.; Wang, S. Fragile Watermarking With Error-Free Restoration Capability. IEEE Trans. Multimed. 2008, 10, 1490–1499. [Google Scholar] [CrossRef]
  27. Bravo-Solorio, S.; Li, C.T.; Nandi, A. Watermarking method with exact self-propagating restoration capabilities. In Proceedings of the IEEE International Workshop on Information Forensics and Security (WIFS), Costa Adeje, Spain, 2–5 December 2012; pp. 217–222. [Google Scholar] [CrossRef]
  28. Hore, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 20th International Conference on Pattern Recognition (ICPR), Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  29. Ismail Avcıbas, B.S. Statistical Analysis of Image Quality Measures; Technical Report; Department of Electrical and Electronic Engineering, Bogaziçi University: İstanbul, Turkey, 1999. [Google Scholar]
  30. Garcia-Hernandez, J.J.; Gomez-Flores, W.; Rubio-Loyola, J. Analysis of the impact of digital watermarking on computer-aided diagnosis in medical imaging. Comput. Biol. Med. 2016, 68, 37–48. [Google Scholar] [CrossRef] [PubMed]
  31. Watson, A.B. DCT quantization matrices visually optimized for individual images. In Human Vision, Visual Processing and Digital Display IV; SPIE: Bellingham, WA, USA, 1993; Volume 1913, pp. 202–216. [Google Scholar]
  32. Ahumada, A.J.; Peterson, H.A. Luminance-Model-Based DCT Quantization for Color Image Compression. In Human Vision, Visual Processing and Digital Display III; SPIE: Bellingham, WA, USA, 1992; Volume 1666, pp. 365–374. [Google Scholar]
  33. Rodriguez, T.F.; Cushman, D.A. Optimized Selection of Benchmark Test Parameters for Image Watermark Algorithms based on Taguchi Methods and Corresponding Influence on Design Decisions for Real-World. In SPIE-IS&T Electronic Imaging; SPIE: Bellingham, WA, USA, 2003; Volume 5020, pp. 215–228. [Google Scholar]
  34. Laboratory, W.V. Break Our Watermarking Systems 2, 2007. Image Data Set. Available online: http://bows2.ec-lille.fr (accessed on 13 June 2022).
  35. Coltuc, D.; Tudoroiu, A. Multibit versus Multilevel Embedding in High Capacity Difference Expansion Reversible Watermarking. In Proceedings of the 20th European Signal Processing Conference (EUSIPCO 2012), Bucharest, Romania, 27–31 August 2012; pp. 1791–1795. [Google Scholar]
Figure 1. Framework for robust RWS with signal restoration.
Figure 1. Framework for robust RWS with signal restoration.
Applsci 12 07242 g001
Figure 2. General fragile reversible watermarking scheme.
Figure 2. General fragile reversible watermarking scheme.
Applsci 12 07242 g002
Figure 3. General self-recovery scheme.
Figure 3. General self-recovery scheme.
Applsci 12 07242 g003
Figure 4. Fragile reversible watermarking schemes performance versus payload under PSNR metric.
Figure 4. Fragile reversible watermarking schemes performance versus payload under PSNR metric.
Applsci 12 07242 g004
Figure 5. Fragile reversible watermarking schemes performance versus payload under the Watson metric.
Figure 5. Fragile reversible watermarking schemes performance versus payload under the Watson metric.
Applsci 12 07242 g005
Figure 6. Content replacement attack: (a) 3.2%, (b) 10%, (c) 15%, and (d) 20%.
Figure 6. Content replacement attack: (a) 3.2%, (b) 10%, (c) 15%, and (d) 20%.
Applsci 12 07242 g006
Figure 7. Example of image perfect restoration after a 10% attack. (a) Watermarked image, (b) attacked image, and (c) restored image.
Figure 7. Example of image perfect restoration after a 10% attack. (a) Watermarked image, (b) attacked image, and (c) restored image.
Applsci 12 07242 g007
Table 1. Self-recovery schemes’ performance in perceptual, robustness, and time complexity terms.
Table 1. Self-recovery schemes’ performance in perceptual, robustness, and time complexity terms.
AlgorithmPSNR dBWatsonMax AttackTime Complexity
Zhang and Wang [26]29.570.1353.2%10 msec
Bravo-Solorio et al. [27]37.900.06720.0%49.7 min
Table 2. Perceptual distortion performance for fragile reversible watermarking (F1, F2, and F3) and self-recovery schemes (S1 and S2).
Table 2. Perceptual distortion performance for fragile reversible watermarking (F1, F2, and F3) and self-recovery schemes (S1 and S2).
IDWorkPSNR (dB)Watson
μ σ minmax μ σ minmax
F1[23]47.03±2.8439.6856.400.1433±0.00320.00070.0905
F2[24]65.35±0.0365.2165.480.0016±0.00150.00060.0271
F3[25]68.84±0.0468.7269.000.0003±0.00020.00020.0027
S1[26]29.57±4.1020.1746.980.1350±0.06750.01690.4146
S2[27]37.90±0.1236.7838.710.0679±0.03160.04220.7103
Table 3. Perceptual distortion performance for framework configuration. The best configuration in terms of average PSNR and Watson values is highlighted in bold.
Table 3. Perceptual distortion performance for framework configuration. The best configuration in terms of average PSNR and Watson values is highlighted in bold.
ConfigurationsPSNR (dB)WatsonPayload (bits)
μ σ minmax μ σ minmax
F1-S129.02±4.5120.2540.340.1433±0.08300.02710.39535k
F1-S237.34±0.3336.3437.900.0688±0.04970.04970.45465k
F2-S129.56±4.0920.1746.760.1355±0.06750.01740.414510k
F2-S237.90±0.1336.5338.740.0678±0.03130.04190.702210k
F3-S132.39±3.8723.1046.130.0985±0.04590.02020.297710k
F3-S237.89±0.1236.9038.900.0678±0.01990.04200.335710k
Table 4. Perceptual distortion performance of [35].
Table 4. Perceptual distortion performance of [35].
IDExpansion nPSNR (dB)Watson
μ σ minmax μ σ minmax
C1226.51±6.767.9349.920.1673±0.16570.01182.8645
C21215.24±3.376.2527.830.7331±0.31170.14024.3541
Table 5. Perceptual distortion performance for framework configuration using [35] as the fragile reversible watermarking scheme.
Table 5. Perceptual distortion performance for framework configuration using [35] as the fragile reversible watermarking scheme.
ConfigurationsPSNR (dB)Watson
μ σ minmax μ σ minmax
C1-S121.74±5.117.8940.700.3164±0.23090.03513.3922
C1-S225.78±5.887.9237.650.1983±0.16860.05392.9249
C2-S112.40±2.596.1522.661.0504±0.37840.27534.9221
C2-S215.21±3.346.5527.440.7391±0.31160.15184.3353
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Garcia-Hernandez, J.J.; Feregrino-Uribe, C.; Menendez-Ortiz, A.; Robledo-Cruz, D.W. Evaluation of a Framework for Robust Image Reversible Watermarking. Appl. Sci. 2022, 12, 7242. https://doi.org/10.3390/app12147242

AMA Style

Garcia-Hernandez JJ, Feregrino-Uribe C, Menendez-Ortiz A, Robledo-Cruz DW. Evaluation of a Framework for Robust Image Reversible Watermarking. Applied Sciences. 2022; 12(14):7242. https://doi.org/10.3390/app12147242

Chicago/Turabian Style

Garcia-Hernandez, Jose Juan, Claudia Feregrino-Uribe, Alejandra Menendez-Ortiz, and Dan Williams Robledo-Cruz. 2022. "Evaluation of a Framework for Robust Image Reversible Watermarking" Applied Sciences 12, no. 14: 7242. https://doi.org/10.3390/app12147242

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop