Reversible Data Hiding for Color Images Using Channel Reference Mapping and Adaptive Pixel Prediction

: Reversible data hiding (RDH) is a technique that embeds secret data into digital media while preserving the integrity of the original media and the secret data. RDH has a wide range of application scenarios in industrial image processing, such as intellectual property protection and data integrity verification. However, with the increasing prevalence of color images in industrial applications, traditional RDH methods for grayscale images are inadequate to meet the requirements of image fidelity. This paper proposes an RDH method for color images based on channel reference mapping (CRM) and adaptive pixel prediction. Initially, the CRM mode for a color image is established based on the pixel variation correlation between the RGB channels. Then, the pixel local complexity context is adaptively selected using the CRM mode. Next, each pixel value is adaptively predicted based on the features and characteristics of adjacent pixels and reference channels, and then data is embedded by expanding the prediction error. Finally, we compare seven existing RDH algorithms on the standard image dataset and the Kodak dataset to validate the advantages of our method. The experimental results demonstrate that our approach achieves average peak signal-to-noise ratio (PSNR) values of 63.61 and 60.53 dB when embedding 20,000 and 40,000 bits of data, respectively. These PSNR values surpass those of other RDH methods. These findings indicate that our method can effectively preserve the visual quality of images even under high embedding capacities.


Introduction
Color images constitute a vital component of digital information transmission and storage [1][2][3].With the rapid advancement of digital technology and the advent of the Industry 4.0 era [4][5][6], the scope of applications for color images continues to expand.They are widely popular in the entertainment sector and play a pivotal role in the industrial domain.Industrial applications typically require high-quality transmission and storage of color images to support decision-making [7], quality control [8], and monitoring production processes [9].However, safeguarding sensitive information contained within images presents challenges due to the complexity of industrial environments and the stringent requirements for accurate information transmission.Reversible data hiding (RDH) technology has garnered significant attention within information security [10][11][12][13].This technology's fundamental attribute lies in its capacity to embed confidential data into images without compromising the original image's integrity, a crucial feature in industrial applications necessitating the restoration of the original image.In particular, we need to deal with a large amount of color image data in the industrial field, such as images captured by surveillance cameras, product inspection images, and quality control images.These images often contain critical industrial process information and production details, making the utilization of RDH to safeguard sensitive information within industrial images notably valuable.Consequently, this paper proposes a novel RDH method explicitly designed for color images, which enables the concealment of industrial secrets within color images while ensuring both the reversibility of data embedding and extraction and the enhancement of color image fidelity.Furthermore, it safeguards the integrity and security of industrial information.
RDH techniques for grayscale images have been extensively researched, primarily categorizing four categories: methods based on lossless compression [14], histogram shifting (HS) [15], difference expansion (DE) [16], and prediction error expansion (PEE) [17].Although extensive research and development has been done on RDH for grayscale images, color images are more prevalent in daily life and industrial production.Some researchers have directly applied RDH methods for grayscale images to color images.Nevertheless, this approach neglects the correlation between the three color channels in color images, resulting in a reduced capacity for data embedding.Therefore, in recent years, RDH methods tailored to color images that leverage inter-channel correlations have been proposed.For instance, Yang et al. [18] designed a high-capacity data embedding and extraction method based on the characteristics of color filter array images.In the paper [19], Li et al. utilized an improved predictor to predict each channel of a color image and employed interchannel correlation for adaptive data embedding, thereby enhancing embedding efficiency.Yao et al. [20] employed guided filtering for prediction in color images and utilized PEE for data embedding, further improving pixel prediction accuracy.The approach proposed by Ou et al. [21] customized effective payload distribution based on the characteristics of each color channel to optimize the process.It also adopted an adaptive embedding strategy to enhance data hiding capacity while minimizing image distortion.Hou et al. [22] introduced a method that maintains grayscale invariance while achieving RDH in color images, ensuring that hidden data can be extracted without affecting the perceived brightness of the image.In the latest research results, Chang et al. [23] dynamically adjusted the embedding mappings of three prediction error histograms based on the inter-channel correlations, thus optimizing embedding performance and image quality.Kong et al. [24] presented an RDH model based on multi-channel difference value ordering.Mao et al. [25] combined the concept of channel unity embedding with pixel value ordering (PVO).Bhatnagar et al. [26] proposed an RDH method for color images based on skewed histograms and cross-channel correlation.Kumar et al. [27] proposed a color image steganography scheme using gray invariant in AMBTC compression domain.In [28], Kumar et al. discussed and reviewed the existing pixel predictors.
While some progress has been made, there remains a need for a deeper exploration of inter-channel correlations in RDH methods for color images to enhance their embedding capacity.The current research exhibits three primary limitations: 1.
Most methods are extensions of single-channel RDH approaches and need to adequately consider the correlations among color image channels, resulting in limited improvements in embedding capacity.

2.
Many approaches rely solely on techniques such as prediction error and PVO for data embedding, failing to leverage the untapped potential of other data hiding spaces within color images, such as color space transformation and color quantization.

3.
The majority of methods employ fixed pixel prediction strategies and parameter settings without dynamic adjustments based on specific image pixel conditions, leading to an imbalance between embedding capacity and pixel distortion.
The preceding analysis indicates that RDH techniques for color images hold significant potential for applications in industrial production and information security.However, existing RDH algorithms for color images still need to be revised, including restricted data embedding capacity, increased image distortion, and underutilized inter-channel correlations.Therefore, our work aims to improve the RDH algorithm's performance and achieve a balance between the embedding capacity and the visual quality of the images.This goal involves improving the local complexity calculation method and using channel reference mapping and an adaptive pixel prediction strategy.This paper proposes a novel RDH method for color images that addresses these challenges based on channel reference mapping and adaptive pixel prediction.The primary innovations and contributions of our paper are delineated as follows: 1.
A novel channel reference mapping (CRM) method is proposed, leveraging trends and correlations among the pixels in the three channels to establish inter-channel reference relationships.These reference relationships are incorporated into pixel-wise local complexity computation and pixel value prediction, thus effectively exploiting the inherent inter-channel connections and reducing pixel distortion during data embedding.

2.
An adaptive local complexity computation algorithm is proposed.Based on the CRM mode, the current channel's pixel-wise local complexity computation context is adaptively selected according to the values of reference channel pixels.Adaptive context selection leads to a more accurate assessment of local complexity.

3.
An adaptive pixel prediction strategy is proposed.By considering each pixel's neighborhood features and channel characteristics, appropriate predictors and prediction contexts are chosen, thereby enhancing the accuracy of pixel prediction while mitigating image distortion.
The rest of this paper is structured as follows.Section 2 briefly overviews the related works.Section 3 presents the proposed CRM-based RDH method in detail.Section 4 reports the experimental comparison results and analysis that compare our method with other existing algorithms.Section 5 summarizes the paper.

Related Work
This section briefly overviews the local complexity method proposed for color images in [21], as detailed in Section 2.1.Additionally, we introduce the pixel prediction method proposed in [19], and specific details can be found in Section 2.2.

Local Complexity Calculation Method [21]
The selection of pixels for data embedding is a pivotal factor influencing RDH performance, impacting indicators such as data embedding capacity, embedding distortion, and security.While embedding strategies differ between grayscale and color images, opting for low-complexity pixels is generally advantageous as data carriers.This choice has the potential to enhance the overall performance of RDH algorithms.Ou et al. [21] employed inter-channel correlation for computing pixel complexity.The pixel smoothness can be described more accurately by considering the texture similarity among the channels.To reduce pixel distortion, pixels with lower complexity are prioritized for data embedding.P x and its surrounding eight pixels are illustrated in Figure 1.The complexity calculation for each channel's pixel P x is as: where ∆ x represents the complexity of P x .The local complexity calculation methods for the R, G, and B channels are the same.To further enhance the accuracy of pixel complexity calculation, the concepts of the current channel and reference channels are introduced.The current channel is the channel in which the pixel for calculating local complexity is located, and the other two channels are reference channels.With the introduction of reference channels, the local complexity of the current channel pixel P x is denoted as LC(x) and is given by: where ∆ c x represents the complexity of the current channel, and ∆ r 1 x and ∆ r 2 x represent the complexities of reference channels.∆ c x , ∆ r 1 x , and ∆ r 2 x are calculated by Equation (1).Ou et al.'s [21] local complexity calculation method uses a fixed channel ratio, which is not conducive to adapting to different color image features.Therefore, how to adjust the local complexity calculation method adaptively according to the channel correlation of the image is a worthwhile research problem.

Pixel Prediction Method [19]
Li et al. [19] proposed a pixel prediction method based on the similarity of edge information across channels.They applied different prediction strategies for edge textures with varying levels of complexity.The edge texture, depicted in Figure 1, is accomplished through the pixel's eight adjacent pixels as: where P r x represents the pixel of the reference channel, and P r u represents the eight adjacent pixels of P r x .To account for the variation of texture complexity in different directions, Li et  x is smooth, and the four pixels adjacent to P c x are used for pixel prediction.The prediction formula for pixel P c x is: where P c 1 , P c 2 , P c 3 , and P c 4 are the four pixels adjacent to x is rough, and the texture in a particular direction may be relatively smooth.In such instances, we use the pixels in the direction with the minimum texture complexity to calculate the pixel prediction value.For example, if W d = W h , the predicted pixel P c x is expressed as P c x = (P c 2 + P c 4 )/2.This method determines the pixel prediction value and expands the prediction errors to embed the data.Li et al.'s [19] pixel prediction strategy utilizes the similarity of edge information between channels, but ignores the difference of texture information within channels.Therefore, how to adjust the pixel prediction strategy dynamically according to the channel texture of the image is a worthwhile research problem.

Proposed Method
In this section, we provide an extensive exposition of our proposed method, encompassing an overview of the CRM-based RDH framework, the formulation of channel reference mapping, the computation of local complexity, adaptive pixel prediction, and the implementation procedure.

An Overview of the CRM-Based RDH Framework
The proposed CRM-based RDH framework is composed of two main components: an embedding process and an extraction process.Figure 2   On the sender side, we propose a data hiding method based on channel reference mapping (CRM) and adaptive pixel prediction.First, we use the correlation between the three channels of the cover color image I to construct a CRM mode.Next, we calculate the local complexity of each pixel adaptively according to the CRM mode and select the pixels with low complexity for data embedding.After that, we perform adaptive pixel prediction on them and embed data into them.Then, we embed auxiliary information to ensure the reversibility of data hiding.Finally, we combine the pixels of the three channels to obtain the marked color image I ′ .
On the receiver side, we reverse the process on the sender side.First, we recover the auxiliary information from the received color image I ′ with hidden data and use it to reconstruct the CRM mode.Then, we calculate the local complexity of each pixel adaptively according to the CRM mode and recover the pixels with low complexity.After that, we use the pixels with low complexity to perform pixel prediction and extract the embedded data.Finally, we recombine the pixels of the three channels to restore the cover color image I.

Channel Reference Mapping Establishment
The characteristics and texture structures among RGB channels exhibit remarkable similarities, manifested as a fundamental consistency in the pixel value variations across the three channels.This similarity has been extensively validated through a plethora of experiments.We take the images "Lena" and "Peppers" as examples and select 500 pixels from each image.As depicted in Figure 3, the pixel value trends of these pixels show similar monotonicity in the RGB channels.To analyze the correlation among the RGB channels, we employed the Pearson correlation coefficient as an evaluation metric.The Pearson correlation coefficient is a commonly used statistical method that quantifies the strength and direction of the linear relationship between two variables.Its formula is presented as Equation ( 6), where cor(X, Y) represents the correlation coefficient between the X and Y channels.The size of the color image is M rows by N columns, with M × N denoting the total number of pixels in the color image.X t and Y t correspond to the pixel values of the X and Y channels at the t-th position, while X and Y denote the means of all pixels in the X and Y channels, respectively.The value of cor(X, Y) ranges from −1 to 1. Positive values indicate a positive correlation, negative values indicate a negative correlation, and a value of zero indicates no correlation.The larger the absolute value of cor(X, Y), the stronger the correlation between X and Y.The Pearson correlation coefficients between the RGB channels are denoted as cor(R, G), cor(R, B), and cor(G, B), respectively.We represent these three coefficients a set We denote the maximum and second maximum values in the set K as K max and K smax , respectively.Subsequently, we use the functions chan(K max ) and chan(K smax ) to calculate the two channels involved when the correlation coefficients take the values K max and K smax , respectively.We express chan(K max ) and chan(K smax ) as where ch 1 and ch 2 represent the two channels involved when the correlation coefficient takes the value K max , while ch 3 and ch 4 correspond to the two channels associated with K smax .As detailed in Table 1, we can construct six mapping modes between the three channels of a color image, denoted by M 1 to M 6 .Each mapping mode consists of three mapping functions describing the relationship between the current and reference channels.Specifically, the function f 1 (X) = Y represents the first mapping function, indicating the correspondence between the current channel X and the reference channel Y.The function f 2 (Y) = Z represents the second mapping function, illustrating the association between the current channel Y and the reference channel Z.The function f 3 (Z) = X ′ constitutes the third mapping function employed to describe the relationship between the current channel Z and the reference channel X ′ .Since data embedding performs in the sequence of channels X, Y, and Z, channel X has already completed data embedding and is denoted as X ′ when data embedding is performed on channel Z.

Adaptive Local Complexity Calculation
After obtaining the CRM mode, the data embedding is performed in a per-channel, per-pixel order while scanning the color image from left to right and top to bottom.We employ a method based on local complexity to select suitable embedding pixels, which ensures both image quality and the effectiveness of the concealment.Local complexity is an indicator that measures the extent of change for each pixel, effectively reflecting image details and texture characteristics.A higher local complexity value indicates more information at a particular location, potentially leading to more significant distortions.Therefore, we select pixels with lower local complexity for data embedding to reduce distortion.There are various methods for local complexity calculation, among which the simplest method is to use the difference between the maximum and minimum values in the local complexity context of that pixel.However, this method only considers the four neighboring pixels of the current pixel, disregarding pixel variations over a more extensive range, potentially failing to capture local image features accurately.Hence, we propose an adaptive local complexity computation method leveraging the CRM mode to resolve this limitation, which considers the influence of the current channel and the reference channel over a more extensive range.This method dynamically adjusts parameters according to different image characteristics, leading to a more accurate reflection of pixel complexity.
We denote the current channel being processed as C and the reference channel for the current channel as S. Specifically, the current pixel being processed is C i,j , where i and j represent the row and column numbers, respectively.Similarly, S i,j denotes the pixel in the reference channel at row i and column j.To more accurately denote the local complexity of C i,j , we extend its local complexity context to a 3 × 3 pixel block centered at C i,j .Likewise, the local complexity context of S i,j in the reference channel is denoted as a 3 × 3 pixel block centered at S i,j .The local complexity contexts of C i,j , and S i,j are shown in Figure 4a,b, respectively.The local complexity contexts for C i,j and S i,j are denoted as NC(C i,j ) and NS(S i,j ), and each of them contains 8 pixels, i.e., NC(C i,j and S i,j = {S i−1,j , S i,j+1 , S i+1,j , S i,j−1 , S i−1,j+1 , S i+1,j+1 , S i+1,j−1 , S i−1,j−1 }.Within NC(C i,j ), the maximum and minimum values are represented as NC max and NC min , while within NS(S i,j ), they are represented as NS max and NS min .We integrate the pixel variation in the local complexity contexts of the current channel and the reference channel and define the local complexity LC(C i,j ) of the current pixel C i,j as where λ 1 and λ 2 represent the influence weight of the current channel and the reference channel, respectively, We introduce a parameter named local complexity threshold, denoted by T, to determine whether a pixel is suitable for data embedding.When the local complexity satisfies LC(C i,j ) ≤ T, it indicates that the pixel C i,j has smooth characteristics and is suitable for data embedding.On the contrary, C i,j belongs to the rough category and is unsuitable for data embedding.

Adaptive Pixel Prediction
We employ the PEE [17] and pixel-based PVO [29] for data embedding.PEE entails an initial prediction of smooth pixels, from which prediction errors are obtained, followed by the subsequent concealment of secret data within these prediction errors.To enhance prediction accuracy and augment the embedding capacity of prediction errors, we propose an adaptive pixel prediction approach.This method automatically selects appropriate prediction contexts based on varying pixel features.
The diamond predictor is a widely used pixel prediction technique to obtain the prediction context of the current pixel C i,j , which includes the four adjacent pixels located at the current pixel's top, bottom, left, and right.However, the prediction range of the diamond predictor is limited and cannot capture a wider range of pixel relationships, especially in complex image scenes.Some pixels may be affected by more distant or complex pixels, but these pixels are not considered, resulting in inaccurate predictions.We propose an adaptive prediction context generation method based on a diamond predictor to address this issue.The diamond prediction context of the current pixel C i,j is denoted as Rh(C i,j ), and the diamond prediction context of the reference pixel S i,j is denoted as Rh(S i,j ).Initially, Rh(C i,j ) includes the four adjacent pixels of the current pixel, i.e., Rh(C i,j ) = {C i−1,j , C i,j+1 , C i+1,j , C i,j−1 }, and Rh(S i,j ) includes the adjacent pixels of the reference pixel, i.e., Rh(S i,j ) = {S i−1,j , S i,j+1 , S i+1,j , S i,j−1 }.To further expand the scope of the prediction context and consider the consistency of pixel value change trends between the reference channel and the current channel, we utilize the diamond prediction context of the reference channel to assist in the generation of the prediction context of the current channel.We employ MaxRh(S i,j ) and MinRh(S i,j ) to represent the maximum and minimum values in Rh(S i,j ), respectively, and use Pos(MaxRh(S i,j )) and Pos(MinRh(S i,j )) to represent their positions in the channel.Specifically, Pos(MaxRh(S i,j )) = (τ 1 , τ 2 ) and Pos(MinRh(S i,j )) = (τ 3 , τ 4 ).Subsequently, we use the position information of the maximum and minimum values in the reference channel's diamond prediction context to expand the current channel's diamond prediction context.This extension is denoted as Rh(C τ 1 ,τ 2 ) and Rh(C τ 3 ,τ 4 ).Afterwards, PC(C i,j ) is utilized to represent the prediction context of the current pixel, defined as To provide a more detailed explanation of the prediction context generation strategy, we illustrate the generation of PC(C i,j ) as an example in Figure 5. Suppose that the maximum and minimum values in the reference channel Rh(S i,j ) are S i−1,j and S i+1,j , respectively, we have Pos(MaxRh(S i,j )) = (τ 1 , τ 2 ) = (i − 1, j) and Pos(MinRh(S i,j )) = (τ 3 , τ 4 ) = (i + 1, j).In this case, Rh(C τ 1 ,τ 2 ) = Rh(C i−1,j ) in the current channel, which includes the pixels C i−2,j , C i−1,j−1 , C i−1,j+1 , and C i,j , i.e., the pixels marked in yellow in Figure 5a.Similarly, Rh(C τ 3 ,τ 4 ) = Rh(C i+1,j ) in the current channel, which includes the pixels C i+1,j−1 , C i+1,j+1 , C i+2,j , and C i,j , i.e., the pixels marked in orange in Figure 5a.In summary, the prediction context PC(C i,j ) for C i,j consists of The pixel prediction context, denoted as PC(C i,j ), is a set comprising neighboring pixels surrounding the current pixel, employed for predicting the current pixel.We use PC max and PC min to denote the maximum and minimum values in the prediction context set, which represent the upper and lower bounds of C i,j .To ensure the reversibility of data embedding and extraction, we categorize C i,j into one of four sets as If C i,j does not belong to any of these four sets, then the pixel will not be used for data embedding.This approach helps prevent image distortion during data embedding.In such cases, the predicted value of C i,j can be represented by where C i,j represents the predicted value of C i,j .

Data Embedding and Data Extraction
On the sender side, we utilize the reference channel's pixel information to calculate the current pixel's local complexity and perform pixel prediction for the smooth pixels.Subsequently, we expand the prediction error to embed secret data.The prediction error of C i,j , denoted as e i,j , can be calculated as If C i,j falls within Q 1 or Q 4 , we observe that e i,j ≥ 0. If C i,j falls within Q 2 or Q 3 , then we have e i,j ≤ 0. We select the prediction errors with the highest distribution in the prediction error histograms for expansion to maximize the embedding capacity of the cover color image.Specifically, we use the bins with e i,j = 0 for data embedding, and the prediction error expansion is defined as where e i,j represents the expanded prediction error and b denotes the secret data to be embedded, with b ∈ {0, 1}.Finally, the marked pixel C i,j is calculated as The process of data extraction is the reverse of data embedding.On the receiving end, we utilize the same CRM mode as data embedding to extract secret data pixel by pixel and channel by channel.The channel order for extraction is the reverse of that used for embedding.Similarly, data extraction for each channel follows a bottom-to-top and rightto-left sequence.For each pixel, we initially calculate the local complexity of the current channel's pixels by leveraging the reference channel, following the same approach as used in embedding.Subsequently, we adaptively compute the prediction context PC( C i,j ) for smooth pixels, with PC ′ max and PC ′ min represent the maximum and minimum values within PC( C i,j ).Afterwards, we categorize C i,j into one of the following sets: Consequently, we predict the pixel value and derive the predicted pixel value C i,j as The subsequent step entails the computation of the expanded prediction error, denoted as e i,j , and it is formulated as Currently, the original prediction error, represented as e i,j , can be restored using Simultaneously, the secret data b = 0 is extracted when e i,j = 0 and b = 1 when e i,j ∈ {0, 1}.Finally, the original pixel value C i,j is restored to (19)

Implementation of the Proposed CRM-Based Method
At the sending end, the data embedding process is specified as follows.
(1) Pretreatment and LM generation.To avoid the pixel value overflow and underflow caused by data embedding, we design three location maps (LM r , LM g , and LM b ), corresponding to the red, green, and blue channels of the cover color image, respectively.Each location map shares the same dimensions as the cover color image, i.e., M × N. Initially, all elements of the location maps are set to 0. Then, we perform pixel-by-pixel preprocessing on each channel of the cover color image, changing the pixel values from 255 to 254, changing the pixel values from 0 to 1, and setting the elements of the corresponding location map to 1.
(2) Payload embedding.After obtaining the location maps (LM r , LM g , and LM b ) for three channels, we employ arithmetic coding algorithms to compress these maps.The lengths of the compressed location maps are denoted as L cr , L cg , and L cb , and their total length is represented as L clm , where L clm = L cr + L cg + L cb .The payload consists of two parts, the first part is the secret data, and the second part is a binary sequence derived from the least significant bits (LSBs) of the first (150 + L clm ) pixels from the top and bottom two rows of each channel in the cover color image.According to the established CRM mode M δ(3) , we embed the payload into the cover color image in a channel-by-channel and pixel-by-pixel manner.
(3) Auxiliary information and compressed LM embedding.After completing the embedding of the payload, we use the LSB replacement method to embed the auxiliary information into the first 150 pixels of the first row of the red channel of the cover color image.The auxiliary information consists of the following parts: L cr (20 bits), L cg (20 bits), L cb (20 bits), M δ(3) (10 bits), T (8 bits), and the positions where the payload embedding ends in the three channels, denoted by (R r , R c ), (G r , G c ), and (B r , B c ), respectively.Each of the six position information occupies 12 bits, totaling 72 bits.Following the auxiliary information, we embed the compressed LM into the cover color image.
At the receiving end, the processing steps for data extraction are detailed below.
(1) Auxiliary information and compressed LM extraction.Initially, we extract auxiliary information from the red channel of the marked color image by the LSB replacement method, which is located in the first row's first 150 pixels.Then, we utilize the extracted L c r, L cg , and L cb to extract the compressed LM.
(2) Payload extraction and pixel recovery.We employ arithmetic coding algorithms to decompress LM.Following the CRM mode M δ(3) , we extract the payload pixel by pixel in reverse order from the point where data embedding concluded.Simultaneously, we restore the original cover pixels.
(3) Image recovery.We separately restore the secret data and the LSBs of the original pixels from the extracted payload.Additionally, we perform LSB replacement for the first 150 pixels of the red channel.This process ultimately achieves the full restoration of the original color image and the secret data.

Experimental Results and Analysis
In this section, we evaluate the performance of the proposed CRM-based RDH method on various color images.First, we verify the method's effectiveness on six classic color images.Then, we perform performance tests on the method using the Kodak dataset.Finally, we analyze and summarize the performance of the method.

Color Image Datasets
We use the USC-SIPI (http://sipi.usc.edu/database/database.php?volume=misc (accessed on 20 December 2023)) and Kodak (http://r0k.us/graphics/kodak/(accessed on 20 December 2023)) datasets as experimental images to evaluate the algorithm's performance.We compare the proposed algorithm with the state-of-the-art algorithms, measuring image fidelity using peak signal-to-noise ratio (PSNR) values.Figure 6 illustrates six classic color images of size 512 × 512 from the USC-SIPI dataset: Lena, Airplane, Lake, Peppers, Splash, and House.The Kodak dataset contains 24 color images of size 512 × 768 or 768 × 512, on which we test the algorithm's performance.The secret data used in the experiments is a bit sequence composed of 0 and 1 generated by a random function.We run our method and the compared methods on a PC equipped with Intel Core i7-9700K CPU (Intel Corporation, Santa Clara, CA, USA), 16 GB memory, and Windows 10 operating system, using MATLAB R2020a environment.

Performance Comparison on Classic Color Images
To assess the performance of CDPP [21], GF-CI [20], GI-CI [22], BRG-EP [30], AT-DHM [23], OPC-PVO [31], and the proposed CRM-based RDH method in terms of image fidelity, we performed a set of comparison experiments on six classic USC-SIPI images shown in Figure 6.We utilize the PSNR in decibels (dB) to gauge the fidelity between the cover color image and the marked color image.A higher PSNR value signifies higher similarity and superior image quality.We embed 20,000 and 40,000 bits of secret data in each of the six images shown in Figure 6, respectively, and present the PSNR values after data embedding in Tables 2 and 3.The results from Tables 2 and 3 demonstrate that our proposed method achieves higher image quality on most images with the same payload.Our method exhibits superior image quality compared to the GI-CI and BRG-EP.For ex-ample, when embedding 20,000 bits, our method yields an average PSNR improvement of 11.18 dB over GI-CI and 6.54 dB over BRG-EP across the six images.This superiority is attributed to our method's comprehensive exploitation of the relationships between the RGB channels and the establishment of appropriate reference relationships, which excel in local complexity computation and pixel prediction.However, the recently proposed OPC-PVO method could perform better than our method on the Peppers and Splash images.Our method adopts pixel-by-pixel data embedding after establishing the channel reference relationship.In contrast, the OPC-PVO method adopts a pixel-block embedding method, which controls the size of the pixel block to ensure higher pixel quality.To further validate the superiority of our proposed method in terms of image fidelity, we conduct an experiment in which we plot the PSNR values against the amount of data embedded.The experiment starts with 20,000 bits, increasing in increments of 2000 bits, as depicted in Figure 7.This figure illustrates the trend of PSNR values with respect to data embedding quantities.Our method consistently maintains a high PSNR value even as the volume of embedded data increases.This observation implies that our method effectively preserves image quality and mitigates the distortion resulting from data embedding.Our method outperforms other methods on six classic USC-SIPI images.For example, when we embed 30,000 bits of data into the Lena image, our method achieves a PSNR of 60.82 dB, significantly surpassing the performance of competing methods.Specifically, it outperforms CDPP [21], GF-CI [20], and ATDHM [23] by 1.74, 0.27, and 0.69 dB, respectively.These results underscore the efficacy of our approach in leveraging inter-channel correlations within the RGB color space and intra-channel pixel-level correlations.Ultimately, this approach enhances both image quality and data-hiding capabilities.The aforementioned experiments contribute to a more comprehensive understanding of the proposed algorithm's performance under varying data loads.Additionally, they assist in determining the optimal data embedding capacity for achieving the highest image fidelity.These findings will further substantiate the effectiveness of our algorithm in practical applications., GF-CI [20], AT-DHM [23], and the proposed method.

Performance Comparison on Kodak Images
To further validate the effectiveness of the proposed method in data embedding, we conduct an extensive experiment using the Kodak dataset.Specifically, we compare our method with two other methods, GF-CI [20] and CUE [25], by measuring PSNR values.We conduct the experiment with an embedding capacity of 30,000 bits and present the comparative results in Figure 8. Figure 8 illustrates a clear advantage of our proposed method in terms of fidelity for most images.In particular, when embedding 30,000 bits of data, our method outperforms GF-CI [20] and CUE [25] by an average of 1.10 dB and 0.93 dB, respectively, across 24 images.As shown in Figure 8, the PSNR value of the GF-CI algorithm on the 20-th image is significantly lower than that of the other images, mainly due to the complex texture features of the image.During pixel preprocessing, more auxiliary information is needed to prevent the pixel values from overflowing or underflowing, increasing the image's distortion.Notably, for the 6-th, 15-th, and 24-th images, our method exhibits slightly lower PSNR values compared to CUE.This is attributed to the prominent texture features in these images, where the accuracy of pixel prediction in our adaptive approach is not as high as in smoother images, leading to lower PSNR values in these cases.Nonetheless, when we consider overall performance, our method demonstrates superior image fidelity compared to existing methods.
In addition to PSNR values, we also use running time as an indicator to evaluate the performance of the algorithms.We compared six algorithms on 24 color images from the Kodak dataset, including GF-CI [20], GI-CI [22], BRG-EP [30], ATDHM [23], OPC-PVO [31], and the proposed method.Table 4 shows the running time of each algorithm when the embedding capacity is 10,000 bits.The running time in Table 4 is the average of multiple experiments, including the time for data embedding and extraction.From Table 4, we can see that the running time of the proposed method is close to that of the GF-CI and GI-CI methods and much lower than that of the BRG-EP, ATDHM, and OPC-PVO methods.These results indicate that our method has high running efficiency while achieving high-quality data embedding.Table 4. Comparisons in terms of running time (unit: ms) on Kodak dataset for GF-CI [20], GI-CI [22], BRG-EP [30], ATDHM [23], OPC-PVO [31], and the proposed method with the payload of 10,000 bits.

Performance Analysis
Our proposed CRM-based method has significant advantages in image fidelity, mainly benefiting from our full use of pixel value monotonicity between channels and establishing the reference mapping relationship between the three channels of the color image.At the same time, using the established CRM mode, we adaptively calculate the local complexity and predicted pixel values, improving the accuracy of pixel complexity calculation and prediction.Of course, our method also has some limitations.In some images with apparent texture features, selecting smooth pixels for data embedding could be more conducive, which reduces image fidelity.Improving the fidelity of texture images is the direction we need to improve and optimize in the future.

Conclusions
This paper proposes a CRM-based RDH method for color images, which includes the channel reference mapping method, the adaptive local complexity computation algo-rithm, and the adaptive pixel prediction strategy.The method proficiently utilizes reference mappings among the RGB channels to attain efficient data embedding and extraction.Furthermore, it adaptively selects suitable pixels for data embedding based on local complexity and predicted pixel values.The experimental results demonstrate that our approach consistently achieves high PSNR values across the majority of datasets, indicating its effectiveness in preserving the visual quality of images.The proposed method exhibits a notable advantage in image fidelity, primarily attributed to the establishment of a reference mapping relationship among the three color channels of the color image.This leverages the monotonicity of pixel values between channels, establishes channel correlations, and adaptively computes local complexity and predicted pixel values, enhancing both pixel complexity and prediction accuracy.Nevertheless, our method has some limitations.In images with pronounced texture features, it is challenging to select smooth pixels for data embedding, leading to a reduction in image fidelity.Therefore, we plan to explore local complexity calculation methods in our future research, such as the method based on local entropy, to improve the accuracy and robustness of local complexity.In addition, we intend to optimize the pixel prediction strategy, including improving the channel reference mapping method and adjusting the parameters of adaptive pixel prediction to enhance the accuracy and flexibility of pixel prediction.In our future research, we will devote ourselves to further improving the algorithm's performance and expanding the depth and breadth of our research.

Figure 1 .
Figure 1.P x and its eight adjacent pixels.
furnishes an illustrative depiction of the CRM-based RDH framework.

Figure 2 .
Figure 2.An overview of the CRM-based RDH framework.

Figure 3 .
Figure 3.The monotonicity of pixel values at the same positions in the RGB channels.

Figure 4 .
Figure 4.The local complexity contexts of C i,j and S i,j .

Figure 5 .
Figure 5. Adaptive pixel prediction context generation fot C i,j .

Figure 8 .
Figure 8. PSNR performance of the proposed method and existing methods on Kodak images with 30,000 bits of data.
denote the minimum value among W h , W v , W es , and W ws , expressed asW d = min{W h , W v , W es , W ws }.A threshold θ is set to quantify the local complexity of pixels.If (W − W d ) ≤ θ, the current pixel P c

Table 1 .
Six channel reference mapping modes for color images.
the pixel values of the RGB channels; M 1 , M 2 , M 3 , M 4 , M 5 , M 6 : six channel reference mapping modes;

Table 2 .
Comparisons in terms of PSNR (dB) on six classic color images with a payload of 20,000 bits.

Table 3 .
Comparisons in terms of PSNR (dB) on six classic color images with a payload of 40,000 bits.