Next Article in Journal
A Hybrid Simulated Annealing Approach for Loaded Phase Optimization in Digital Lasers for Structured Light Generation
Previous Article in Journal
The Dependence of Spatial Aliasing on the Amount of Defocus and Spherical Aberration in a Model Eye
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Implicit DFC: Blind Reference Frame Estimation in Screen-to- Camera Communication Using First-Order Statistics

Department of Electronic Engineering, Yeungnam University, Gyeongsan 38541, Republic of Korea
*
Author to whom correspondence should be addressed.
Photonics 2025, 12(10), 1004; https://doi.org/10.3390/photonics12101004
Submission received: 27 July 2025 / Revised: 25 August 2025 / Accepted: 26 August 2025 / Published: 13 October 2025

Abstract

Display-field communication (DFC) is an imperceptible screen-to-camera technology that embeds and recovers data from the frequency domain of an image frame. Conventional DFC requires a reference frame for each data frame to estimate the channel, a method that, while reliable, is not bandwidth-efficient. Similarly, iterative DFC requires the transmission of pilot symbols for channel estimation. In this paper, we propose an implicit DFC (iDFC) scheme that eliminates the need for reference frames by estimating them using the first-order statistics of the received image. The system employs discrete Fourier-transform-based subcarrier mapping and adds data directly to the frequency coefficients of the host image. At the receiver, statistical estimation enables blind channel equalization without sacrificing the data rate. The simulation results show that iDFC achieves an achievable data rate (ADR) of up to 1.52 × 10 5 bps, a significant enhancement of approximately 97% and 11% compared to conventional and iterative DFC schemes, respectively. Furthermore, the analysis reveals a critical trade-off between communication robustness and visual imperceptibility; allocating 70% of signal power to the image maintains high visual quality but results in a symbol error rate (SER) floor of 1.5 × 10 1 , whereas allocating only 10% improves the SER to below 10 2 at the cost of visible artifacts. The findings also identify QPSK as the optimal modulation order that maximizes the data rate, showing that higher-order schemes can be detrimental due to system impairments such as signal clipping. The proposed iDFC scheme presents a more efficient and robust solution for high-capacity DFC applications by balancing the competing demands of data throughput and visual fidelity.

1. Introduction

With the increasing number of screens and cameras in handheld devices, display-to-camera (D2C) communication [1,2] is becoming an alternative data transfer technology that can be compared to traditional QR codes [3,4] and 2D barcodes [5,6]. It can be considered as a sub-field of visible-light communication (VLC) [7,8] and optical camera communications (OCC) [9]. VLC primarily relies on data transfer through LEDs and photodiodes and has been explored for indoor wireless communications [10], underwater data transmission [11], and vehicular communication [12]. In OCC, data are transmitted by modulating the light intensity captured by the camera sensor. D2C communication involves data exchange between a visual display and a camera. In this approach, the display acts as a transmitter, emitting visual signals that are captured by the camera’s imaging sensor. The camera then decodes these signals to extract the embedded information. By utilizing the existing display and camera hardware in devices such as smartphones, tablets, and laptops, D2C communication has gained increasing interest and found applications in different domains, including mobile device interaction, augmented/virtual reality (AR/VR), interactive installations, digital signage, advertising, and even smart TVs. Furthermore, D2C communication can operate in real time, enabling interactive and dynamic data exchange between the display and the camera. This capability supports a wide range of potential applications across different domains, such as gaming, navigation, wayfinding, remote collaboration, education, retail, and training. For example, D2C communication enables a more immersive AR/VR experience. In museums or art exhibition settings, D2C communication allows users to access detailed information about a specific art.
D2C communication encompasses all kinds of screen–camera communication, including both obtrusive [13] and unobtrusive [1,2,14] modes. However, this paper focuses on unobtrusive D2C communication, where the data are embedded such that they remain invisible to the naked eye and can only be decoded by the receiver’s camera. This type of communication enables normal human viewing of the screen and data transfer simultaneously. It becomes challenging when embedding data in image frames unobtrusively, as it must not compromise the user’s normal viewing experience. In other words, the displayed image frames must not disrupt viewers who are not interested in receiving additional information. Hence, by employing the properties of the human visual system and image frames, data must be embedded covertly so that significant visual artifacts are not introduced on the screen. Data-embedding approaches in D2C communication can be broadly categorized into spatial- and spectral-domain-based mechanisms [15,16]. Spatial-domain embedding involves modifying the pixel values or arrangement within image or video frames. For instance, in MobiScan [14], data are first encrypted using different keys and then embedded into a set of frames by the visual pattern module proposed by the authors. In [1], the authors propose modulating the display brightness with a signal function that encodes the data to be transmitted. This method is simple to implement but more vulnerable to attacks [17,18]. Spectral-domain embedding involves manipulating the frequency or transform coefficients of a signal, typically using techniques such as the discrete Fourier transform, discrete cosine transform, or wavelet transform [19,20]. Information is embedded by modifying the coefficients in the spectral domain, rather than directly modifying the pixel values. Although more complex to implement, it offers increased robustness and security and is well suited for invisible embedding.
Display-field communication (DFC) [16] is a spectral-domain data-embedding scheme that embeds data into the frequency coefficients of an image frame [21,22]. Specifically, DFC utilizes a selected frequency range, called sub-bands, in the Fourier domain to embed data such that artifacts remain imperceptible to the human eye. This is achieved through multiplicative modulation of the image’s pixel values. Although DFC supports dual-mode, full-frame, visible communication, it requires a non-embedded reference frame for every data frame to enable decoding at the receiver [21]. This reference frame serves two purposes: (1) acting as a full-pilot frame for channel estimation and decoding and (2) preserving the visual integrity of the displayed content. However, transmitting one reference frame per data frame reduces the achievable data rate (ADR) by nearly 50%.
To address this, we previously proposed pilot-pixel-assisted modulation [23,24], which removes the need for reference frames by iteratively reconstructing them at the receiver. This approach involves transmitting pilot pixels and reconstructing the spectral image through an iterative process, utilizing the pilot symbol estimates. The main idea is to multiplex known pilot symbols with the unknown data pixels. The receiver first obtains tentative pixel estimates at the positions of the pilot pixels and then computes the data pixel estimates using interpolation. In addition to known pilot symbols, the receiver utilizes auxiliary pilot symbols, which are reliable symbol estimates obtained through piecewise cubic interpolation of the information symbols. Provided that an initial channel estimate is available, iterative channel estimation utilizes data symbol estimates fed back by the channel decoder. At subsequent iterations, channel estimates are refined by utilizing fed-back symbol estimates, which serve as auxiliary pilot symbols. The efficiency of the iterative approaches, however, depends on the reliability of the data estimates. While iterative DFC nearly doubles the data rate of conventional DFC [21], the pilot pixels still consume valuable bandwidth and reduce the data rate.
In this paper, we propose a blind reference frame estimation technique for DFC, referred to as implicit DFC (iDFC). Unlike prior approaches, iDFC requires neither reference frames nor explicit pilot symbols for channel estimation. Instead, we directly add the data symbols onto the frequency-domain coefficients, and the channel (reference frame) is estimated without sacrificing the data rate. The original image, which uses a fraction of the transmission power from the information sequence, is utilized for receiver channel estimation. Because there are no pilots, the reference frame estimation is carried out using the first-order statistics of the received image. At the receiver, the reference frame is reconstructed using a maximum likelihood (ML) estimator, which relies on statistical averaging of the received signal. This blind estimation allows for subsequent channel equalization and data recovery. The simulation results show that the proposed scheme offers increased bandwidth efficiency (overall data rate) and achieves performance far better than that of conventional and iterative DFC. We also verify that the proposed scheme shows robust performance and offers comparatively low computational complexity. The main contributions of this paper can be summarized as follows:
  • We propose a new iDFC framework that significantly improves bandwidth efficiency by completely eliminating the need for separate reference frames or explicit pilot symbols used in previous DFC methods.
  • We introduce a robust, blind channel estimation technique. This method reconstructs the reference frame at the receiver by using the first-order statistics of the received data-embedded image, allowing for effective channel equalization without any data-rate overhead.
  • We demonstrate through simulations that by removing the overhead associated with reference frames, the proposed iDFC scheme boosts the achievable data rate (ADR) by approximately 97% compared to conventional DFC and 11% compared to iterative DFC systems.
  • We provide an in-depth analysis of the system, quantifying the critical trade-off between visual imperceptibility and communication robustness (SER) as a function of the power allocation factor. Our analysis also identifies an optimal modulation order (QPSK) that maximizes the data rate under practical system impairments such as perspective distortion and channel noise.
The remainder of this paper is organized as follows. Section 2 details the proposed hidden-pilot-based DFC system model and the data-embedding process and describes the hidden pilot insertion technique, including the mechanism for data and power allocation. Section 3 describes the channel impairments—particularly the perspective distortion modeling and correction. In Section 4, we describe the data detection and decoding process. Section 5 presents a performance evaluation of the proposed iDFC scheme in terms of various system design parameters. We also compare our proposed method to conventional and iterative DFC schemes. Finally, Section 6 concludes the paper.

2. Overview of the Proposed iDFC Scheme

Our proposed end-to-end iDFC scheme for invisible data transmission from a display to a camera is depicted in Figure 1. The overall process can be broken down into four main stages: data embedding, transmitting, capturing and distortion correction, and data detection. The core innovation lies in eliminating the need for separate, bandwidth-consuming reference frames. Because no reference frame is required, this scheme can also be used for offline decoding—for example, extracting data from a photograph of a printed, data-embedded picture. First, the original display image is taken from its standard spatial domain and converted into the frequency domain. The result is a representation of the image in terms of its frequency coefficients. The information to be sent is converted into modulated data symbols. The key step is the additive embedding process, where the data symbols are scaled according to a power ratio and then added directly to the frequency coefficients of the original image. This process is known as spectral superposition. An inverse DFT is then performed to transform the image back into the spatial domain. The resulting data-embedded frame is visually indistinguishable from the original, making the embedded data completely unobtrusive to a human viewer. The data-embedded image is displayed on a screen, which acts as the transmitter in the communication link.
At the receiver, a camera captures the image being displayed on the screen. This process inevitably introduces noise and geometric distortions. Before the data can be read, the captured frame must be digitally processed. A crucial step is correcting for perspective distortion, which occurs when the camera is not perfectly parallel to the screen. The system models this distortion and applies a transformation to restore the image’s original rectangular geometry. Other channel effects, such as brightness and background noise, are also handled at this stage. The receiver takes the corrected image and transforms it back into the frequency domain. It then computes the first-order statistics of the received image to estimate the characteristics of the reference image—that is, how the image was distorted during transmission and capture. Using the reference frame estimate, the receiver can now accurately reverse the decoding of the data symbols. These symbols are then demodulated to reconstruct the original information.

2.1. System Model

The signal flow diagram of our proposed iDFC system is illustrated in Figure 2. The process can be divided into the transmitter, the channel, and the receiver stages.

2.1.1. Transmitter

The process begins with the binary information data. These data are fed into a modulator (e.g., a QAM modulator), which maps the bits to complex symbols. The resulting matrix of modulated symbols is denoted by X . To ensure that the final spatial-domain image is real-valued, a Hermitian symmetry constraint is applied to X . Concurrently, the original input image, a spatial-domain signal denoted by I t , is transformed into its frequency-domain representation, I F , by applying a column-wise discrete Fourier transform (DFT). The core of the scheme lies in the data-embedding block. Here, the frequency-domain image I F is additively combined with the scaled data matrix X . This spectral superposition results in a new frequency-domain matrix D F , which now contains both the original image information and the hidden data. Finally, to prepare the image for display, the combined matrix D F is converted back into the spatial domain using an inverse DFT (IDFT), yielding the final data-embedded image, D t . This image, which is visually indistinguishable from the original, is then sent to the display screen, which acts as the transmitter.

2.1.2. Channel and Receiver

The communication link between the display and the camera is subject to real-world impairments, primarily perspective distortion (due to the camera’s angle) and additive white Gaussian noise (AWGN) from ambient light and camera sensor noise. The camera at the receiver captures the distorted screen, producing the received spatial-domain image, Y t . After correcting for geometric distortions, the image Y t is converted to the frequency domain via DFT, resulting in Y F . The key innovation occurs in the channel estimation block. The received signal Y F is processed using our proposed statistical method to generate a blind estimate of the transmitted reference frame’s spectrum. This estimated reference is denoted by D ^ F . Finally, in the data retrieval and demodulator block, the estimated reference D ^ F is subtracted from the received signal Y F to isolate the embedded data symbols. These symbols are then demodulated to recover the original output data, completing the communication process.

2.2. Data Embedding

The first step in our data-embedding process is to convert the spatial-domain image, which consists of pixel intensity values, into its frequency-domain representation. This is achieved by applying a one-dimensional DFT to each column of the image matrix independently. The transformation for the entire image I t is expressed as follows [21]:
I F = i F 1 , i F 2 , , i F Q = F · i t 1 , F · i t 2 , , F · i t Q = F · I t ,
where I t R P × Q is the original spatial-domain image with P rows and Q columns. The resulting matrix, I F C P × Q , is its frequency-domain representation. Here, F C P × P represents the DFT matrix, and the vectors i t q and i F q represent the q-th columns of the spatial image I t and frequency image I F , respectively. The elements of the DFT matrix are given by
F p , k = 1 P e j 2 π p k P , for p , k { 0 , 1 , , P 1 } .
In the resulting frequency-domain image I F , each point represents a particular frequency contained in the spatial-domain image. The results of the 1D-DFT exhibit low-frequency components on both sides of each column vector i F q , while high-frequency components are symmetrically located in the central region. In other words, the low-frequency components, which contain most of the image’s structural information, are located at the top and bottom of each column vector, while the high-frequency components, which correspond to finer details and textures, are situated in the center.
For data embedding, we consider high-frequency sub-bands of the spectral-domain image. This choice is motivated by the fact that most of an image’s visual characteristics reside at lower frequencies, while details and noise tend to be present at higher frequencies. Since the human visual system is more sensitive to lower frequencies, the use of low-frequency components for data embedding would create visible artifacts and distract the human observer [21]. If X = x 1 , x 2 , , x Q represents a P × Q data matrix to be embedded, the structure for each column vector x q is defined as follows:
x q = 0 1 × ( κ 1 ) s q T 0 1 × γ flip s q * T 0 1 × ( κ 2 ) T ,
where the parameters are defined as follows:
  • κ : The starting row index for data embedding in the frequency domain.
  • s q : Column vector [ s q ( 1 ) , , s q ( L ) ] T containing the L complex data symbols to be embedded in the q-th column of the image.
  • L: The number of data symbols embedded per column.
  • γ : The number of zero-padding elements used for spacing between the data and their conjugate, calculated as γ = P 2 κ 2 L + 3 .
  • flip ( s q * ) : The conjugate symmetric counterpart of the data vector, given by [ s q * ( L ) , , s q * ( 1 ) ] T . This is required to ensure that the IDFT of the combined signal results in a real-valued image suitable for display.
  • 0 1 × N : A row vector of N zeros used for padding to ensure that data are only placed in the designated high-frequency sub-bands.
The parameters κ and L must satisfy κ + L = P / 2 + 2 to properly position the data in the high-frequency region of the spectrum. In this manner, the data structure encompasses an L × Q rectangular region, as well as an additional L × Q conjugate symmetric region in the frequency-domain representation. The remaining elements are populated with 0 s to ensure that only the specified high-frequency components are altered during the additive embedding process.
Figure 3a presents the original image, a grayscale Lena image, used for data embedding in our simulations. Figure 3b illustrates the location of the data-embedding region within the high-frequency sub-bands. This image is a map of the 2D frequency domain, showing exactly where the data are placed. The two horizontal white bands represent the data-embedding regions, while the black areas correspond to the frequency coefficients that are left unchanged. The bands are located in the high-frequency areas, avoiding the low-frequency information at the top and bottom of the frequency map.
In the proposed scheme, the original frequency-domain coefficients of the image itself ( I F ) serve as the unknown channel (reference image frame). The data embedding is performed by additively combining the modulated data symbols directly with image coefficients at high-frequency locations. The matrix of data symbols to be embedded, X , is constructed as defined by the structure in Equation (3). To control the embedding strength, these data symbols are scaled by a power allocation factor, α . The final data-embedded image matrix, D F , in the frequency domain is then generated by adding the scaled data matrix to the original frequency-domain image:
D F = α I F + 1 α X ,
where α is the image-to-data power ratio ( 0 < α < 1 ). This factor determines the power of the embedded data relative to the power of the original image content. By keeping α high, we ensure that the embedding remains imperceptible. Once the data-embedded frequency-domain matrix D F is created, it is converted back to the spatial domain using the following IDFT operation:
D t = F H · D F ,
where ( · ) H denotes the Hermitian transpose operator. Note that the above equation introduces a nonlinear distortion during the operations of clipping and quantization.
The primary advantage of the proposed iDFC scheme is its high spectral efficiency. By eliminating the need for separate reference frames, the ADR is more than doubled compared to conventional DFC. This creates a fundamental trade-off: to maintain visual imperceptibility, the power allocated to the data must be kept low relative to the image content, which can lead to a higher symbol error rate (SER). A key benefit of our approach is that all the pixels are now used to transmit the data, and the receiver performs estimation across all data-bearing symbols, which eliminates the need for channel interpolation and makes the scheme robust in dynamic channel conditions. In other words, there is no need to send the reference frames, as in conventional DFC [21], or to insert pilot symbols, as in iterative DFC [23].

3. Perspective Distortion

In practical DFC scenarios, perspective distortion arises when the camera is not perfectly aligned with the display. This misalignment causes geometric deformation in the captured image, which can misplace embedded data in the spectral domain and degrade symbol recovery. The geometric relationship between the original display’s coordinates and the captured image’s coordinates can be modeled using a 2D projective transformation, also known as a homography matrix H . The general form of a homography matrix H R 3 × 3 is given as [25]
H = h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 h 33 .
The eight parameters of H are calculated by finding at least four corresponding points between the source display and the captured image (e.g., the corners of the screen). This transformation is applied to each pixel coordinate p = [ x , y , 1 ] T in the display image to obtain its corresponding coordinate p in the distorted image as follows:
p = H · p .
To decode the embedded information, these display boundaries must first be detected in the captured frame using computer vision techniques such as Harris corner detection. Once the corner points are found and the homography H is estimated, the receiver recovers the original data alignment by applying the inverse transformation H 1 to warp the distorted frame back to its original rectangular grid. During this rectification process, interpolation is used to estimate pixel values at non-integer coordinates. The preprocessed image can then be demodulated with improved reliability. While the proposed iDFC method exhibits partial robustness to minor distortions, significant warping can still degrade the statistical estimation used for channel correction. Therefore, accurate rectification is a crucial preprocessing step.
Figure 4 demonstrates the impact of perspective distortion on a D2C link and the effectiveness of the subsequent geometric restoration. The distortion in Figure 4a, caused by a projective transformation, skews the image’s rectangular grid into a trapezoidal shape. The restored image (cf. Figure 4b) recovers its correct alignment by compensating for this deformation. This correction is a crucial pre-processing step for a practical DFC receiver, as it ensures the integrity of the spatial data before the information is extracted.

4. Data Detection

The key advantage of the implicit pilot scheme is the ability to use all pixels for both channel estimation and data detection. However, data signals may cause interference in the channel estimation process. This interference degrades the estimation quality and can create a correlation between the channel estimates and the data. Note that realizing an ideal separation between the data and channel coefficients at the receiver is difficult [26], and the recovered reference image signal inevitably consists of data interference, which, in turn, affects the accuracy of the channel estimation. For our analysis, we assume a simplified channel model. We consider a perfect alignment between the data-transmitting screen and the camera. That is, all the light-emitting pixels from the screen are at the focus of the camera pixels, resulting in no loss of data signal energy on the pixel. Therefore, in the given scenario, we consider the AWGN channel as the loss in signal energy on a pixel; this is also attributed to the noise in that pixel. Moreover, the AWGN also captures the primary impairment from the camera sensor and background illumination, which corrupts the intensity values in the captured image. Therefore, the received spatial-domain image Y t can be expressed as
Y t = D t + N t ,
where N t is the AWGN matrix. The received spatial-domain image is then converted to the frequency domain as
Y F = F · D t + F · N t = D F + N F .
All subsequent processing is performed in the frequency domain.
Since data are embedded in only L specific rows of the transmitted image, we can analyze these rows individually. Let the set of row indices be { l 1 , l 2 , , l L } . The relationship for each row vector is given as
y l k = d l k + n l k , for k = 1 , 2 , , L .
Assuming that the noise samples are i.i.d. Gaussian random
f Y l k ( y l k ) = 1 2 π σ 2 Q / 2 exp 1 2 σ 2 y l k d l k 2 .
The joint p.d.f. of L observations is given as
f Y l 1 Y l 2 Y l L ( y l 1 y l 2 y l L ) = 1 2 π σ 2 L Q / 2 exp 1 2 σ 2 k = 1 L y l k d l k 2 .
Note that d l k is an unknown and is the summation of data and the image frequency-domain coefficient. Now, considering the above joint p.d.f. as a function of the unknown parameter d l k gives us the likelihood function. In this paper, we use the framework of the log-likelihood estimate to find the values of the unknown parameter for which the likelihood function is at the maximum. Hence, Equation (12) is defined as the likelihood function P ( y l ; d l ) , defined as the likelihood of observation y l parameterized by d l , where y l = y l 1 , y l 2 , , y l L T . Hence, the log likelihood function is defined as
L y ; d = ln P ( y ; d ) = ln 1 2 π σ 2 L Q / 2 exp 1 2 σ 2 k = 1 L y l k d ¯ 2 ,
assuming d l 1 d l 2 d l L = d ¯ , where d ¯ is the constant row vector that we want to estimate. Now, maximizing the log-likelihood function with respect to the single unknown vector d ¯ yields the maximum likelihood (ML) estimate as follows (see Appendix A):
d ^ M L = 1 L k = 1 L y l k .
This shows that the ML estimate of the transmitted signal is the sample mean of the observed row vectors. As proven in Appendix B, this estimator is unbiased, and its variance is inversely proportional to L, which mathematically confirms how averaging reduces the effect of noise. Expanding the above mathematics to the whole received image will yield an estimated reference frame spectrum D ^ F C P × Q , where the data-embedded rows are replaced with the estimate d ^ M L . Then, using the subtraction receiver, we can extract the data components as follows:
X ^ F = Y F D ^ F .
After the data symbols are extracted, QAM demodulation is applied to recover the transmitted bits.

5. Simulation Results

For the simulations, we assume perfect frame synchronization between the display and camera. Moreover, the camera’s frame rate is assumed to be greater than or equal to the display’s frame rate to ensure that every frame is captured successfully. We use a standard 256 × 256 grayscale Lena image as the host content. The input information sequence is composed of i.i.d. equiprobable QAM symbols. A total of 20 frequency sub-bands are allocated for data transmission, positioned symmetrically within the high-frequency region. The power of the embedded data is controlled by the data-to-signal power ratio, α , which is varied to analyze its impact on performance. Moreover, the DFC channel was simulated as an AWGN channel. A subtraction receiver was implemented. This receiver first estimates the reference frame from the received signal using the ML estimate. The estimation is carried out using statistical averaging of the symmetric frequency rows, under the assumption that the embedded data are zero-mean. The subtraction receiver then subtracts the estimated pilot spectrum to isolate the data components, which are subsequently demodulated. The default simulation parameters are mentioned in Table 1.
First, we look at the effect of perspective distortion. As shown in Figure 5, the ‘after correction’ curve has a waterfall shape, and the error floor is higher than the ‘before correction’ case. This performance gap is caused by the pixel interpolation that occurs during the distortion correction process. This shows that the perspective correction step is an absolutely necessary step in DFC. Without correction, the geometric distortion is so severe that the decoder cannot locate the data correctly, making its performance no better than a random guess.
Next, let us look at the visual imperceptibility of data embedding for different image-to-data power ratios, α , depicted in Figure 6. An alpha of 0.1 means that 10% of the power is allocated to the image (and 90% to data). This visual comparison is a qualitative assessment of imperceptibility—how well the hidden data are concealed from the human eye. When the image power is low, the most visible distortion is observed in Figure 6a. In other words, since the data power is high (90%), the embedded data dominates the frequency components, resulting in visible artifacts. This confirms that very low image power compromises image quality. On the other hand, if we increase the image power, there is a significant improvement in imperceptibility. In Figure 6b,c, distortion is still visible upon close inspection, but the structure of the original image is better preserved. Finally, Figure 6d, which has 70% image power, has the least perceptible distortion. The image appears smooth and nearly identical to the original (cf. Figure 3a).
Figure 7 illustrates the SER performance of the proposed system as a function of the SNR for various image power allocation ratios, α . The plot demonstrates a clear and direct relationship between the image power allocation and system performance. As the α increases from 0.1 to 0.7, the overall SER performance gets progressively worse at all SNR values. This is because as α is increasing, the power loaded onto the data symbols decreases simultaneously, resulting in a worse SER. In other words, when the image power is low, the data power is high. This strong data signal is robust against channel noise, resulting in a low SER. More importantly, the low overall signal amplitude causes minimal clipping, so the signal distortion is low. In addition, the results clearly demonstrate the impact of power allocation between the image signal and the data signal. First, we observe that all curves exhibit an error floor. This floor arises from inherent system errors, including clipping and quantization noise after IDFT, imperfect channel estimation using image averaging, and image-data interference in non-orthogonal embedding. The SER improves as the SNR increases up to a certain point, after which the curves flatten out. Further increases in the SNR do not improve the SER. This indicates that at high SNRs, the performance is limited by a factor other than channel noise.
The above results reveal a trade-off between visual imperceptibility and detection performance in the proposed iDFC system. As illustrated in Figure 7, reducing the image power significantly improves the SER due to the stronger data signal. However, this comes at the cost of noticeable visual artifacts (cf. Figure 6). Conversely, increasing the image power enhances visual quality by preserving the host image spectrum, but it weakens the embedded data, leading to an increased SER. This trade-off necessitates careful tuning of the power allocation factor α to strike a balance between communication reliability and visual transparency, depending on the target application constraints.
Next, we computed the ADR of the system, which represents the effective throughput of useful information. Mathematically, it can be related to the maximum theoretical data rate and SER as follows:
ADR = R max ( 1 SER ) ,
where R max is the maximum theoretical data rate in bits per second (bps). It can be determined by the system’s configuration as follows:
R max = N data log 2 ( M ) R f ,
where N data represents the total number of data-carrying pixels per frame, M denotes the modulation order, and R f stands for the frame rate of the display in frames per second.
Figure 8 illustrates the ADR as a function of the SNR for different image power allocations. First, it is observed that for every power allocation, the ADR increases with the SNR until it reaches a saturation point at high SNR values. This indicates that beyond a certain point, improving the channel quality with a higher SNR yields no further increase in data throughput. This saturation occurs because the SER hits an error floor, as shown in Figure 7. At high SNRs, the performance is no longer limited by channel noise but by the various distortions caused by the system behavior. Second, the plot shows a clear inverse relationship between the image power and the maximum ADR. The configuration with the lowest image power achieves the highest ADR and vice versa. A lower image power means a higher data power. This leads to a better SER and, hence, ADR. In other words, the system successfully transmits more useful, error-free data.

5.1. Data Rate vs. Imperceptibility Trade-Off

To provide clear, quantitative guidance for selecting the power allocation factor α , we evaluated the trade-off between the ADR and visual imperceptibility. We used the peak signal-to-noise ratio (PSNR) as the objective metric for image quality, where a higher PSNR value indicates less distortion and better imperceptibility. Figure 9 presents the Pareto frontier for our proposed iDFC system, illustrating the fundamental trade-off. The downward slope of the curve clearly demonstrates the inverse relationship between data throughput and visual fidelity. Each point on the curve represents an optimal operating point for a given image-to-data power ratio, α . The top-left portion of the curve corresponds to low α values. At α = 0.1, the system achieves its maximum data rate, exceeding 1.5 × 10 5 bps. This is because a low α allocates most of the signal power to the data, making the transmission robust against noise and minimizing the symbol error rate. However, this high-power data signal significantly distorts the host image, resulting in the lowest image quality, with a PSNR below 30 dB. On the other hand, the bottom-right portion of the curve corresponds to high α values. At α = 0.9, the system achieves the best visual quality, with a PSNR approaching 48 dB, making the embedded data virtually invisible to the human eye. This high fidelity is achieved by allocating most of the power to the image itself. The trade-off is a significantly weaker data signal, which leads to a higher error rate and reduces the ADR to its lowest point.
This Pareto frontier serves as a valuable practical guide for system implementation. The choice of α depends directly on the priorities of the target application. For applications where visual quality is critical, such as digital art displays or premium advertising, we should choose a high α value (e.g., α 0.7 ) to ensure high imperceptibility. For applications where maximizing the data rate is paramount and slight visual artifacts are acceptable, a low α value (e.g., α ≤ 0.3) is the optimal choice. A balanced approach is represented by the “knee” of the curve (around α = 0.5), which offers a strong data rate of approximately 1.4 × 10 5 bps while maintaining a respectable PSNR of 38 dB.

5.2. Comparative Evaluation with Related Schemes

Figure 10 compares the achievable data rates of three DFC schemes: conventional DFC [21], iterative DFC [23], and the proposed iDFC. For a fair comparison, the best-case scenarios of all schemes that provide the lowest SERs are considered. In particular, the results of the seventh iteration are utilized for the iterative DFC [23], and α = 0.1 is utilized for the iDFC scheme. The results demonstrate that the iDFC method achieves the highest throughput, making it the most bandwidth-efficient DFC scheme. As mentioned above, in the conventional DFC, a separate, unaltered reference image frame is explicitly inserted between every data-embedded image frame to support data decoding, which resulted in 50% image overhead. Therefore, it uses half of its transmission time just to send reference information. In iterative DFC, we eliminate the use of reference frames by embedding images within the same frame and using auxiliary image refinement at the receiver to enhance image estimation over multiple iterations. This method is a significant improvement over the conventional DFC because every frame carries data, effectively improving the potential throughput, particularly in the high-SNR region. However, its performance is still limited by the use of dedicated pixels as image pixels, resulting in 10% image overhead. Moreover, the scheme is computationally complex due to the overhead of the iterative process, where each iteration has matrix operations. In the proposed iDFC scheme, as we are not transmitting explicit pilot symbols and the implicit frequency-domain image works as a reference image, there are no dedicated pixels working as pilot pixels. In other words, the data are multiplexed with the implicit image pixels in all sub-bands. We can observe from the figure that it significantly boosts the data rate, particularly in the low-to-mid-SNR region (10–40 dB). Even though the embedded data act as interference, through efficient power utilization, this interference can be reduced.
To contextualize the proposed iDFC, we compared it against widely referenced screen-to-camera communication schemes in Table 2. We observe that iDFC achieves the highest ADR in the low-to-mid-SNR region while incurring zero image overhead.
Figure 11 illustrates the SER performance of the proposed iDFC system across different modulation orders (M). It can be observed that a higher M yields a higher SER. This is due to the denser constellation points in higher-order QAM, which are more susceptible to noise and distortion. In addition, the SER saturates at high SNRs, forming an error floor due to the presence of non-Gaussian errors. Overall, lower-order modulations, such as BPSK ( M = 2 ) and QPSK ( M = 4 ), are robust and exhibit sharp SER reductions above 20 dB, making them suitable for low-SNR or high-fidelity applications. Although higher-order modulation allows more bits to be transmitted per symbol, maintaining the same average signal energy requires constellation points to be placed closer together. As a result, the Euclidean distance between adjacent symbols decreases, making the system more prone to symbol errors in the presence of noise or distortion. These effects become more pronounced as the SNR increases. The results highlight the trade-off between spectral efficiency and robustness in modulation design for DFC systems.
Figure 12 illustrates how the ADR varies with both the SNR and the modulation order. The data rate initially increases with both the SNR and M but then drops beyond M = 4 at high SNRs. This results in a ridge-like peak, showing that there is an optimal M for each SNR that maximizes the data rate. For a fixed modulation order, increasing the SNR improves the data rate until it saturates. This is because the SER drops with the SNR, allowing more symbols to be decoded correctly. Regarding the effect of M, increasing M increases the spectral efficiency (more bits per symbol). However, beyond a point ( M = 4 ), symbol errors due to noise and tightly packed constellation points dominate, reducing the net data rate. This drop is more visible at higher M and moderate SNRs, where the decoder struggles to distinguish adjacent constellation points. This underscores the trade-off between spectral efficiency and robustness in practical iDFC systems.

5.3. Complexity Analysis

The primary computational difference between the conventional DFC [21] and our proposed iDFC lies at the receiver. The complexity is dominated by the DFT operations, which, using the fast Fourier transform algorithm, is O ( N log N ) for an N-point transform [30,31,32]. For a P × Q image processed column-wise, the complexity becomes O ( Q P log P ) . Table 3 breaks down the key computational steps at the receiver for decoding one data frame.
As shown in Table 3, the dominant computational load for both schemes comes from the DFT operation. The subsequent estimation and data retrieval steps (e.g., averaging, division, subtraction) have a much lower complexity of O ( L Q ) , where L is the number of data-bearing rows, and these are negligible compared to the DFT. The key difference is that conventional DFC requires two full DFT operations for every data frame it decodes. In contrast, our proposed iDFC scheme requires only a single DFT operation. By eliminating the need for a separate reference frame and instead performing a computationally lightweight statistical estimation, iDFC reduces the dominant computational load by approximately 50%. This makes our proposed method significantly more efficient and better suited for real-time applications on resource-constrained devices.

5.4. Practicality Considerations

The proposed iDFC scheme is designed considering several practical advantages, which can be described as follows:
  • The system operates using the existing, off-the-shelf screens and cameras found in billions of consumer devices such as smartphones, monitors, and digital signage. No specialized or additional hardware is required, making the system widely deployable and cost-effective.
  • A unique practical benefit of iDFC is its ability to support offline decoding. Because the system does not depend on a separate, temporally linked reference frame, data can be extracted from a single, static photograph of a data-embedded screen or even a printed picture. This is not possible with conventional DFC.
  • The system design explicitly accounts for perspective distortion, a critical impairment in real-world scenarios. As demonstrated, the geometric correction step is crucial and effectively restores the signal, enabling robust communication even when the camera is not perfectly aligned with the screen.
  • As detailed in the previous section, iDFC is significantly more computationally efficient than conventional methods, making it better suited for real-time processing on resource-constrained mobile devices.

6. Conclusions

This paper introduced and evaluated a novel implicit DFC scheme, which is a novel self-referenced screen-to-camera communication framework that leverages the inherent spectral characteristics of display images to enable imperceptible data embedding and blind channel estimation. This approach eliminates the need for the separate reference frames used in conventional DFC or the explicit, sparse pilots and refinement steps used in iterative DFC. The statistical properties of the host image content itself are used for the reference frame estimation. An analysis of perspective distortion effects confirmed the importance of geometric correction in practical deployment scenarios. The simulation results demonstrate that the system achieves a favorable trade-off between imperceptibility and the SER, with optimal performance occurring at moderate image power allocations. Moreover, the ADR was shown to be sensitive to both modulation order and image power, with QPSK providing the best balance between reliability and throughput. Overall, the proposed scheme is an elegant and highly efficient solution for robust, high-speed data communication in unobtrusive screen-to-camera communication.

Author Contributions

Conceptualization, P.S. and S.-Y.J.; methodology, P.S.; software, P.S.; validation, P.S. and S.-Y.J.; formal analysis, P.S.; investigation, P.S.; resources, S.-Y.J.; data curation, P.S.; writing—original draft preparation, P.S.; writing—review and editing, P.S. and S.-Y.J.; visualization, P.S. and S.-Y.J.; supervision, S.-Y.J.; project administration, S.-Y.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Gyeongsangbuk-do RISE (Regional Innovation System & Education) Regional Growth & Innovation LAB.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author(s).

Acknowledgments

During the preparation of this manuscript/study, the author(s) used ChatGPT 4o for the purposes of generating text, assisting in analysis, and interpretation. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. ML Estimate

Appendix A.1. Derivation for the ML Estimate

This appendix details the derivation of the ML estimate of the transmitted signal vector based on L observed row vectors. The joint p.d.f. of L i.i.d. Gaussian row vectors, which serves as the likelihood function P , is expressed as
P ( y ; d ) = 1 2 π σ 2 L Q / 2 exp 1 2 σ 2 k = 1 L y l k d l k 2 .
Now, under the assumption d l 1 = d l 2 = = d l L = d ¯ , the likelihood function can be expressed as
P ( y ; d ¯ ) = 1 2 π σ 2 L Q / 2 exp 1 2 σ 2 k = 1 L y l k d ¯ 2 .
To simplify the maximization problem, we take the natural logarithm of the likelihood function. Finding the maximum of the log-likelihood function is equivalent to maximizing the original function. Therefore, we have
L ( d ¯ ; y ) = ln P ( y ; d ¯ ) = ln 1 2 π σ 2 L Q / 2 + ln exp 1 2 σ 2 k = 1 L y l k d ¯ 2 = L Q 2 ln ( 2 π σ 2 ) 1 2 σ 2 k = 1 L y l k d ¯ 2 .
Maximizing the function is equivalent to minimizing the sum of squared errors as follows:
d ¯ min = arg min d ¯ k = 1 L y l k d ¯ 2 .
We find the minimum by taking the gradient of the sum of squared errors with respect to d ¯ and setting it to the zero vector as follows:
d ¯ k = 1 L y l k d ¯ 2 = 0 k = 1 L d ¯ y l k d ¯ 2 = 0 .
Using the gradient identity x a x 2 = 2 ( a x ) , we get
k = 1 L 2 ( y l k d ¯ ) = 0 k = 1 L y l k k = 1 L d ¯ = 0 .
Finally, solving for d ¯ gives us the ML estimate d ^ ML :
d ^ ML = 1 L k = 1 L y l k .

Appendix B. First-Order Statistics of the ML Estimate

To check the reliability of the ML estimate, we have to understand its behavior, primarily in terms of first-order statistics.

Appendix B.1. Mean

To find the mean of ML estimate, we take the expectation E [ · ] of both sides of Equation (A7) as follows:
E [ d ^ ML ] = E 1 L k = 1 L y l k = 1 L k = 1 L E [ y l k ] .
Using the signal model y l k = d l k + n l k , we can write
E [ y l k ] = E [ d l k ] + E [ n l k ] = d l k ,
as d l k is deterministic, and the noise n l k has zero mean. From our MLE derivation, we assumed d l k = d ¯ for all k. Therefore,
E [ y l k ] = d ¯ .
Substituting this result back into our main equation gives
E [ d ^ ML ] = 1 L k = 1 L d ¯ = d ¯ .
From Equation (A11), we can observe that the mean of our estimate is equal to the true value. Therefore, the sample mean is an unbiased estimator.

Appendix B.2. Variance

The variance of a vector estimate is its covariance matrix. From Appendix B.1, using the fact that the estimator is unbiased, we define covariance as
Var ( d ^ ML ) = E ( d ^ ML d ¯ ) ( d ^ ML d ¯ ) T .
The error term ( d ^ ML d ¯ ) can be expressed as
d ^ ML d ¯ = 1 L k = 1 L y l k d ¯ .
Substituting the signal model y l k = d ¯ + n l k , we get
d ^ ML d ¯ = 1 L k = 1 L ( d ¯ + n l k ) d ¯ = 1 L L d ¯ + 1 L k = 1 L n l k d ¯ = 1 L k = 1 L n l k .
As can be seen from Equation (A14), the error of the estimate is simply the average of the noise vectors. Therefore, the variance can be expressed as
Var ( d ^ ML ) = E 1 L k = 1 L n l k 1 L j = 1 L n l j T = 1 L 2 k = 1 L j = 1 L E n l k n l j T .
Since the noise vectors n l k are independent for k j , their cross-correlation is E n l k n l j T = 0 . Therefore, only the terms where k = j remain are the following:
Var ( d ^ ML ) = 1 L 2 k = 1 L E n l k n l k T .
Each term E n l k n l k T is the covariance matrix of the noise, i.e., σ 2 I . Thus,
Var ( d ^ ML ) = 1 L 2 k = 1 L σ 2 I = σ 2 L I .
This confirms that by averaging L rows, the noise variance of the estimate is reduced by a factor of L.

References

  1. Jo, K.; Gupta, M.; Nayar, S.K. DisCo: Display-camera communication using rolling shutter sensors. ACM Trans. Graph. (TOG) 2016, 35, 1–13. [Google Scholar] [CrossRef]
  2. Wang, A.; Li, Z.; Peng, C.; Shen, G.; Fang, G.; Zeng, B. Inframe++: Achieve simultaneous screen-human viewing and hidden screen-camera communication. In Proceedings of the 13th Annual International Conference on Mobile Systems, Applications, and Services, Florence, Italy, 18–22 May 2015; pp. 181–195. [Google Scholar]
  3. Chu, H.K.; Chang, C.S.; Lee, R.R.; Mitra, N.J. Halftone QR codes. ACM Trans. Graph. (TOG) 2013, 32, 1–8. [Google Scholar] [CrossRef]
  4. Garateguy, G.J.; Arce, G.R.; Lau, D.L.; Villarreal, O.P. QR images: Optimized image embedding in QR codes. IEEE Trans. Image Process. 2014, 23, 2842–2853. [Google Scholar] [CrossRef]
  5. Chen, C.; Zhou, B.; Mow, W.H. RA code: A robust and aesthetic code for resolution-constrained applications. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 3300–3312. [Google Scholar] [CrossRef]
  6. Chen, C.; Huang, W.; Zhou, B.; Liu, C.; Mow, W.H. PiCode: A new picture-embedding 2D barcode. IEEE Trans. Image Process. 2016, 25, 3444–3458. [Google Scholar] [CrossRef]
  7. Chi, N.; Zhou, Y.; Wei, Y.; Hu, F. Visible light communication in 6G: Advances, challenges, and prospects. IEEE Veh. Technol. Mag. 2020, 15, 93–102. [Google Scholar] [CrossRef]
  8. Matheus, L.E.M.; Vieira, A.B.; Vieira, L.F.M.; Vieira, M.A.M.; Gnawali, O. Visible Light Communication: Concepts, Applications and Challenges. IEEE Commun. Surv. Tutor. 2019, 21, 3204–3237. [Google Scholar] [CrossRef]
  9. Cahyadi, W.A.; Chung, Y.H.; Ghassemlooy, Z.; Hassan, N.B. Optical camera communications: Principles, modulations, potential and challenges. Electronics 2020, 9, 1339. [Google Scholar] [CrossRef]
  10. Mapunda, G.A.; Ramogomana, R.; Marata, L.; Basutli, B.; Khan, A.S.; Chuma, J.M. Indoor visible light communication: A tutorial and survey. Wirel. Commun. Mob. Comput. 2020, 2020, 1–46. [Google Scholar] [CrossRef]
  11. Ali, M.F.; Jayakody, D.N.K.; Li, Y. Recent trends in underwater visible light communication (UVLC) systems. IEEE Access 2022, 10, 22169–22225. [Google Scholar] [CrossRef]
  12. Memedi, A.; Dressler, F. Vehicular visible light communications: A survey. IEEE Commun. Surv. Tutor. 2020, 23, 161–181. [Google Scholar] [CrossRef]
  13. Han, H.; Xie, K.; Wang, T.; Zhu, X.; Zhao, Y.; Xu, F. RescQR: Enabling Reliable Data Recovery in Screen-Camera Communication System. IEEE Trans. Mob. Comput. 2023, 23, 3510–3522. [Google Scholar] [CrossRef]
  14. Zhang, X.; Liu, J.; Ba, Z.; Tao, Y.; Cheng, X. MobiScan: An enhanced invisible screen-camera communication system for IoT applications. Trans. Emerg. Telecommun. Technol. 2022, 33, e4151. [Google Scholar] [CrossRef]
  15. Tamang, L.D.; Kim, B.W. Spectral domain-based data-embedding mechanisms for display-to-camera communication. Electronics 2021, 10, 468. [Google Scholar] [CrossRef]
  16. Singh, P.; Kim, Y.J.; Kim, B.W.; Jung, S.Y. Display Field Communication: Enabling Seamless Data Exchange in Screen–Camera Environments. Photonics 2024, 11, 1000. [Google Scholar] [CrossRef]
  17. Hussain, M.; Wahab, A.W.A.; Idris, Y.I.B.; Ho, A.T.; Jung, K.H. Image steganography in spatial domain: A survey. Signal Process. Image Commun. 2018, 65, 46–66. [Google Scholar] [CrossRef]
  18. Fridrich, J. Steganography in Digital Media: Principles, Algorithms, and Applications; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  19. Yasmeen, F.; Uddin, M.S. A novel watermarking scheme based on discrete wavelet transform-singular value decomposition. Secur. Priv. 2022, 5, e216. [Google Scholar] [CrossRef]
  20. Zainol, Z.; Teh, J.S.; Alawida, M.; Teh, J.S. A new chaotic image watermarking scheme based on SVD and IWT. IEEE Access 2020, 8, 43391–43406. [Google Scholar] [CrossRef]
  21. Kim, B.W.; Kim, H.C.; Jung, S.Y. Display field communication: Fundamental design and performance analysis. J. Light. Technol. 2015, 33, 5269–5277. [Google Scholar] [CrossRef]
  22. Jung, S.Y.; Kim, H.C.; Kim, B.W. Implementation of two-dimensional display field communications for enhancing the achievable data rate in smart-contents transmission. Displays 2018, 55, 31–37. [Google Scholar] [CrossRef]
  23. Singh, P.; Jung, S.Y. Data decoding based on iterative spectral image reconstruction for display field communications. ICT Express 2021, 7, 392–397. [Google Scholar] [CrossRef]
  24. Kim, B.W.; Singh, P.; Jung, S.Y. Iterative pilot-based reference frame estimation for improved data rate in two-dimensional display field communications. Appl. Sci. 2023, 13, 9916. [Google Scholar] [CrossRef]
  25. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  26. Bohlin, P.; Tapio, M. Performance evaluation of MIMO communication systems based on superimposed pilots. In Proceedings of the 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing, Montreal, QC, Canada, 17–21 May 2004; Volume 4, pp. 425–428. [Google Scholar]
  27. Wang, A.; Peng, C.; Zhang, O.; Shen, G.; Zeng, B. Inframe: Multiflexing full-frame visible communication channel for humans and devices. In Proceedings of the 13th ACM Workshop on Hot Topics in Networks, Los Angeles, CA, USA, 27–28 October 2014; pp. 1–7. [Google Scholar]
  28. Zhang, K.; Wu, C.; Yang, C.; Zhao, Y.; Huang, K.; Peng, C.; Liu, Y.; Yang, Z. ChromaCode: A fully imperceptible screen-camera communication system. In Proceedings of the 24th Annual International Conference on Mobile Computing and Networking, New Delhi, India, 29 October–2 November 2018; pp. 575–590. [Google Scholar]
  29. Tran, V.; Jayatilaka, G.; Ashok, A.; Misra, A. Deeplight: Robust & unobtrusive real-time screen-camera communication for real-world displays. In Proceedings of the 20th International Conference on Information Processing in Sensor Networks (Co-Located with CPS-IoT Week 2021), Nashville, TN, USA, 18–21 May 2021; pp. 238–253. [Google Scholar]
  30. Cooley, J.W.; Tukey, J.W. An algorithm for the machine calculation of complex Fourier series. Math. Comput. 1965, 19, 297–301. [Google Scholar] [CrossRef]
  31. Oppenheim, A.V.; Schafer, R.W. Discrete-Time Signal Processing, 3rd ed.; Pearson Education India: Bengaluru, India, 2009. [Google Scholar]
  32. Proakis, J.G.; Manolakis, D.K. Digital Signal Processing: Principles, Algorithms, and Applications, 4th ed.; Pearson Education India: Bengaluru, India, 2007. [Google Scholar]
Figure 1. Overview of the proposed iDFC communication.
Figure 1. Overview of the proposed iDFC communication.
Photonics 12 01004 g001
Figure 2. Signal flow diagram of the proposed iDFC system.
Figure 2. Signal flow diagram of the proposed iDFC system.
Photonics 12 01004 g002
Figure 3. Data embedding in iDFC: (a) Original image. (b) Data-embedding region (white bands) in the frequency-domain image.
Figure 3. Data embedding in iDFC: (a) Original image. (b) Data-embedding region (white bands) in the frequency-domain image.
Photonics 12 01004 g003
Figure 4. Effect of perspective distortion in the D2C link: (a) Distorted image. (b) Restored image.
Figure 4. Effect of perspective distortion in the D2C link: (a) Distorted image. (b) Restored image.
Photonics 12 01004 g004
Figure 5. SER before/after perspective correction (M = 2).
Figure 5. SER before/after perspective correction (M = 2).
Photonics 12 01004 g005
Figure 6. Visual imperceptibility of data-embedded images for different image-to-data power ratios α : (a) α = 0.1 . (b) α = 0.3 . (c) α = 0.5 . (d) α = 0.7 .
Figure 6. Visual imperceptibility of data-embedded images for different image-to-data power ratios α : (a) α = 0.1 . (b) α = 0.3 . (c) α = 0.5 . (d) α = 0.7 .
Photonics 12 01004 g006
Figure 7. Symbol error rate performance according to the different values of image power ratios.
Figure 7. Symbol error rate performance according to the different values of image power ratios.
Photonics 12 01004 g007
Figure 8. Achievable data rate performance according to the different values of image power ratios.
Figure 8. Achievable data rate performance according to the different values of image power ratios.
Photonics 12 01004 g008
Figure 9. Pareto frontier illustrating the trade-off between the achievable data rate and visual imperceptibility. Each point corresponds to a specific power allocation factor, α .
Figure 9. Pareto frontier illustrating the trade-off between the achievable data rate and visual imperceptibility. Each point corresponds to a specific power allocation factor, α .
Photonics 12 01004 g009
Figure 10. Comparison of the data rate of the proposed implicit DFC scheme with the conventional DFC [21] and iterative DFC [23] schemes.
Figure 10. Comparison of the data rate of the proposed implicit DFC scheme with the conventional DFC [21] and iterative DFC [23] schemes.
Photonics 12 01004 g010
Figure 11. Symbol error rate performance for different modulation orders.
Figure 11. Symbol error rate performance for different modulation orders.
Photonics 12 01004 g011
Figure 12. Achievable data rate performance versus modulation order (M) versus the SNR.
Figure 12. Achievable data rate performance versus modulation order (M) versus the SNR.
Photonics 12 01004 g012
Table 1. Default simulation parameters.
Table 1. Default simulation parameters.
ParameterValue
Host imageGrayscale ‘Lena’
Image resolution ( P × Q ) 256 × 256 pixels
ChannelPerspective distortion + AWGN
Receiver typeSubtraction with ML estimation
Input data symbolsi.i.d. equiprobable QAM symbols
ModulationBPSK (M = 2), QPSK (M = 4), 8-QAM, 16-QAM
Data sub-bands (L)20
Image-to-data power ratio α [ 0.1 , 0.9 ]
Display frame rate30 fps
Table 2. Comparison of the proposed scheme with other screen–camera communication schemes.
Table 2. Comparison of the proposed scheme with other screen–camera communication schemes.
SchemeEmbedding DomainChannel Estimation/
Synchronization
Data OverheadImperceptibility MethodReported Throughput
InFrame [27]/
InFrame++ [2]
Spatial/TemporalVisible locators and alignment patternsLowLeverages the flicker-fusion property of human vision; requires high refresh rate displays (120 Hz).150–360 kbps
ChromaCode [28]Spatial/TemporalCode borders and featured linesLowOutcome-based adaptive lightness changes to guarantee imperceptibility.120 kbps (goodput); >700 kbps (raw)
MobiScan [14]Spatial/TemporalLocators + smartphone sensorsLowBased on flicker-fusion principles similar to InFrame++. 190 kbps
DeepLight [29]Spatial/TemporalDNN-based screen extractor (no locators needed)NoneModulates only the Blue color channel, making it highly imperceptible even on standard 60 Hz displays.1–4.7 kbps
Proposed iDFCSpectral (DFT)Blind (first-order statistics)NoneData are embedded in high-frequency bands; quality is tunable via the power ratio α .∼152 kbps
Notes: The reported results reflect each paper’s settings (display, camera, distance, coding).
Table 3. Comparison of the complexity of the proposed scheme with the conventional DFC scheme.
Table 3. Comparison of the complexity of the proposed scheme with the conventional DFC scheme.
SchemeConventional DFC [21]Proposed iDFCComplexity
Frame processingRequires processing two frames (a reference frame and a data frame).Requires processing only one frame (the data frame).N/A
DFT operationsPerforms a full DFT on the reference frame AND the data frame.Performs a single full DFT on the received data frame. O ( Q P log P ) per DFT
Channel estimationUses the DFT of the reference frame directly.Performs statistical averaging over L data-bearing rows O ( L Q )
Data retrievalElement-wise division.Element-wise subtraction O ( L Q )
Dominant complexity 2 O ( Q P log P ) O ( Q P log P )
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Singh, P.; Jung, S.-Y. Implicit DFC: Blind Reference Frame Estimation in Screen-to- Camera Communication Using First-Order Statistics. Photonics 2025, 12, 1004. https://doi.org/10.3390/photonics12101004

AMA Style

Singh P, Jung S-Y. Implicit DFC: Blind Reference Frame Estimation in Screen-to- Camera Communication Using First-Order Statistics. Photonics. 2025; 12(10):1004. https://doi.org/10.3390/photonics12101004

Chicago/Turabian Style

Singh, Pankaj, and Sung-Yoon Jung. 2025. "Implicit DFC: Blind Reference Frame Estimation in Screen-to- Camera Communication Using First-Order Statistics" Photonics 12, no. 10: 1004. https://doi.org/10.3390/photonics12101004

APA Style

Singh, P., & Jung, S.-Y. (2025). Implicit DFC: Blind Reference Frame Estimation in Screen-to- Camera Communication Using First-Order Statistics. Photonics, 12(10), 1004. https://doi.org/10.3390/photonics12101004

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop