Next Article in Journal
Enhanced Preparative-Scale Extraction from Graševina Grape Pomace Using Ultrasound-Assisted Extraction and Natural Deep Eutectic Solvents
Next Article in Special Issue
Reversible Data Hiding Algorithm in Encrypted Images Based on Adaptive Median Edge Detection and Matrix-Based Secret Sharing
Previous Article in Journal
A Novel Computational Paradigm for Reconstructing Solid CAD Features from a Segmented Manifold Triangular Mesh
Previous Article in Special Issue
Reversible Data Hiding in Encrypted Images Based on Preprocessing-Free Variable Threshold Secret Sharing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IW-NeRF: Using Implicit Watermarks to Protect the Copyright of Neural Radiation Fields

Chinese University of Engineering of the Chinese People’s Armed Police Force, Xi’an 710086, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2024, 14(14), 6184; https://doi.org/10.3390/app14146184
Submission received: 28 May 2024 / Revised: 10 July 2024 / Accepted: 12 July 2024 / Published: 16 July 2024
(This article belongs to the Special Issue Recent Advances in Multimedia Steganography and Watermarking)

Abstract

The neural radiance field (NeRF) has demonstrated significant advancements in computer vision. However, the training process for NeRF models necessitates extensive computational resources and ample training data. In the event of unauthorized usage or theft of the model, substantial losses can be incurred by the copyright holder. To address this concern, we present a novel algorithm that leverages the implicit neural representation (INR) watermarking technique to safeguard NeRF model copyrights. By encoding the watermark information implicitly, we integrate its parameters into the NeRF model’s network using a unique key. Through this key, the copyright owner can extract the embedded watermarks from the NeRF model for ownership verification. To the best of our knowledge, this is the pioneering implementation of INR watermarking for the protection of NeRF model copyrights. Our experimental results substantiate that our approach not only offers robustness and preserves high-quality 3D reconstructions but also ensures the flawless (100%) extraction of watermark content, thereby effectively securing the copyright of the NeRF model.

1. Introduction

Digital watermarking is a copyright protection technology that embeds copyright identifiers into digital media using algorithms [1,2,3]. In the event of copyright disputes, copyright owners can extract copyright information from digital media using the inverse operation of the embedding algorithm to confirm ownership. The concept of neural radiance fields (NeRFs), introduced by Mildenhall et al. [4], utilizes multilayer perceptrons (MLPs) to implicitly represent 3D scenes and render 2D images by synthesizing new perspectives. Due to the NeRF’s robust representational capabilities, high generalization, and ease of learning, it is poised to become a mainstream technique in digital media representation. However, training NeRF models is challenging, and the theft of such models can lead to significant losses for their owners. Therefore, safeguarding NeRF model copyrights is of paramount importance.
Traditional watermarking algorithms [5,6,7] typically rely on specific mathematical functions to embed watermarks into media by altering them. However, these methods often struggle to strike a balance between imperceptibility, robustness, and watermark capacity. The integration of deep learning in watermarking has revolutionized the field, eliminating the need for manually designing complex mathematical functions and showcasing superior performance. In deep learning watermarking, copyright holders encode watermark information into carrier images using an encoder and extract the watermark information from noisy images through a decoder [8], simulating a black-box scenario. While existing deep learning-based watermarking algorithms excel in robustness, imperceptibility, and embedding capacity, they are primarily tailored for multimedia data, like images [9], audio [10], and videos [11], with limited focus on watermarking implicit data such as NeRF.
Uchida et al. [12] pioneered embedding watermarks in deep neural network models, proposing a model watermarking scheme that embeds watermarks in model parameters using a regularization technique. Model watermarking has since evolved rapidly, proving effective in safeguarding network model copyrights. Leveraging the idea of model watermarking, researchers are exploring solutions to protect NeRF copyrights. For instance, Li et al. [13] introduced the StegaNeRF approach for information hiding, while Luo et al. [14] developed the CopyRNeRF scheme, aiming to safeguard NeRF model copyrights while preserving rendering quality and bit accuracy. However, these techniques may suffer from limited watermark capacity when using bit strings or images as watermark information, necessitating comprehensive representation of copyright holders’ logos in embedded watermark data.
To overcome the low watermark capacity and security vulnerabilities in existing NeRF watermarking techniques, we propose an implicit representation-based watermarking algorithm. This method encodes watermark information as continuous functions using implicit representation neurons, embedding these data into carrier networks via a key. By representing watermark information as neural network models instead of bit strings or images, we significantly enhance the watermark capacity. Our approach involves representing watermark information as continuous functions, embedding these parameters into an untrained network using a key, training the carrier network to encapsulate carrier information while sharing the trained watermark-containing carrier model across the network, and enabling copyright owners to extract watermark information using the key if model theft is suspected.
In summary, this work contributes the following:
  • The first application of implicit representations in NeRF copyright protection to address existing NeRF watermarking challenges.
  • A key-based carrier network construction method for lossless watermark information extraction.
  • Validation of our method across various datasets, ensuring model quality and successful watermark extraction.
  • Testing the robustness of our model to demonstrate that any attempts to remove the watermark would render the NeRF model unusable.

2. Related Work

2.1. Digital Watermarking for 2D Data

Traditionally, digital watermarking algorithms focused on embedding watermarks into two-dimensional data, categorized into spatial domain-based algorithms and transform domain-based algorithms depending on the embedding domains utilized. In spatial domain algorithms, classic examples include least significant bit (LSB) and patchwork algorithms. LSB algorithms [5,6,7] encode watermark information in the least significant bit of image pixels, while patchwork algorithms. utilize statistical characteristics of the carrier image for encoding. Transform domain-based watermarking algorithms [15,16,17] involve transforming the carrier image into a different domain and embedding the watermark by modifying the transformation domain coefficients based on the watermark value. While traditional watermarking algorithms excel in watermark extraction quality, they often lack robustness and struggle to withstand various attacks. The advancement of deep learning has revolutionized digital watermarking, with significant strides made in image watermarking methods leveraging deep learning techniques.
For instance, Kandi et al. [18] implemented watermark embedding using a convolutional neural network autoencoder, while Zhu et al. [19] introduced an end-to-end watermarking framework named HiDDeN. Building upon these innovations, Zhang et al. [20] delved into attack networks within the universal deep learning watermarking paradigm, enhancing the overall robustness of watermarking algorithms. Notably, deep learning-based watermarking algorithms have witnessed remarkable progress, showcasing their potential for robust and efficient watermark embedding and extraction. However, these existing methods do not cater to safeguarding the copyrights of 3D models effectively.

2.2. Digital Watermarking for 3D Data

Within the realm of 3D digital watermarking, traditional methods involve embedding watermarks in various representations of 3D models, such as point clouds [21], voxels [22], or triangular meshes [23], through transformations like translation, rotation, scaling [24], and parameter modifications [25]. Notably, Hou et al. [26] introduced a technique employing layered artifacts of 3D-printed objects for watermarking, while Hamidi et al. [27] enhanced robustness by leveraging wavelet transform to amplify grid saliency. Recognizing the complexities of traditional 3D digital watermarking approaches, researchers have turned towards integrating deep learning techniques in this domain. Wang et al. [28] pioneered a deep learning-based 3D mesh watermarking network, introducing a more versatile framework for 3D mesh watermarking. Building upon this, Yoo et al. [29] successfully embedded messages within 3D meshes and retrieved them from 2D renderings. However, existing methodologies predominantly cater to explicit 3D models, posing challenges when applying them to neural radiance field (NeRF) models lacking specific 3D structures, thereby limiting their efficacy in ensuring copyright protection for NeRF representations.

2.3. Model Watermarking Algorithm

With the burgeoning advancements in deep learning technology, a plethora of sophisticated deep learning models have surfaced, prompting the development of diverse watermarking strategies tailored to safeguard these models. These approaches can be broadly categorized into two groups: watermark embedding based on network weights and watermark insertion based on the classification labels of trigger sets. Uchida et al. [12] pioneered a method for embedding watermarks through network weights, utilizing a parameter regularizer to infuse watermarks into model parameters. Building on Uchida et al’s work, Wang et al. [30] enhanced this technique by introducing an additional neural network that maps network weights to the watermark space, although this is susceptible to ambiguity attacks. In response to these vulnerabilities, Fan et al. [31] devised a novel passport-based ownership verification scheme for deep neural networks (DNNs), showcasing resilience against network modifications and ambiguity attacks. Rouhani et al. [32] proposed DeepSigns, an end-to-end protective framework empowering developers to embed digital watermarks into pertinent deep learning models prior to dissemination. On the other hand, employing trigger set classification labels for watermark incorporation aims to introduce a backdoor into the network architecture, delivering exclusive activation rights to the copyright verifier. Adi et al. [33] introduced a framework integrating author signatures during DNN training to establish a backdoor-based watermarking mechanism, enabling verification of author identity through predefined signature patterns in watermarked DNNs. Additionally, Shafieinejad et al. [34] proposed a backdoor-inspired approach for watermarking DNNs. While these methodologies are tailored for specific network models and not directly applicable to neural radiance field (NeRF) copyright protection, insights from model watermarking can inform the design of watermark algorithms tailored for NeRF security.

2.4. Watermark Algorithm for Neural Radiation Field

The remarkable performance of neural radiance fields (NeRFs) in 3D reconstruction has garnered significant attention from researchers. Li et al. [13] introduced StegaNeRF, a method for embedding secret information in NeRF rendering. Their approach involved an optimization framework, initially training a standard NeRF model and subsequently conducting additional training to achieve the rendering of secret images while preserving the original visual quality. Similarly, Liu et al. [35] embedded secret messages within the implicit representation function of masked data, enabling direct extraction of these messages via a shared key between the sender and receiver. In contrast, Luo et al. [14] proposed an anti-distortion rendering scheme, replacing the original color representation in NeRF with a watermarked color representation to ensure stable extraction of watermark information in 2D rendered images. Additionally, Chen et al. [36] utilized the newly synthesized view of NeRF as a pivotal element for copyright verification, employing parameterized methods to train a watermark extractor for model validation. Notably, the majority of the aforementioned schemes rely on watermark extractors for copyright protection, which may entail certain security risks. Consequently, the development of an implicit representation watermarking algorithm specifically tailored for NeRF models has been pursued to address these concerns.

3. Preliminaries

The neural radiance field (NeRF) model utilizes an MLP network trained with the input of three-dimensional coordinate positions x = ( x , y , z ) and spatial point directions d = ( θ , φ ) , where θ and φ represent horizontal and vertical azimuths, respectively. The network outputs the color of spatial points c = ( r , g , b ) and the density of corresponding positions (voxels) σ . In the specific implementation, the position information of x and d is encoded initially. Subsequently, x is fed into the MLP network to produce σ and a 256-dimensional intermediate feature. This intermediate feature, along with d, is jointly input into a fully connected layer for color prediction, culminating in the generation of a two-dimensional image through volume rendering. The network structure of the NeRF is depicted in Figure 1.
Despite the high adaptability and strong learning capabilities of fully connected MLP networks, they exhibit notable parameter and structural redundancies. To validate the substantial redundancy within the NeRF model network, an 8-layer MLP network with a hidden layer size of 128 was trained. Following this, parameters of the trained NeRF model were pruned based on very small absolute values. The experimental outcomes, illustrated in Figure 2, demonstrate that even with a 50% parameter trim, the NeRF model retains high reconstruction quality. Hence, leveraging the inherent redundancy in MLP networks is considered in the design of our proposed solution.

4. Proposed Method

4.1. Framework

Our study centers on INR watermarking, utilizing the implicit representation of a NeRF as a case study to elucidate our approach, as depicted in Figure 3.
In our methodology, we denote the initial network conveying watermark information as F σ ( · ) , the augmented model network with the watermark as G θ ( · ) , the watermark information dataset as H i , i [ 1 , k ] , and the carrier model’s training dataset as Q i , i [ 1 , k ] . The overall framework of our approach, delineated in Figure 4, is partitioned into three sequential stages. Initially, the content owner trains H i using F σ ( · ) to derive the watermark information network F ^ σ ( · ) and integrates the model parameters of F ^ σ ( · ) into G θ ( · ) via the key K. Subsequently, the content owner trains G θ ( · ) with the carrier dataset to construct a NeRF model G ^ θ ( · ) encompassing watermarks. In the final phase, during a copyright dispute, the content owner leverages the key K to extract the original parameters of F ^ σ ( · ) from G ^ θ ( · ) for watermark information recovery. The embedding and extraction processes of the watermark information described above can be expressed through Equations (1) and (2):
G θ ( · ) = E m b ( F ^ σ ( · ) , K )
F ^ σ ( · ) = E x t ( G ^ θ ( · ) , K )

4.2. Data Representation and Transformation

In the initial phase, we depict the watermark information data using implicit neural representations. The structure of the image noise-resilient (INR) model is a multilayer perceptron (MLP) network comprising n hidden layers of size D, employing the activation function σ ( · ) . The output of the INR model can be formulated as shown in Equation (3):
y = W ( n ) ( g n 1 o o g 1 o g 0 ) ( h 0 ) + b ( n ) , where h i + 1 = g i ( h i ) = σ ( W ( i ) h i + b ( i ) )
Here, h i , i { 0 , 1 , 2 , , n } represents the input of the i-th layer, and y denotes the corresponding output value of h 0 . Additionally, g i signifies the hidden layer, W ( i ) represents the weight matrix of the i-th layer, and b ( i ) denotes the bias parameter of the i-th layer. The INR network architecture utilizes a 10-layer MLP network structure with the ReLU activation function. In the context of a given input H, the output set M corresponding to Y is associated with a loss function as in Equation (4):
L w = min ( h , y ) M F σ ( h ) y 2 2
Here, F σ ( · ) denotes the neural representation of the data. For instance, in the context of the NeRF model and a given set of images from diverse perspectives H i , the set M comprises three-dimensional coordinates, camera pose h = ( x , y , z , θ , φ ) , corresponding RGB values, and opacity y = ( r , g , b , σ ) .

4.3. Watermark Information Embedding Stage

This article is influenced by advancements in neural network watermarking, focusing on embedding watermark representations within carrier networks. The methodology leverages network unfolding techniques for watermark embedding, ensuring the preservation of watermark network parameters within a newly constructed neural network structure. By adopting implicit representations, especially through straightforward network architectures like MLP, embedding watermark parameters using a designated key is achieved without compromising performance. The new network design is centered on the watermark network, with inspiration drawn from models like the NeRF. Three key network deployment strategies are introduced, depicted in Figure 5, to optimize the embedding process efficiently while maintaining the integrity of the watermark information within the carrier network, as follows:
(1) Horizontal expansion. This strategy entails the insertion of new layers subsequent to the watermark network, preserving the original network structure, to construct a larger network, as illustrated in Figure 5a. (2) Vertical expansion. An alternative approach to expanding the network is to maintain the layer count unchanged. Given that both the watermark and extended networks embody the NeRF model, aside from the input and output layers, we augment solely the count of neurons within the hidden layers, as depicted in Figure 5b. (3) Mixed expansion. Mixed expansion encompasses both horizontal and vertical expansion, concurrently increasing the layer count and the neuron count within existing hidden layers. This constitutes a versatile operation, as demonstrated in Figure 5c, where the entirety of the watermark network’s content and parameters are encapsulated within a single network. Indeed, both horizontal and vertical expansion can be encapsulated under the umbrella of mixed expansion, and our objective is to construct the carrier network’s structure using expansion methodologies that are as simplistic as is feasible.
Employing watermark networks and carrier networks to encapsulate watermark information and NeRF models, we naturally leverage mixed dilation to construct carrier networks. Within implicitly represented 3D scenes, upon the training of the watermark network, we can assemble a carrier network through the watermark network. Due to the necessity of embedding the parameters of the trained watermark network within the carrier network, we designate a shared key K, comprising a randomized binary sequence of zeros and ones, as presented in Figure 2. Within our scheme, the key is defined as K = { k 0 , k 1 , k 2 , , k m } , wherein k 0 represents the count of layers of the watermark network neurons within the carrier network, with a binary value of 1 indicating that the layer includes neurons of the watermark network and 0 indicating otherwise. The sequence { k 1 , k 2 , , k m } denotes the location of the watermark network neurons within the carrier network, where each layer corresponds to a binary stream k m of length d k m , specifying the neuron count in that layer. Each bit within this sequence represents a neuron, with 1 indicating it belongs to the watermark information and 0 indicating it belongs to the carrier information. For example, in Figure 5c, the corresponding key K is { 01110 , 00110 , 101101 , 01101 } . This method significantly reduces the length of the key through k 0 .

4.4. Training of Carrier Networks

We utilized Equation (4) to enable the implicit representation of the carrier NeRF model through the carrier network G θ ( · ) , where θ encompasses two distinct parameter categories: those of the watermark network and others representing carrier information. In ensuring the lossless extraction of F ^ σ ( · ) from G ^ θ ( · ) , it becomes imperative to immobilize the parameters associated with the watermark information. Subsequently, training is exclusively directed towards the parameters embodying carrier information, thereby optimizing F ^ σ ( · ) with precision. In pursuit of this objective, we introduced a binary mask M to facilitate selective optimization of G θ ( · ) in Equation (5):
M [ p ] = 1 , θ [ p ] φ 0 , e l s e
Among them, φ represents the parameter representing carrier information in θ , and θ [ p ] represents the p-th parameter in θ . Similar to Equation (6), we use the image set Q i to train the carrier network and define the loss of the carrier network as L c . The formula for L c is as follows:
L c = min ( q , y ) M G θ ( q ) y 2 2
Let λ be the learning rate and ⊙ be the product of elements, and update G θ ( · ) using the gradient descent method, as shown in Equation (7).
θ = θ λ M θ L c
Through training, we obtained the NeRF model G ^ θ ( · ) containing watermarks.

4.5. Watermark Information Extraction

After releasing model G ^ θ ( · ) online, regular users can engage with and explore 3D environments. Maintaining the watermark information parameters unaltered during the training phase of the carrier network G θ ( · ) ensures that the party responsible for copyright protection can restore the watermark information network F ^ σ ( · ) without any loss postacquisition of the watermark network parameter details utilizing the key K. Leveraging key K, we not only ascertain the layers within F ^ σ ( · ) housing watermark information but also identify the neurons containing watermark information along with their pertinent parameter specifics. In essence, through the parameters P F ^ σ ( · ) and structure S F ^ σ ( · ) in key K, the watermark network can be regenerated. The overall process of our method is shown in Algorithm 1.
Algorithm 1 Training process of IW-NeRF.
  1:
Data: Watermark information dataset H i , i [ 1 , k ] , Carrier model training dataset Q i , i [ 1 , k ] , Random key K, learning rate η
  2:
Output: Watermark information network model F ^ σ ( · )
Initial network containing watermark information G θ ( · )
Watermarked NeRF model G ^ θ ( · ) .
Optimizing the parameter θ of model G θ ( · ) through Q i
Compute mask M for θ as in Equation (5)
  3:
for each training iteration t do
    Compute Watermark information loss as Equation (4)
    Compute Carrier network information loss as Equation (6)
    Update θ as Equation (7)
  4:
end for

5. Experiments

5.1. Experimental Settings

Dataset. We assessed our algorithm using the NeRF Semantic and LLFF datasets sourced from the NeRF dataset. The LLFF dataset encompasses diverse scenes such as flowers, ferns, fortresses, and rooms, while the NeRF Semantic dataset features 360-degree scenes like LEGO structures, drums, and chairs. For evaluating the efficacy of our approach within the NeRF Semantic dataset, akin to a NeRF [1], our training regimen involves feeding in 100 views per scene. To gauge the visual fidelity of our methodology, we handpicked 20 images per scene from the test dataset. Furthermore, we conducted renders of 200 views for each scene to validate the watermark extraction precision across varying camera angles. Our comprehensive experimental methodology entails presenting all outcomes as averaged results for clarity and consistency.
Implicit neural representation setting. Our approach was implemented utilizing PyTorch, and both the watermark information and carrier network were structured employing an MLP network design. We employed 12 hidden layers of size 128 to encode the watermark information and 22 hidden layers of size 256 for the carrier network. The hyperparameters used include a learning rate of 5 × 10 4 , a batch size of 512, utilization of the Adam optimizer with default values of β 1 = 0.9 , β 2 = 0.999 , and regularization parameter λ = 1 × 10 8 . The experiments were conducted on an NVIDIA A100 GPU, with training of both the watermark network and carrier network executed using a stochastic gradient descent algorithm over 20,000 epochs.
Baselines. To the best of our knowledge, there exists limited research on watermarking specifically tailored for NeRF applications. Thus, we conducted a comparative analysis with established watermarking techniques adapted for NeRF to ensure a comprehensive evaluation: (1) LSB [6] + NeRF: here, we employed the LSB algorithm to embed watermark data into the dataset images before NeRF model training; (2) DeepStega [37] + NeRF: preceding NeRF model training, we leveraged the two-dimensional watermarking approach of DeepStega for image processing; (3) HiDDeN [19] + NeRF: prior to NeRF model training, the image underwent processing utilizing the HiDDeN scheme; (4) StegaNeRF [13]; and (5) CopyNeRF [14].
Distortion evaluation metric. In order to evaluate the distortion between the carrier data and the carrier data containing watermarks, as well as the distortion between the watermark data and the extracted watermark data, we visualized the implicitly represented data and used several evaluation metrics. For 2D images rendered from 3D models, we evaluate them using peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and perceived loss (LPIPS).
(1)
PSNR (peak signal-to-noise ratio).
PSNR is an indicator used to evaluate image quality based on the concept of root mean square error (MSE), which represents the peak signal-to-noise ratio of image signals. The higher the PSNR value, the better the quality of the evaluated image and the smaller the error. The calculation methods for MSE and PSNR are shown in Equations (8) and (9), respectively.
MSE = 1 W × H i = 1 W j = 1 H ( X i , j Y i , j ) 2
PSNR = 20 × lg ( s c ) 10 × lg ( MSE )
In the formulas, X and Y represent two images with a size of W × H, respectively, and sc represents the scaling factor, usually taken as 2.
(2)
SSIM (structural similarity).
SSIM is a metric that evaluates image quality by measuring the similarity between two images, with higher values indicating higher similarity. The calculation method of SSIM is shown in Equation (10).
SSIM ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 )
In the above equation, μ x and σ x are the mean and variance of image X, μ y and σ y are the mean and variance of image Y, and σ x y is the covariance of X and Y. c 1 = ( k 1 L ) 2 , c 2 = ( k 2 L ) 2 is a constant, k 1 = 0.01 , k 2 = 0.03 , L is the dynamic range of pixel values, and if the data are uint8-type, L takes a value of 255.
(3)
LPIPS (learned perceptual image patch similarity).
LPIPS is a method of measuring image similarity, which does not use mathematical formulas to implement but evaluates the perceptual differences between two images through deep learning models. The lower the value of LPIPS, the more similar the two images are; and the higher the value, the greater the difference. We use a pretrained LPIPS model Alex for evaluation.

5.2. Reconstruction and Watermark Extraction Quality

We conducted a qualitative comparison of the reconstruction quality against all baselines, with the results depicted in Figure 6. The visual analysis from Figure 6 indicates that all methods exhibit commendable reconstruction quality. While other schemes may outperform ours in reconstruction quality due to their limited capacity stemming from text or image information embedding, our approach stands out in terms of embedded capacity. Despite the reduction in reconstruction quality caused by the embedding of watermark network parameters, likely attributed to fixed parameters impacting the carrier network’s fitting effect, the overall NeRF model information remains retrievable visually. Notably, our designed key facilitates lossless recovery of watermark information from a carrier network containing watermarks, offering a significantly larger embedded capacity compared to alternative schemes. For a quantitative assessment, we further analyzed the rendering quality and watermark information extraction effects across various schemes, as detailed in Table 1 and Table 2.
By examining the outcomes outlined in Table 1 and Table 2, a notable observation is the ineffectiveness of the two-dimensional watermarking techniques applied to the NeRF model in retrieving watermark information. The alteration of information embedded via the two-dimensional watermarking method due to NeRF’s view synthesis led to the method’s failure. Conversely, the utilization of three-dimensional watermarking approaches facilitated accurate extraction of watermark information. Notably, while other algorithms incorporate bit strings into the model, our methodology visualizes the 3D reconstruction and watermark extraction effects. Our experimentation involved utilizing diverse 360-degree scenes such as ‘Lego’, ‘drums’, and ‘chairs’ from the NeRF Semantic dataset for training the watermark network and carrier network. The resulting experimental findings, as illustrated in Figure 7, showcase various images representing original watermark information, carrier information, samples from both the watermark and carrier networks, and extracted watermark network samples. Despite our method slightly affecting the NeRF model’s training performance, it demonstrates the ability to extract embedded watermark information from the carrier network without loss. Noteworthy is the incorporation of a neural network capable of implicitly representing a range of copyright materials (e.g., images, sounds, videos) in real-world scenarios.

5.3. Algorithm Capacity

Our approach involves embedding the parameters of a watermark network within a carrier model network, allowing for the assessment of algorithm capacity through scalability analysis. The expansion rate, as defined in Equation (11), serves as a metric to quantify this scalability.
e = N c a r r i e r N w a t e r m a r k 1
In this context, N w a t e r m a r k and N carrier represent the parameter counts within the original watermark network and carrier network, respectively. Consequently, with e 0 , the expansion rate is intricately linked to the sizes of the watermark network and carrier network. A lower expansion rate correlates with higher capacity. To assess the influence of watermark capacity on the outcomes of our proposed scheme, we executed the subsequent pair of experiments.
Watermark network layer modification. In our approach, the utilization of INR to represent watermark information enables a smaller watermark neural network to achieve superior carrier data reconstruction during the fitting process. By adjusting the size of the watermark network, we can influence the performance of the carrier network fitting. Specifically, we maintained a fixed size of 22 hidden layers for the carrier network and employed hidden layer networks with 16, 14, 12, 10, and 8 layers for training the watermark information. The hidden layer width for the carrier network was set at 256, while for the watermark network, it was set at 128. Our experiments were carried out using the LLFF dataset. Figure 8 illustrates the performance of carrier models trained with varying watermark networks, showcasing rendered images by both the carrier and watermark networks in the respective columns. Furthermore, Figure 9 displays the SSIM values between the rendered and real images of the carrier network across different watermark network levels. Table 3 depicts the impact of the expansion rate on the PSNR values between the rendered and real images of the carrier network. Notably, our experimental results highlight that with a consistent size for the carrier network, reducing the number of layers in the watermark network leads to an increase in the expansion rate and the complexity of the carrier network, consequently enhancing the quality of the carrier model.
Carrier network layer modification. To further optimize 3D reconstruction outcomes, adjustments to both the watermark network size and the carrier network dimensions were explored. The watermark network was kept at a fixed size of 10 layers, while carrier networks ranging from 18 to 26 layers were utilized in training the carrier model. The hidden layer width was set to 128 for the watermark network and 256 for the carrier network. PSNR values illustrating the comparison between images rendered by the carrier model trained on different layers of the carrier network and the original images is depicted in Figure 10. Additionally, Figure 11 showcases SSIM values representing the comparison between the rendered image and the actual image of the carrier network across varying layers. The impact of the expansion rate on the PSNR values between the rendered and real images of the carrier network is detailed in Table 4. The experimental findings reveal a positive correlation between the depth of the carrier network and the expansion rate, leading to an enhancement in the reconstruction quality of the carrier model. However, it is worth noting that an increase in the number of layers of the carrier network introduces higher computational complexity and training challenges, potentially reducing training efficiency. For future research directions, the adoption of multiresolution hash encoding [38] is under consideration to boost training speed.

5.4. The Robustness of the Model

In practical scenarios, the vulnerability of our model to malicious attacks necessitates an assessment of the robustness of our watermarking scheme against various model alterations. To evaluate the resilience of our algorithm, we employed two common modifications. Model pruning is a vital technique, especially for MLP networks known for their robust fitting capabilities but often burdened with numerous parameters due to deep network layers and high neuron counts. The abundance of parameters not only escalates computational demands in deep learning but also presents an opportunity for optimization through pruning. The primary objective of network pruning is to reduce redundant parameters while preserving the original network’s performance. Our experiments encompassed L 1 unstructured pruning, L n structured pruning, random structured pruning, and random unstructured pruning methods. During the pruning process, the parameter with the minimum absolute pruning rate p% (ranging from 10% to 90%) was set to 0 for the NeRF model embedded with watermarks. Subsequently, we compared the extracted watermark information postpruning with the prepruning data to assess its impact on our watermark framework. Ideally, a model thief aims to render the extracted watermark information inaccurate postpruning while retaining the model’s performance. Figure 12 and Figure 13 illustrate the influence of pruning rates on watermark extraction performance under various pruning methodologies. Notably, even with L 1 unstructured pruning removing 90% of parameters, our embedding method demonstrated high accuracy in watermark extraction (with an SSIM value of 0.8829 between the extracted and original watermark information). Conversely, L n structured pruning, random structured pruning, and random unstructured pruning methods exhibited a significant reduction in extracted watermark quality beyond a 10% pruning rate. However, while these pruning methods may compromise watermark quality, they also detrimentally impact the model’s performance, rendering the stolen model ineffective. Thus, our watermark algorithm exhibits robustness against pruning modifications, ensuring that any attempts to maintain the model’s quality post-theft come at the cost of significantly reduced model performance.
Model fine-tuning is a crucial aspect in the context of large-scale NeRF models. Training such models from scratch demands a substantial training dataset, and the lack of adequate data can impact model performance. Hence, in practical scenarios, fine-tuning existing models becomes a more feasible approach when data scarcity is a concern. Typically, fine-tuning is preferred when there is minimal variance in experimental performance between the original dataset and the dataset trained using a pretrained model. Consequently, for potential plagiarists, employing fine-tuning methods to train a new model based on the stolen model with limited training data becomes a viable strategy. This new model not only mirrors the performance of the original model but also diverges in experimental outcomes from the original model. In our experimental setup, each dataset underwent a partitioning of the test data into two segments: the first half facilitated the fine-tuning of the trained NeRF model, while the second half served to evaluate the performance of the newly derived model. Subsequently, we utilized the fine-tuned model to assess the resilience of our watermark scheme against modifications induced by the fine-tuning process.
Table 5 presents the PSNR and SSIM metrics comparing the original watermark information with the extracted watermark information post-model fine-tuning. The experimental findings indicate that fine-tuning has a minimal impact on watermark accuracy, regardless of whether synthetic or real datasets are utilized. This resilience to fine-tuning-induced modifications can be attributed to the presence of numerous redundant neurons within the MLP network, enhancing the robustness of our watermark algorithm against such adjustments.

5.5. Security Assessment

Contrary to discrete representations, implicit neural representations introduce uncertainty, leading to varied weight distributions for each dataset across rendering and watermark extraction processes. To assess the algorithm’s security, we conducted a comprehensive statistical analysis and visual examination of the weight distributions within the watermark network model and the carrier network post-watermark embedding. The watermark network comprised 10 layers, while the carrier network featured 22 layers. By comparing the corresponding layers’ weights in both networks, we observed distinct differences. Visualization techniques were employed to juxtapose the weights of networks with equivalent layer numbers in the watermark and carrier networks, depicted in Figure 14. The disparity in weight distributions between the watermark and carrier networks was evident, signifying the successful incorporation of watermark network parameters into the carrier network.
Within our watermarking framework, the pivotal role of the key in extracting watermark information underscores the need for a comprehensive evaluation of key sensitivity. Analogous to cryptographic protocols, the correct key serves as the sole means through which copyright stakeholders can access the embedded watermark data. In contrast to traditional cryptography, the performance of neural networks in extracting watermarks is subject to various influencing factors, including network topology, training datasets, parameter configurations, and optimization strategies, which may introduce performance variability. From an experimental standpoint, the fundamental criterion lies in the successful extraction of the watermark using the correct key, as even minor deviations in the key could result in extraction failures. Figure 15 showcases the average peak signal-to-noise ratio (PSNR) values between the original watermark and the extracted watermark across diverse datasets, highlighting the impact of varying error bits within the 2180-sized key. Additionally, Figure 16 visually demonstrates the consequences of erroneous key bits on the output quality of the extracted watermark network, showcasing a rapid decline in PSNR values as the number of incorrect key bits escalates, leading to perceptible image blurring. This observation underscores the sensitivity of the watermark extraction process to even minor alterations in the key. For individuals lacking direct access to the key, resorting to guesswork may introduce errors, with approximately 50% of incorrect key bits resulting in failed watermark extraction attempts. Future endeavors will focus on refining key generation methodologies to bolster the security of the watermarking scheme.

6. Limitations

The utilization of a network model, as opposed to a conventional bit string or image, in our watermarking scheme introduces a trade-off in the quality of the NeRF model. Our embedding and extraction processes operate akin to a white-box model, necessitating access to the NeRF model’s parameters for successful watermark extraction. This requirement may constrain the practical applicability of our algorithm, particularly in scenarios where parameter accessibility is restricted. Moving forward, our research endeavors will focus on enhancing the watermark embedding methodology to refine the overall efficiency and robustness of our watermarking scheme.

7. Conclusions

In light of the rapid advancements and diverse applications of NeRF technology, the preservation of copyright for NeRF models has emerged as a critical imperative. This study introduces a novel NeRF model watermarking scheme known as implicit neural representation watermarking for NeRF (IW-NeRF). Within this framework, the watermark information is intricately embedded into the NeRF model through implicit neural representations, ensuring a seamless and robust integration of copyright protection mechanisms within the model architecture. Subsequently, the extraction of the watermark information is facilitated through a designated key, effectively fortifying the copyright protection framework for NeRF models.
Advantages. The experimental findings demonstrate the distinct advantages of our proposed watermark scheme: (1) Security: Within our scheme, the watermark information is implicitly represented and then randomly embedded into the NeRF model using a key derived from network parameters. The extensive key space, stemming from the multitude of network parameters, ensures the security of our approach. (2) Capacity: By characterizing the watermark information implicitly through neural methods and incorporating it into a continuous function, the memory required for signal parameterization becomes independent of spatial resolution, solely influenced by the signal’s underlying complexity. This strategy enables the utilization of continuous functions for achieving high-capacity watermarks. (3) Robustness: Neural networks exhibit inherent redundancy in structure and neuron count. By embedding the watermark information within the NeRF model network, our scheme maintains resilience against minor alterations to the model; any significant attacks on the model would consequently impact its performance rather than compromise the embedded watermark. (4) Universality: Implicit neural representations possess the capability to depict diverse data forms. While our experimental focus entailed the utilization of the NeRF model as watermark information, the watermark network framework proposed in our study extends to various media types such as images and videos. Noteworthy is the pioneering introduction of implicit neural representation within the realm of digital watermarking in this research.

Author Contributions

Writing—original draft, L.C.; Formal analysis, C.S.; Writing—review & editing, J.L.; Software, W.S.; Methodology, W.D.; Resources, F.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the National Natural Science Foundation of China, with fund number 62272478.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the author 18792537291@139.com. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare that they have no conflicts of interest to report regarding the present study.

References

  1. Li, D.; Yang, Z.; Jin, X. Zero watermarking scheme for 3D triangle mesh model based on global and local geometric features. Multimed. Tools Appl. 2023, 82, 43635–43648. [Google Scholar] [CrossRef]
  2. Wu, H.; Liu, G.; Yao, Y.; Zhang, X. Watermarking Neural Networks with Watermarked Images. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 2591–2601. [Google Scholar] [CrossRef]
  3. Feng, L.; Zhang, X. Watermarking Neural Network with Compensation Mechanism. In Knowledge Science, Engineering and Management; Lecture Notes in Computer Science; Li, G., Shen, H.T., Yuan, Y., Wang, X., Liu, H., Zhao, X., Eds.; Springer: Cham, Swtzerland, 2020; pp. 363–375. [Google Scholar] [CrossRef]
  4. Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. NeRF: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 2022, 65, 99–106. [Google Scholar] [CrossRef]
  5. Kuang, X.; Ling, W.A.; Ke, L.S.; Lei, G.; Ping, P.J.; Yue, L.Z.; Ping, L.F. Watermark embedding and extraction based on LSB and four-step phase shift method. In Proceedings of the 2019 7th International Conference on Information Technology: IoT and Smart City, Shanghai, China, 20–23 December 2019; pp. 243–247. [Google Scholar]
  6. Muyco, S.D.; Hernandez, A.A. Least significant bit hash algorithm for digital image watermarking authentication. In Proceedings of the 2019 5th International Conference on Computing and Artificial Intelligence, Bali, Indonesia, 17–20 April 2019; pp. 150–154. [Google Scholar]
  7. Van Schyndel, R.G.; Tirkel, A.Z.; Osborne, C.F. A digital watermark. In Proceedings of the 1st International Conference on Image Processing, Austin, TX, USA, 13–16 November 1994; Volume 2, pp. 86–90. [Google Scholar]
  8. Chen, L.; Liu, J.; Dong, W.; Sun, W. Image Hiding Scheme Based on Dense Residual Networks. Sci. Technol. Eng. 2024, 24, 03719-08. [Google Scholar]
  9. Singh, H.K.; Singh, A.K. Digital image watermarking using deep learning. Multimed. Tools Appl. 2024, 83, 2979–2994. [Google Scholar] [CrossRef]
  10. Charfeddine, M.; Mezghani, E.; Masmoudi, S.; Amar, C.B.; Alhumyani, H. Audio watermarking for security and non-security applications. IEEE Access 2022, 10, 12654–12677. [Google Scholar] [CrossRef]
  11. Luo, X.; Li, Y.; Chang, H.; Liu, C.; Milanfar, P.; Yang, F. DVMark: A deep multiscale framework for video watermarking. arXiv 2023, arXiv:2104.12734. [Google Scholar] [CrossRef] [PubMed]
  12. Uchida, Y.; Nagai, Y.; Sakazawa, S.; Satoh, S. Embedding Watermarks into Deep Neural Networks. In Proceedings of the 2017 ACM on International Conference on Multimedia Retrieval, Bucharest, Romania, 6–9 June 2017; pp. 269–277. [Google Scholar] [CrossRef]
  13. Li, C.; Feng, B.Y.; Fan, Z.; Pan, P.; Wang, Z. StegaNeRF: Embedding Invisible Information within Neural Radiance Fields. arXiv 2022, arXiv:2212.01602. [Google Scholar]
  14. Luo, Z.; Guo, Q.; Cheung, K.C.; See, S.; Wan, R. CopyRNeRF: Protecting the CopyRight of Neural Radiance Fields. arXiv 2023, arXiv:2307.11526. [Google Scholar]
  15. Chen, W.; Zhu, C.; Ren, N.; Seppänen, T.; Keskinarkaus, A. Screen-cam robust and blind watermarking for tile satellite images. IEEE Access 2020, 8, 125274–125294. [Google Scholar] [CrossRef]
  16. Singh, O.P.; Singh, A.K. Image fusion-based watermarking in IWT-SVD domain. In Advanced Machine Intelligence and Signal Processing; Springer: Berlin/Heidelberg, Germany, 2022; pp. 163–175. [Google Scholar]
  17. Thomas, R.; Sucharitha, M. Contourlet and Gould transforms for hybrid image watermarking in RGB color images. Intell. Autom. Soft Comput. 2022, 33, 879–889. [Google Scholar] [CrossRef]
  18. Kandi, H.; Mishra, D.; Gorthi, S.R.S. Exploring the learning capabilities of convolutional neural networks for robust image watermarking. Comput. Secur. 2017, 65, 247–268. [Google Scholar] [CrossRef]
  19. Zhu, J.; Kaplan, R.; Johnson, J.; Fei-Fei, L. Hidden: Hiding data with deep networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 657–672. [Google Scholar]
  20. Zhang, C.; Karjauv, A.; Benz, P.; Kweon, I.S. Towards robust deep hiding under non-differentiable distortions for practical blind watermarking. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 5158–5166. [Google Scholar]
  21. Zhang, G.; Liu, J. Research on Watermark Algorithm for 3D Color Point Cloud Models. Comput. Technol. Dev. 2023, 33, 62–68. [Google Scholar]
  22. Li, B.; Wu, J.; Zhang, B. Improved Image Digital Watermark Algorithm for Zero Parallax Pixel Reorganization. Comput. Simul. 2023, 40, 244–248. [Google Scholar]
  23. Cui, J.; Zhang, G. Anti simplification blind watermarking algorithm based on vertex norm 3D mesh model. Comput. Eng. Des. 2023, 44, 692–698. [Google Scholar] [CrossRef]
  24. Xiong, X.; Wei, L.; Xie, G. A robust color image watermarking algorithm based on 3D-DCT and SVD. Comput. Eng. Sci. Gongcheng Kexue 2015, 37, 8. [Google Scholar]
  25. Pham, G.N.; Lee, S.H.; Kwon, O.H.; Kwon, K.R. A 3D Printing Model Watermarking Algorithm Based on 3D Slicing and Feature Points. Electronics 2018, 7, 23. [Google Scholar] [CrossRef]
  26. Hou, J.U.; Kim, D.G.; Lee, H.K. Blind 3D mesh watermarking for 3D printed model by analyzing layering artifact. IEEE Trans. Inf. Forensics Secur. 2017, 12, 2712–2725. [Google Scholar] [CrossRef]
  27. Hamidi, M.; Chetouani, A.; El Haziti, M.; El Hassouni, M.; Cherifi, H. Blind robust 3D mesh watermarking based on mesh saliency and wavelet transform for copyright protection. Information 2019, 10, 67. [Google Scholar] [CrossRef]
  28. Wang, F.; Zhou, H.; Fang, H.; Zhang, W.; Yu, N. Deep 3D mesh watermarking with self-adaptive robustness. Cybersecurity 2022, 5, 24. [Google Scholar] [CrossRef]
  29. Yoo, I.; Chang, H.; Luo, X.; Stava, O.; Liu, C.; Milanfar, P.; Yang, F. Deep 3D-to-2D Watermarking: Embedding Messages in 3D Meshes and Extracting Them from 2D Renderings. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 10021–10030. [Google Scholar] [CrossRef]
  30. Wang, J.; Wu, H.; Zhang, X.; Yao, Y. Watermarking in Deep Neural Networks via Error Back-propagation. Electron. Imaging 2020, 32, 22-1–22-9. [Google Scholar] [CrossRef]
  31. Fan, L.; Ng, K.W.; Chan, C.S. Rethinking Deep Neural Network Ownership Verification: Embedding Passports to Defeat Ambiguity Attacks. 2019. Available online: https://proceedings.neurips.cc/paper_files/paper/2019/file/75455e062929d32a333868084286bb68-Paper.pdf (accessed on 16 September 2019).
  32. Rouhani, B.D.; Chen, H.; Koushanfar, F. DeepSigns: A Generic Watermarking Framework for IP Protection of Deep Learning Models. arXiv 2018, arXiv:1804.00750. [Google Scholar]
  33. Adi, Y.; Baum, C.; Cisse, M.; Keshet, J.; Pinkas, B. Turning Your Weakness Into a Strength: Watermarking Deep Neural Networks by Backdooring. 2018. Available online: https://www.usenix.org/system/files/conference/usenixsecurity18/sec18-adi.pdf (accessed on 13 February 2018).
  34. Shafieinejad, M.; Wang, J.; Lukas, N.; Li, X.; Kerschbaum, F. On the Robustness of the Backdoor-based Watermarking in Deep Neural Networks. arXiv 2019, arXiv:1906.07745. [Google Scholar]
  35. Liu, J.; Luo, P.; Ke, Y. Hiding Functions within Functions: Steganography by Implicit Neural Representations. arXiv 2023, arXiv:2312.04743. [Google Scholar]
  36. Chen, L.; Liu, J.; Ke, Y.; Sun, W.; Dong, W.; Pan, X. MarkNerf: Watermarking for Neural Radiance Field. arXiv 2023, arXiv:2309.11747. [Google Scholar] [CrossRef]
  37. Baluja, S. Hiding Images in Plain Sight: Deep Steganography. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  38. Müller, T.; Evans, A.; Schied, C.; Keller, A. Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. 2022, 41, 1–15. [Google Scholar] [CrossRef]
Figure 1. The network structure of the NeRF model.
Figure 1. The network structure of the NeRF model.
Applsci 14 06184 g001
Figure 2. The reconstruction quality of NeRF model under different pruning rates.
Figure 2. The reconstruction quality of NeRF model under different pruning rates.
Applsci 14 06184 g002
Figure 3. We present an implicit representation watermarking algorithm tailored for NeRFs. This approach encapsulates the watermark information implicitly within the NeRF model through a specified key, facilitating subsequent watermark extraction using the same key.
Figure 3. We present an implicit representation watermarking algorithm tailored for NeRFs. This approach encapsulates the watermark information implicitly within the NeRF model through a specified key, facilitating subsequent watermark extraction using the same key.
Applsci 14 06184 g003
Figure 4. The overall framework of our algorithm IW-NeRF.
Figure 4. The overall framework of our algorithm IW-NeRF.
Applsci 14 06184 g004
Figure 5. Three expansion strategies for constructing carrier networks.
Figure 5. Three expansion strategies for constructing carrier networks.
Applsci 14 06184 g005
Figure 6. Comparison of reconstruction quality of NeRF models with different baselines.
Figure 6. Comparison of reconstruction quality of NeRF models with different baselines.
Applsci 14 06184 g006
Figure 7. The NeRF reconstruction and watermark extraction effects of our scheme.
Figure 7. The NeRF reconstruction and watermark extraction effects of our scheme.
Applsci 14 06184 g007
Figure 8. The reconstruction and watermark extraction performance of NeRF models under different watermark network layers with a fixed number of carrier network layers.
Figure 8. The reconstruction and watermark extraction performance of NeRF models under different watermark network layers with a fixed number of carrier network layers.
Applsci 14 06184 g008
Figure 9. The SSIM value between the images rendered by the carrier NeRF model and the images in the original dataset of NeRF under different watermark network layers.
Figure 9. The SSIM value between the images rendered by the carrier NeRF model and the images in the original dataset of NeRF under different watermark network layers.
Applsci 14 06184 g009
Figure 10. The 3D reconstruction quality of the NeRF model under different carrier network layers after a fixed number of watermark network layers.
Figure 10. The 3D reconstruction quality of the NeRF model under different carrier network layers after a fixed number of watermark network layers.
Applsci 14 06184 g010
Figure 11. SSIM values between the images rendered by the carrier NeRF model and the carrier dataset images under different carrier network layers.
Figure 11. SSIM values between the images rendered by the carrier NeRF model and the carrier dataset images under different carrier network layers.
Applsci 14 06184 g011
Figure 12. The effect of different pruning methods on watermark extraction.
Figure 12. The effect of different pruning methods on watermark extraction.
Applsci 14 06184 g012
Figure 13. SSIM values between the original watermark image and the extracted watermark image under different pruning methods.
Figure 13. SSIM values between the original watermark image and the extracted watermark image under different pruning methods.
Applsci 14 06184 g013
Figure 14. Comparison of weight visualization between watermark network and carrier network.
Figure 14. Comparison of weight visualization between watermark network and carrier network.
Applsci 14 06184 g014
Figure 15. The average PSNR between the original watermark and the extracted watermark under different incorrect bit keys.
Figure 15. The average PSNR between the original watermark and the extracted watermark under different incorrect bit keys.
Applsci 14 06184 g015
Figure 16. Visual effects of extracting watermarks under different key error bits.
Figure 16. Visual effects of extracting watermarks under different key error bits.
Applsci 14 06184 g016
Table 1. Quantitative analysis of reconstruction quality and watermark extraction effect of NeRF models under different baselines (Synthetic-chair).
Table 1. Quantitative analysis of reconstruction quality and watermark extraction effect of NeRF models under different baselines (Synthetic-chair).
MethodNeRF RenderingWatermark Extraction
PSNR ↑SSIM ↑LPIPS ↓Acc (%) ↑SSIM ↑
Standard NeRF33.230.91430.1113N/AN/A
LSB + NeRF27.450.84460.1410N/AN/A
DeepStega + NeRF26.410.84570.1429N/AN/A
HiDDeN + NeRF27.880.89640.1418N/AN/A
StegaNeRF30.310.98470.02761000.9643
CopyRNeRF30.540.96890.0327100N/A
IW-NeRF (ours)21.320.63640.28521001
Table 2. Quantitative analysis of reconstruction quality and watermark extraction effect of NeRF models under different baselines (LLFF-trex).
Table 2. Quantitative analysis of reconstruction quality and watermark extraction effect of NeRF models under different baselines (LLFF-trex).
MethodNeRF RenderingWatermark Extraction
PSNR ↑SSIM ↑LPIPS ↓Acc (%) ↑SSIM ↑
Standard NeRF27.760.85460.1453N/AN/A
LSB + NeRF27.590.84350.1399N/AN/A
DeepStega + NeRF26.980.83560.1269N/AN/A
HiDDeN + NeRF27.640.88650.1512N/AN/A
StegaNeRF28.210.84530.14231000.9698
CopyRNeRF30.670.96830.0457100N/A
IW-NeRF (ours)22.540.65390.27981001
Table 3. After modifying the number of layers in the watermark network, we download the PSNR value between the rendered image and the real image of the network with different expansion rates.
Table 3. After modifying the number of layers in the watermark network, we download the PSNR value between the rendered image and the real image of the network with different expansion rates.
Expansion Rate1.752.142.663.404.50
PSNR19.8821.6722.9923.4125.86
Table 4. After modifying the number of layers in the carrier network, we download the PSNR value between the rendered image and the real image of the body network with different expansion rates.
Table 4. After modifying the number of layers in the carrier network, we download the PSNR value between the rendered image and the real image of the body network with different expansion rates.
Expansion Rate2.603.003.403.804.20
PSNR20.9421.5823.2124.3425.66
Table 5. The watermark extraction effect on different datasets after model fine-tuning.
Table 5. The watermark extraction effect on different datasets after model fine-tuning.
DatasetHot DogLegoFernFlower
PSNR (dB)46.3445.2844.5746.13
SSIM0.98430.97890.96480.9881
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, L.; Song, C.; Liu, J.; Sun, W.; Dong, W.; Di, F. IW-NeRF: Using Implicit Watermarks to Protect the Copyright of Neural Radiation Fields. Appl. Sci. 2024, 14, 6184. https://doi.org/10.3390/app14146184

AMA Style

Chen L, Song C, Liu J, Sun W, Dong W, Di F. IW-NeRF: Using Implicit Watermarks to Protect the Copyright of Neural Radiation Fields. Applied Sciences. 2024; 14(14):6184. https://doi.org/10.3390/app14146184

Chicago/Turabian Style

Chen, Lifeng, Chaoyue Song, Jia Liu, Wenquan Sun, Weina Dong, and Fuqiang Di. 2024. "IW-NeRF: Using Implicit Watermarks to Protect the Copyright of Neural Radiation Fields" Applied Sciences 14, no. 14: 6184. https://doi.org/10.3390/app14146184

APA Style

Chen, L., Song, C., Liu, J., Sun, W., Dong, W., & Di, F. (2024). IW-NeRF: Using Implicit Watermarks to Protect the Copyright of Neural Radiation Fields. Applied Sciences, 14(14), 6184. https://doi.org/10.3390/app14146184

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop