Next Article in Journal
Reverse Poisson Counting Process with Random Observations
Previous Article in Journal
Advancing Link Prediction with a Hybrid Graph Neural Network Approach
Previous Article in Special Issue
Tail Conditional Expectation and Tail Variance for Extended Generalized Skew-Elliptical Distributions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Can the Collateral Value of a Data Asset Be Increased by Insurance?

1
Key Laboratory of Advanced Theory and Application in Statistics and Data Science-MOE, School of Statistics, East China Normal University, Shanghai 200062, China
2
Shanghai Data Exchange, Shanghai 200135, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(22), 3596; https://doi.org/10.3390/math13223596
Submission received: 30 September 2025 / Revised: 1 November 2025 / Accepted: 3 November 2025 / Published: 10 November 2025

Abstract

As an emerging production factor, data assets are gaining strategic prominence, yet their application in collateralized financing faces persistent challenges, including illiquidity and risk evaluation complexities. This study introduces an innovative P max model to enhance the Collateral Value of data assets through insurance mechanisms, systematically demonstrating the feasibility conditions under which risk transfer optimizes asset valuation and delineating implementation pathways to integrate data insurance with asset-backed financing. Building on the theoretical framework of Value-at-Risk (VaR), this study develops a dynamic valuation model to assess the value of the collateral before and after insurance. Our analysis shows that insurance coverage for potential losses significantly enhances financing viability when premiums satisfy P max . Empirical analysis employing Monte Carlo simulations reveals a nonlinear positive correlation between pledgees’ risk tolerance thresholds and the maximum acceptable premium P max . This study bridges theoretical gaps in understanding insurance-value relationships for data assets while providing conceptual foundations and operational blueprints to standardize data markets and foster financial innovation.

1. Background and Problem

As a crucial strategic resource in the digital economy era, the potential in value of data is being unleashed in the process of iterative technological upgrades. Through its profound permeation in circulation and transaction ecosystems, data have transcended the value boundaries of traditional production factors. Its value configuration is undergoing a paradigmatic transition from fundamental resource to strategic assets with transformative significance, progressively evolving into the central dynamic driving systemic reconstruction of social production.
In the digital economy era, data have been strategically positioned as a “new factor of production” and have emerged as a core driver of economic development. Its recognition as a distinct asset class has gained substantial traction within both academic and industrial domains. The 2023 White Paper on Data Asset Management Practice (Version 6.0), published by the China Academy of Information and Communications Technology (CAICT), has formally operationalized technical criteria for data asset management. This authoritative framework underscores that data assets must exhibit defining attributes, including electronic traceability, transactional measurability, capacity to generate socioeconomic value, and unambiguous ownership. Concurrently, the 2023 Guidelines for Data Asset Valuation issued by the China Appraisal Society (CAS) codify data assets as legally owned or controlled data resources that are monetarily quantifiable and capable of delivering direct or indirect economic benefits to identifiable entities.
Tracing the academic lineage, Ref. [1] was the first to systematically propose the concept of “data assets” in academia, defining them as “various types of digital media with usage rights and encoded in binary form.” This definition broke through the functional boundaries of traditional information carriers, elevating data from a technical resource to an asset category for the first time. It emphasized the core features of usage rights and digital encoding, thereby laying the foundation for subsequent theoretical research and industrial practice in data assets. From an information management perspective, Ref. [2] constructed a demand-oriented dynamic assetization model, emphasizing that only data units meeting specific needs can be transformed into assets and that the scale of assets exhibits elasticity with demand fluctuations. The report of [3] demonstrates the productive asset attributes of data from a macroeconomic perspective and proposes pathways for data capitalization. These theoretical developments indicate that data assets not only carry information value, but are also deeply embedded in the social production system in the form of resources and factors of production, demonstrating unique strategic value.
In the academic exploration of data asset valuation methodologies, scholars have developed diverse systematic approaches to assess economic worth. The income approach estimates the present value of future earnings generated by data assets using a reasonable discount rate. The cost approach accounts for process-related expenses, including data collation and maintenance costs. The market approach bases the discounted value on premiums paid by specific business partners. Notably, many studies measure the value of data assets from the perspective of value loss. Ref. [4] discussed the protection of data assets in the context of the digital economy, emphasizing the maximum potential loss of assets within a specific time frame. Ref. [5] discussed how the rise of the data economy is changing the nature of income sources and risks, analyzing how to measure data as an intangible asset from the perspectives of value and risk.
It is worth noting that risk factors constitute a critical dimension that cannot be overlooked in the valuation of data assets. Based on the theory of intangible assets, Ref. [6] explored the valuation and pricing of data assets for internet platform enterprises, with particular emphasis on the impact of risk and uncertainty on data value. Ref. [7] constructed a three-dimensional value valuation model of “development–application–risk” for enterprise data assets based on the concept of data potential and the entire data life cycle, enabling a more comprehensive value assessment. This shows that considering the risks of data assets can significantly improve the accuracy and comprehensiveness of data asset value assessment.
Accurately identifying and scientifically classifying the risks of data assets are crucial prerequisites for building effective risk management mechanisms and achieving reasonable pricing of data assets. By systematically clarifying the sources of risk and their impact mechanisms, it is possible not only to optimize resource allocation efficiency and enhance the utilization and market competitiveness of data assets, but also to comprehensively safeguard and enhance the intrinsic value of data assets. Existing literature on data asset risk mainly focuses on dimensions such as data security risk and liability risk.
In terms of data security risks, Ref. [8] pointed out that data security issues mainly stem from external attacks and internal threats, permeating every stage of data collection, storage, and processing. Ref. [9] proposed that enterprise IT data security mainly faces four core risks: data security, privacy protection, system reliability, and storage security. Ref. [10] analyzed the impact of data breaches on corporate value, revealing that secondary breaches cause greater harm, financial information leaks are the most severe, and replacing the CEO after the first violation can prevent future losses.
At the level of liability risk, Ref. [11] pointed out that while anonymization techniques (such as k-anonymity) are widely used for privacy protection in medical big data, their potential re-identification risks (e.g., through linkage attacks using external data) may lead to sensitive information disclosure, thereby resulting in legal compliance responsibilities and ethical risks. Ref. [12] discussed big technology and data privacy. Ref. [13] highlighted that predictive analytics using big data and AI, which infer sensitive attributes (e.g., health, ethnicity, income) by linking individual behavioral data to anonymized datasets, creates liability risks if such predictions result in privacy breaches or unauthorized disclosures about third parties.
As the market value of data assets as a new production factor becomes increasingly prominent, managing related risks has become a focus of industry attention. Drawing on mature models such as cyber insurance [14,15,16], intellectual property insurance [17,18] and technology transformation insurance, data asset insurance has entered the stage of practical exploration as an innovative risk management tool. This trend not only reflects the recognition of data assets as insurable objects but also highlights the real market demand for risk hedging mechanisms for data elements.
On the other hand, the financial attributes of data assets mean they can be used as collateral to help enterprises obtain funding. Ref. [19] proposed a data valuation model to enable companies to use data assets as collateral for loans in financial transactions. Ref. [20] reported focuses on the mechanisms through which digital assets (such as tokenized bonds, securities, and funds) optimize financing processes and enhance credit trust via Distributed Ledger Technology (DLT).
It is generally believed that by introducing data asset insurance, enterprises can demonstrate to financial institutions that their data assets are effectively protected, thereby reducing the risk assessment of pledge financing. Data asset insurance can not only enhance the market recognition of data assets as collateral, but also, through risk transfer and guarantee mechanisms, significantly improve the operability and security of data asset pledge financing. This provides strong support for activating enterprise data assets and broadening financing channels, while also promoting the healthy development of the data element market.
In financial market risk research, the VaR (Value at Risk) model is widely used to assess and quantify market risk.VaR emerged from the need for a single, comprehensible metric to capture market risk, famously endorsed by J.P. Morgan’s RiskMetrics system [21]. Mathematically, for a random loss variable L and a confidence level α , VaR α ( L ) is the α -quantile of the loss distribution. The link to insurance was established formally in the economic decision-theoretic literature. In the seminal work on premium principles, Ref. [22] laid the groundwork for considering upper-tail risks. This was later advanced by [23], who explicitly defined a mean-shortfall criterion. From this perspective, a risk measure focusing on a quantile naturally leads to an insurance contract with a deductible at that quantile. An insurance policy that indemnifies the loss ( L VaR α ( L ) ) + is a quantile insurance contract. The insured retains all losses below the α -quantile while transferring the tail risk to the insurer. Therefore, VaR is not merely a risk measure; it also serves as the contractual trigger point for a class of optimal insurance contracts. Given the beneficial role of VaR theory in risk management, it has also been widely applied in various specific fields. In the field of railway hazardous materials transportation, Ref. [24] proposed a method based on the VaR model, aiming to minimize the risks of railway hazardous materials transportation by optimizing train configurations. Ref. [25] incorporated weather factors into the VaR model in their research, validating it through historical data and the bootstrap method, thereby offering a new perspective for derivative valuation. The integration of the VaR model with other methods has demonstrated significant advantages in financial market risk management. It not only enhances the precision and flexibility of risk management but also provides robust support for enterprises in navigating complex market environments.
Based on this, our study innovatively evaluates data assets from a risk measurement perspective. We classify data asset risks into three stages: holding, operation, and value-added. Furthermore, we propose a P max model for enhancing the pledged value of data assets through insurance. By introducing the VaR method, this study systematically examines the changes in the pledged value of data assets before and after insurance, clarifying the specific pathways and conditions under which insurance can enhance the market value of data assets by reducing risk and strengthening financing capability. This opens up new ideas for data asset appreciation and market application. The research in this paper not only fills the gap in existing studies on the relationship between data asset insurance and value but also provides theoretical support and practical guidance for data asset pledge financing, promoting the healthy development of the data element market.
The remainder of this paper is organized as follows. Section 2 develops a risk-based valuation framework for data assets, in which risks are classified according to the stages of holding, operation, and value addition. Section 3 presents the theoretical P max model, establishing the feasibility conditions under which insurance arrangements can raise collateral value. This section also examines the standard deviation premium principle within the same framework. Section 4 offers a numerical illustration based on Monte Carlo simulation, which verifies the P max model and explores how risk tolerance influences both the maximum acceptable premium and the corresponding increase in collateral value. The paper concludes with Section 5, which summarizes key findings and discusses their broader implications.

2. Risk-Based Valuation of Data Assets

2.1. Data Assets and Risks

From a mathematical perspective, data assets can be conceptually identified as the totality of their future returns and losses (assuming discounting where appropriate), so that a data asset can be typically represented by X = U L , where U 0 is the totality of its returns and L 0 the totality of its losses. Since these returns and losses are subject to stochastic and uncertain factors—such as market conditions and technological failures—the values of data assets are treated as random variables. We would like to note that, clearly, U and L can be and are not necessarily the positive part X + = max { X , 0 } and negative part X = min { X , 0 } of X extensively referred to in mathematics. For example, if U and L are supposed statistically independent and following exponential distributions Exp ( λ U ) with density λ U e λ U x ( x > 0 ) and Exp ( λ L ) with density λ L e λ L x ( x > 0 ) , then X follows an asymmetric Laplace distribution with density λ U λ L λ U + λ L e λ U x , x 0 e λ L x , x < 0 . Thus X + is distributed as: Pr ( X + = 0 ) = λ U λ U + λ L and X + has density λ U λ L λ U + λ L e λ U x when x > 0 , so that U is absolutely different from X + in distribution.
The sources of risk-related losses are diverse. Specifically, these losses can be categorized diversely in different stages within the data asset lifecycle: holding, operation, and value-addition. This classification not only enables more precise identification and management of stage-specific risks, but also provides a solid theoretical basis for risk quantification and value assessment.
During the holding stage, risks primarily take the form of data leakage or damage. Although data assets at this stage have not yet entered active use or transaction, and thus exert only a minor direct impact on value realization, such losses are typically permanent, jeopardizing the integrity and availability of the data.
The operational stage is characterized by potential risks arising from the circulation and use of data assets, such as data breaches, legal compliance issues, or credit risks. In this stage, as data assets are leveraged to deliver products or services, their economic value and operational effectiveness become directly apparent. The risks encountered here tend to correlate strongly with asset value, and the resulting losses are often considered general losses within the insurance industry.
Risks encountered during the value-addition stage often stem from issues like legal liabilities emerging in the course of data product development. This stage is pivotal for generating further economic value through the development and utilization of data assets. Risks at this stage are most closely correlated with asset value and can substantially constrain value creation, frequently resulting in significant potential losses.
For an illustrative breakdown, see Table 1 below:

2.2. Risk-Based Valuation of Data Assets

From an operational perspective, data asset valuation has typically employed market-based model, economic models and dimensional models [26]. However, from a theoretical standpoint, data asset valuation fundamentally represents a risk measurement exercise because, as established earlier, a data asset can be conceptualized as a random variable encompassing the totality of its future returns and losses. Consequently, as a standard practice in risk management, the determined value depends not only on the distribution of these returns and losses (appropriately discounted), but also critically on the valuator’s specific purposes. This implies that the same data asset may yield different valuations based on the assessment purpose.
When considering the perspective of a pledgee (such as a financial institution), the primary purpose of valuation is to establish appropriate credit limits for the asset holder. For the pledgee, this valuation can be formalized as: determining the maximum of the deterministic quantities v 0 such that Pr ( X v ) α , i.e.,
V = max { v 0 : Pr ( X v ) α } = F X 1 ( α ) ,
where in (1), α represents a small predetermined risk threshold of the pledgee, typically 5% or 1%, and F X is the distribution function of random variable X. The value V (referred to as collateral value) ensures that the probability of the data asset X falling below V does not exceed α . Expressed differently, if a bank extends credit V against the data asset, the probability of the asset’s return dropping below this credit amount (potentially triggering insolvency) remains bounded by the predefined risk level α . Clearly, if Pr ( X 0 ) > α , then the data asset X is not valuable at the risk level α in the pledgee’s perspective. Different risk thresholds typically lead to different values of a same data asset.
From an insurer’s perspective, the analytical focus shifts to potential losses. As aforementioned, an asset is represented as X = U L , where U 0 and L 0 represent respectively the future random return and random loss of asset X and an insurer is not concerned with the whole data asset X itself but possibly a part L I of the loss L such that L I L . L I will be referred to as insured loss. The insurer valuates the data asset’s risk profile L (if Pr ( L = 0 ) < 1 ) to determine an appropriate underwriting premium. The objective becomes: for a given risk threshold β , identify the minimum of the numbers v 0 such that Pr ( L v ) 1 β , i.e.,
p = min { v 0 : Pr ( L v ) 1 β } .
In words, the insurer seeks to ensure that the probability of losses not exceeding p is at least 1 β . The value p can be understood as the value of the data asset X in the insurer’s perspective.
The valuation methodology above corresponds to the VaR approach widely employed in risk management. As demonstrated, banks and insurers arrive at distinct valuation assessments (V and p respectively) for the same data asset due to their fundamentally different purposes. This valuation differential creates the economic basis for insurance instruments to enhance the collateral value of data assets in secured financing arrangements.

3. Increasing Collateral Value of Data Assets Through Insurance

Let X denote a data asset (discounted if necessary), where X is a real-valued random variable and can take either positive or negative values (a profit is positive, a loss is negative). Obviously, a data asset that takes positive values with probability 1 does not require insurance, and a random variable that takes negative values with probability 1 has no collateral value. Therefore, this study only focuses on data assets where the probabilities of taking positive and negative values are both nonzero.
As discussed in Section 2, the data assets are valuated within the VaR risk measurement framework. VaR is based on the definition of loss. The VaR at a threshold α ( 0 , 1 ) is the minimal number y such that the probability of the loss Y : = X not exceeding y is at least ( 1 α ) . Mathematically, VaR α ( X ) is the ( 1 α ) quantile of Y, that is,
VaR α ( X ) = F Y 1 ( 1 α ) = sup { x R : F X ( x ) α } ,
where F X ( x ) = Pr ( X x ) is the probability distribution function of X. This is the general definition of VaR, and these two expressions are equivalent.
From the insurer’s perspective, using VaR to assess the value of a data asset is equivalent to the quantile premium principle. That is, for a given confidence level α , the premium p for a risk or loss L is set as the minimal y such that the probability of L exceeding y does not exceed α . This means that an insurer, according to their risk tolerance (indicated by α ), determines a premium level that covers loss in most cases (with probability 1 α ). Mathematically, for 0 < α < 1 , p is the solution to F L ( p ) = 1 α or Pr { L > p } = α , and can be denoted as p = F L 1 ( 1 α ) .

3.1. Insurance and Collateral Value Based on VaR Measurement

As mentioned above, from the perspective of the insurer, according to the premium pricing principle based on VaR,
p = F L 1 ( 1 α ) or Pr ( L > p ) = α ,
where p is the premium determined by the insurer based on the risk volatility of data assets, and α is the insurer’s risk tolerance. The above formula indicates that the probability of the insured risk loss exceeding the premium equals α .
From the perspective of the pledgee, under the VaR principle, the financing collateral value of the data asset X, denoted V ( X ) , is
Pr X V ( X ) = β or V ( X ) = F X 1 ( β ) ,
where β is the pledgee’s risk tolerance. The formula shows that the probability of the pledgee suffering a loss because the value of the pledged asset X is less than the estimated asset value V ( X ) equals β .

3.2. Changes in Insurance Coverage and the Collateral Value of Data Assets

Assume that X consists of U and the risk loss L, i.e., before insurance, the value of the data asset is
X = U L ,
where L is the loss portion of the data asset that is covered by insurance (other uncovered parts are included in U). Before insurance, from the perspective of the pledgee, the financing pledge value of the data asset is
V β ( X ) = F X 1 ( β ) = F U L 1 ( β ) .
If the risk loss L of the data asset is insured, with a premium of p, the post-insurance data asset becomes X ˜ , where the pre-insurance asset U is reduced by the premium expenditure p, and the original loss L changes to L f ( L ) due to the insurance compensation function (8),
X ˜ = U p L + f ( L ) .
Here, U represents the (discounted) future return of the data asset, p is the premium paid at the time of insurance, L represents the inherent loss of the data asset caused by risk events, and f ( L ) represents the compensation function provided by the insurer after loss occurs.
Suppose the insurance provides full compensation, i.e., f ( L ) = L , then X ˜ = U p . Therefore, after insurance, from the pledgee’s perspective, the financing collateral value of the data asset becomes
V β X ˜ = F X ˜ 1 ( β ) = F U p 1 ( β ) .
Obviously, when V β X ˜ > V β ( X ) , utilizing insurance rights can increase the financing collateral value of the data asset.
Next, we demonstrate the conditions under which V β X ˜ > V β ( X ) . For any y, define
g ( p , y ) = P r ( U L y ) Pr ( U p y ) = [ F ¯ U | L ( x + y ) F ¯ U | L ( p + y ) ] d F L ( x ) .
Note that g ( p , y ) satisfies: if L is “non-degenerate”, then g ( p , y ) is non-decreasing with respect to p; g L min < 0 and g L max > 0 . Define
P max ( y ) = max p : g ( p ) < 0 L min , L max ,
then p P max ( y ) implies Pr ( U L y ) Pr ( U p y ) . In other words, if y is the pre-insurance collateral value of the data asset U L (i.e., Pr ( U L y ) = 1 β ), then it must be that Pr ( U p y ) 1 β ; therefore, the post-insurance collateral value from the perspective of the pledgee (i.e., U p ) will certainly exceed y.
Let y β be the value of y that satisfies Pr ( U L y ) = 1 β , and let P max ( β ) = P max ( y β ) . Then P max ( β ) (abbreviated as P max ) is the maximum premium for which the collateral value of the data asset can be increased. The pledgee’s risk tolerance β varies, and the corresponding P max is also different. In addition, P max is related to the joint distribution of ( U , L ) . As long as the insurance premium p P max , the post-insurance collateral value will exceed the pre-insurance value.
The above principle of enhancement can also be further described from the perspective of the insurer. Let α min ( β ) = P r ( L P max ) , then p P max ( β ) is equivalent to α α min ( β ) (abbreviated as α min below); that is, if the insurer’s risk tolerance is greater than or equal to α min , the financing collateral value of the data asset can be enhanced via insurance rights.
When p P max as stated above, the financing collateral value of the data asset can be increased by the rights of insurance. At this time, the net financing enhancement for the data asset holder (pledge value increase minus premium expenditure) is denoted as follows:
U ( α , β ) = V β X ˜ V β ( X ) = F U p 1 ( β ) F U L 1 ( β ) = F U 1 ( β ) F L 1 ( 1 α ) F U L 1 ( β ) .
The net financing enhancement depends on U , L , and α , β .
Figure 1 illustrates the value composition of data assets and their relationships with insurance and pledge financing. The data asset holder defines the asset value as future returns minus losses (i.e., X = U L ). At this point, the data asset holder assesses the asset for financing, while the pledgee, based on their risk tolerance, expresses the value as V β ( X ) = F X 1 ( β ) = F U L 1 ( β ) , where β represents the pledgee’s risk tolerance.
Under the condition that p < P max , the holder pays an insurance premium p, calculated as p = F L 1 ( 1 α ) , where α denotes the insurer’s risk tolerance. After insurance is procured, the collateral value of the data asset becomes X ˜ = U p . At this point, when seeking financing from a pledgee with the same risk tolerance, the value is expressed as V β X ˜ = F X ˜ 1 ( β ) = F U p 1 ( β ) = F U 1 ( β ) p .
Moreover, Figure 1 demonstrates that, under the protection of the insurance mechanism, the potential loss borne by the data asset holder is reduced, thereby enhancing the value of the pledge financing to a certain extent. The specific increase can be defined as U ( α , β ) = V β X ˜ V β ( X ) = F U p 1 ( β ) F U L 1 ( β ) = F U 1 ( β ) F L 1 ( 1 α ) F U L 1 ( β ) . We name this model the P max model, as it provides an intuitive and integrated framework for understanding insurance, financing, and risk management of data assets.

3.3. Innovative Analysis of the Standard Deviation Premium Principle Under the P max Model Framework

When U and L follow normal distributions, the VaR criterion can be simplified to the standard deviation principle. According to the standard deviation premium principle, the corresponding insurance premium is
p = E ( L ) + η σ L ,
where η represents the insurer’s risk safety loading coefficient. A larger η indicates lower risk tolerance α for the insurer. Here, σ L is the standard deviation of the risk loss and E ( L ) is the expected value of the risk loss of the digital asset.
The valuation of data assets can be considered as the expected value of the data asset X = UL, see Equation (6). Due to the existence of digital asset risk, its valuation is less than the mathematical expectation. We can also use the standard deviation principle to view the valuation of data assets. Thus, before insurance, the asset valuation is
V β ( X ) = E ( U ) E ( L ) k σ U L .
Here, E ( U ) and E ( L ) denote the expected values, σ U L is the standard deviation of UL, and k is a function of β representing the pledgee’s risk safety loading coefficient. A larger k means lower risk tolerance β for the pledgee, making their valuation of data assets more conservative. The standard deviation principle ranks among the most frequently applied and easily interpretable premium principles in insurance practice. Its primary advantage stems from a transparent structure: the premium consists of the expected loss, covering the average claim, augmented by a risk load that scales with the statistical volatility of the loss. This offers a clear mechanism for insurers to price uncertainty. The link to the VaR framework is particularly clear under the normality assumption, where the VaR itself becomes a linear function of the standard deviation. Consequently, in such a context, the standard deviation principle serves as a practical approximation to a VaR-based premium, combining computational simplicity with sound theoretical grounding.
From the same perspective, Equation (8), after insurance, the insured data asset X ˜ is
X ˜ = U p = U E ( L ) η σ L .
Here, it is also assumed that the insurance provides full risk compensation. After insurance, the pledged financing value of the asset as follows:
V β X ˜ = E U E ( L ) η σ L k σ U E ( L ) η · σ L = E ( U ) E ( L ) η σ L k σ U .
Therefore, V β ( X ˜ ) V β ( X ) is equivalent to k σ U L η σ L + k σ U , which is also equivalent to
k η σ L σ U L σ U .
If U and L are independent, then k / η 1 , which means the pledgee’s safety loading coefficient is greater than that of the insurer. As the risk manager, the insurer generally has a higher risk tolerance. Therefore, k / η 1 can generally be satisfied.
Assuming the risk tolerance β of the pledgee is determined in the process of data asset pledge financing, then k is also determined. At this point, k / η 1 is equivalent to η η max ( β ) . Accordingly, the insurer’s risk tolerance α α min , and the premium p P max . That is, when these conditions are met, the use of insurance rights can enhance the value of data assets. This conclusion is consistent with that under the VaR risk measurement approach.

3.4. P max Model Contributions and Comparison

The P max model is the first to formally integrate insurance pricing (from the insurer’s perspective) with collateral valuation (from the pledgee’s perspective) within a unified VaR framework for data assets. It quantitatively establishes the maximum acceptable premium P max as a critical threshold for value enhancement, presents a novel contribution by being the first to formally integrate insurance pricing, from the insurer’s perspective, with collateral valuation, from the pledgee’s perspective, within a unified VaR framework tailored for data assets. This integration quantitatively establishes the maximum acceptable premium P max as a critical threshold for enhancing collateral value, thereby providing a clear criterion for evaluating the financial viability of insuring data assets.
Compared to existing models, prevailing data asset valuation methods primarily focus on the intrinsic value of the data asset itself, only qualitatively considering risks as an adjunct to the core valuation approach. Furthermore, the Value-at-Risk model itself has not yet been systematically applied to the valuation or value measurement of data assets. Therefore, our model innovatively adapts the VaR framework to establish a theoretical foundation for data asset pledging. Crucially, P max model serves as a practical decision-making tool by explicitly answering the critical question: “Under what premium conditions does insuring a data asset become financially viable for increasing its collateral value?” This dynamic integration of risk transfer (insurance) with asset financing represents a key advancement.

4. Numerical Example Verification of the P max Model

This section aims to verify the theoretical feasibility of the P max model through a concrete numerical example and to visually demonstrate the quantitative relationships among key variables, such as the risk tolerance of the pledgee β , the maximum acceptable premium P max and the enhancement of the net value U ( α , β ) , thus making the theoretical framework more tangible. To achieve this, we first need to specify the distributional assumptions for the core variables ( U , L ) .
We assume U N ( 1000 , 200 2 ) , where the mean μ = 1000 and the variance σ 2 = 200 2 . The mean parameter for this assumption is primarily informed by the concept of industry benchmark financial impact from the IBM “Cost of a Data Breach Report 2024”, mapped to the expected value of the data asset’s potential return. The variance parameter captures the volatility in value realization. This setting provides a representative theoretical scenario. It is assumed that there is no loss in 80 % of cases, while in 20 % of cases, a risk loss occurs and the amount of loss typically exhibits heavy-tailed characteristics, which we assume follows a log-normal distribution. Therefore, we can assume that the risk loss L of the data asset follows a mixed distribution: 80 % 0 + 20 % log N ( e 4 , 1 ) .
In this section, we abbreviate V β ( X ) as V ( X ) . To calculate P max . Based on the above distributions of ( U , L ) , we use Monte Carlo simulation to generate a dataset comprising 100,000 pairs of ( U , L ) , then calculate P max for different values of β , and plot a comparison chart of the pledged value of data assets before and after insurance.
Table 2 shows that, under a fixed α , for different values of β , both P max and α min can be calculated. According to the data in the table, as the value of β decreases, P max increases nonlinearly, while α min shows a decreasing trend. This means that as the pledgee’s risk tolerance increases (i.e., β decreases), the critical value of the maximum premium that allows insurance rights to enhance the collateral value becomes higher, implying that the actual required premium also increases.
Further calculations reveal how U ( α , β ) varies with α for different β , as shown in Figure 2 below. The following conclusions can be observed:
(1)
As α increases, the value of U ( α , β ) for all curves exhibits an upward trend; that is, as the insurer’s risk tolerance decreases, the amount by which the financing value increases will also rise.
(2)
The degree to which different β values affect the increase in financing value varies. A larger β (e.g., 0.2) corresponds to the smallest U ( α , β ) at the same α value, while a smaller β (e.g., 0.01) shows a relatively higher value. This indicates that in the data asset value model, the value of β has a significant effect on the amount by which the financing value increases, and this effect is independent of α . This provides important guidance for subsequent risk management and decision-making.
Further analysis, as shown in Figure 3 below, reveals that given a fixed β , the value of V ( X ) for uninsured data assets remains constant. For insured data assets, when assessing the collateral value, the value of V ( X ˜ ) depends on the insurer’s risk tolerance α for a given β . For the insurer, the greater the value of α , the higher the insurer’s risk tolerance, and the greater the collateral value value V ( X ˜ ) of the insured data asset. Therefore, as α increases, the corresponding increase in the collateral value after insurance, U ( α , β ) , also becomes larger. The corresponding values α min and P max can thus be obtained: only when the insurer’s tolerance α is greater than α min is it possible to enhance the collateral value of the data asset through insurance. When α exceeds the probability of loss occurrence, V ( X ˜ ) reaches its maximum, and the increase in the collateral value U ( α , β ) also reaches its maximum.
The parameters and distributions specified for U and L in this example are chosen to clearly demonstrate the model’s mechanics. The core conclusion of the model—the existence of a critical premium P max that enables insurance to enhance collateral value—is robust to the specific parameter choices, as it is rooted in the risk quantification logic of VaR and the risk transfer function of insurance. Naturally, if the distribution or parameters of U were changed (e.g., increased variance indicating higher risk), the specific numerical values in Table 2 and Figure 2 and Figure 3 would change accordingly (e.g., the base collateral value V ( X ) might decrease). However, the qualitative trends, such as the negative correlation between P max and β and the positive correlation between U ( α , β ) and α , are expected to hold. Future research will utilize real-world data to calibrate the distributions and parameters for more thorough robustness checks.

5. Concluding Remarks and Practical Application

This study introduces the P max model, a novel framework that quantifiably integrates insurance mechanisms into the valuation and financing of data assets. Our analysis, grounded in the VaR framework and supported by numerical simulations, yields the following principal findings:
First, by categorizing data asset risks across the holding, operational, and value-addition stages, we establish that the correlation between risk and asset value increases progressively. The insurance mechanism, by transferring these quantified risks, directly targets and mitigates the tail risks that most concern pledgees. Our Monte Carlo simulations (Section 4) visually demonstrate that this risk transfer is the foundational mechanism enabling the enhancement of collateral value.
Second, and most critically, the P max model provides a precise decision-making tool, which represents a key advancement over traditional valuation methods. It moves beyond qualitative risk consideration by explicitly answering the pivotal question: “Under what premium conditions does insurance become financially viable?” Our analytical and numerical results (e.g., Table 2 and Figure 2) confirm that when the premium p P max , which is nonlinearly influenced by the pledgee’s risk tolerance β , the post-insurance collateral value V ( X ˜ ) unequivocally exceeds its pre-insurance counterpart V ( X ) . This dynamic integration of risk transfer with asset financing is a core contribution of our study.
Furthermore, the “market mechanism + insurance” dual credit enhancement model proposed is not merely conceptual. The model’s structure implies that standardized platforms reduce valuation uncertainty (a component of U-L, thereby raising the base value V ( X ) upon which the insurance mechanism then superposes a further layer of protection and value enhancement via the P max condition. This synergy offers a clear pathway to reduce credit risk premiums for financial institutions.
This study is motivated by the core dilemma faced by data assets in collateralized financing: on one hand, data as a new factor of production carries significant strategic value; on the other hand, its characteristics—difficult-to-quantify risks and insufficient liquidity—severely constrain its financialization. The P max model proposed in this paper is designed to systematically address this dilemma, providing an operational pathway of “risk quantification–insurance enhancement–value uplift” for realizing the value of data assets.
It is acknowledged that, from a theoretical perspective, the P max model requires further refinement. For instance, future work could incorporate more realistic insurance contract terms, such as deductibles and coinsurance provisions. Additionally, it is crucial to investigate whether consistent conclusions can be derived under alternative premium principles, particularly when accounting for the heavy-tailed characteristics of data asset losses.
In practice, the model can be implemented through the following steps: First, based on the full lifecycle risk categorization of data assets (holding, operational, value-addition), use the VaR framework to quantify their risk exposure and base collateral value V ( X ) . Second, determine a reasonable premium p according to the insurer’s risk tolerance α , and use the P max model to assess whether purchasing insurance is economically viable. Finally, when p P max , the data asset holder can transform the asset into a more recognized form of collateral X ˜ through insurance, thereby securing a higher financing amount V ( X ˜ ) from financial institutions.
Looking ahead, with the maturation of data asset registration, valuation, and trading systems, as well as the enrichment of data insurance products, the P max model can be further embedded into on-market transactions and securitization processes, forming a closed-loop ecosystem of “market valuation + insurance enhancement + collateralized financing.” This will not only activate the financing function of corporate data assets but also promote the standardization and financial innovation of the data element market, ultimately contributing to the high-quality development of the digital economy.

Author Contributions

Conceptualization, N.Z. and C.Q.; Methodology, N.Z., C.Q. and X.W.; Software, X.W.; Validation, C.Q.; Formal analysis, N.Z. and X.W.; Investigation, C.Q.; Resources, Y.Z.; Data curation, X.W. and Y.Z.; Writing—original draft, N.Z. and C.Q.; Writing—review & editing, N.Z., X.W. and Y.Z.; Supervision, C.Q. and X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This study is supported by the 2025 Annual Research Project of the Insurance Society of China (Grant No.: ISCKT2025-N-1-15).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

The authors gratefully acknowledge the editor and two anonymous referees for their insightful comments, which have significantly improved the quality of this paper.

Conflicts of Interest

Author Yongchao Zhao was employed by the company Shanghai Data Exchange. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. van Niekerk, A.J. A Methodological Approach to Modern Digital Asset Management: An Empirical Study. Allied Acad. Int. Conf. Int. Acad. Case Stud. 2006, 13, 53. [Google Scholar]
  2. Liu, X.; Jiang, Y.; Pan, L. Application of Data Assets. In Proceedings of the 2018 2nd IEEE Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Xi’an, China, 25–27 May 2018; pp. 2414–2418. (In Chinese). [Google Scholar]
  3. Corrado, C.; Haskel, J.; Iommi, M.; Jona-Lasinio, C. Data as a Strategic Asset: Enhancing Productivity and Innovation in the Digital Age; OECD Economics Department Working Papers; OECD: Paris, France, 2022; 1731p. [Google Scholar]
  4. Xiao, B.; Lin, S.; Wu, C.; Sun, B. Leveraging Traditional Financial Asset Rrotection Methods for Digital Asset Security. Int. Rev. Econ. Financ. 2025, 104010. [Google Scholar] [CrossRef]
  5. Veldkamp, L. Valuing Data as an Asset. Rev. Financ. 2023, 17, 1545–1562. [Google Scholar]
  6. Li, A.; Wang, A.; Chi, Y.; Yu, J. Exploration of Data Asset Valuation for Internet Platform Companies Based on the DEA-BCC Model. Procedia Comput. Sci. 2024, 242, 1235–1242. [Google Scholar] [CrossRef]
  7. Xia, W.; Wu, Y.; Yu, H.; Chen, X. Data Asset Evaluation and Pricing: Based on the “Development-Application-Risk” Three-Dimensional Model. Financ. Account. Mon. 2024, 45, 89–95. (In Chinese) [Google Scholar]
  8. Jurgelane-Kaldava, I.; Batenko, A.; Krivins, A.; Durguti, E.; Garskaite-Milvydiene, K. Data Security Risks and their Prevention: Latvian Enterprises in the Context of Selected Countries. Entrep. Sustain. Issues 2024, 11, 78–90. [Google Scholar] [CrossRef] [PubMed]
  9. Subbalakshmi, S.; Madhavi, K. Security Challenges of Big Data Storage in Cloud Environment: A Survey. Int. J. Appl. Eng. 2023, 13, 13237–13244. [Google Scholar]
  10. Peng, J.; Zhang, H.; Mao, J.; Xu, S. Repeated Data Breaches and Firm Value. Econ. Lett. 2023, 224, 111001. [Google Scholar] [CrossRef]
  11. Jain, P.; Gyanchandani, M.; Khare, N. Big Data Privacy: A Technological Perspective and Review. J. Big Data 2016, 3, 25. [Google Scholar] [CrossRef]
  12. Conyon, M.J. Big technology and data privacy. J. Econ. 2022, 46, 1369–1385. [Google Scholar] [CrossRef]
  13. Mühlhoff, R. Predictive privacy: Collective data protection in the context of artificial intelligence and big data. Big Data Soc. 2023, 10, 20539517231166886. [Google Scholar] [CrossRef]
  14. Klumpes, P. Coordination of Cybersecurity Risk Management in the U.K. Insurance Sector. Geneva Pap. Risk Insur.-Issues Pract. 2023, 48, 332–371. [Google Scholar] [CrossRef] [PubMed]
  15. Heath, B. Before the Breach: The Role of Cyber Insurance in Incentivizing Data Security. Georg. Wash. Law Rev. 2018, 86, 1115. [Google Scholar]
  16. Uuganbayar, G.; Yautsiukhin, A.; Martinelli, F.; Massacci, F. Optimisation of cyber insurance coverage with selection of cost effective security controls. Comput. Secur. 2021, 101, 102121. [Google Scholar] [CrossRef]
  17. Koleva, F. Intellectual property insurance as a management tool. Int. J. Intellect. Prop. Manag. 2012, 5, 327–341. [Google Scholar] [CrossRef]
  18. Kumar, J. Insurance coverage in intellectual property litigation. J. Intellect. Prop. Rights 2008, 13, 234–238. [Google Scholar]
  19. Cheong, H.; Kim, B.; Vaquero, I.U. Data Valuation Model for Estimating Collateral Loans in Corporate Financial Transaction. Risk Financ. Manag. 2023, 16, 206. [Google Scholar] [CrossRef]
  20. World Economic Forum. Digital Assets, Distributed Ledger Technology and the Future of Capital Markets; World Economic Forum: Koloni, Switzerland, 2021. [Google Scholar]
  21. Morgan, J.P. Risk Metrics Technical Document, 4th ed.; Morgan Guaranty Trust Company of New York: New York, NY, USA, 1996. [Google Scholar]
  22. Bühlmann, H. Mathematical Methods in Risk Theory; Springer: Berlin/Heidelberg, Germany, 1970. [Google Scholar]
  23. Stone, B.K. A general class of three-parameter risk measures. J. Financ. 1973, 28, 675–685. [Google Scholar]
  24. Hosseini, S.D.; Verma, M. A Value-at-Risk (VAR) approach to routing rail hazmat shipments. Transp. Res. Part D Transp. Environ. 2017, 54, 191–211. [Google Scholar] [CrossRef]
  25. Bilan, Y.; Mentel, G.; Streimikiene, D.; Szetela, B. Weather Risk Management in the Weather-VaR Approach. Assumptions of Value-at-Risk Modeling. Econ. Comput. Econ. Cybern. Stud. Res. 2020, 54, 31–48. [Google Scholar]
  26. Fleckenstein, M.; Obaidi, A.; Tryfona, N. A Review of Data Valuation Approaches and Building and Scoring a Data Valuation Model. Harv. Data Sci. Rev. 2023, 5, 1. [Google Scholar] [CrossRef]
Figure 1. Integrated Diagram of the P max Data Asset Insurance Value Model.
Figure 1. Integrated Diagram of the P max Data Asset Insurance Value Model.
Mathematics 13 03596 g001
Figure 2. Collateral Value Increment from Insurance, U ( α , β ) .
Figure 2. Collateral Value Increment from Insurance, U ( α , β ) .
Mathematics 13 03596 g002
Figure 3. Comparison of Collateral Value Before and After Insurance.
Figure 3. Comparison of Collateral Value Before and After Insurance.
Mathematics 13 03596 g003
Table 1. Categories of Data Asset Risks.
Table 1. Categories of Data Asset Risks.
Holding Data AssetsOperating Data AssetsValue-Addition of Data Assets
Type of RiskFundamental RiskOperational RiskIncremental Risk
Correlation with Asset ValueLow RelativelyHighHigh
Form of LossPermanent LossGeneral LossPotential Loss
Table 2. α min and P max under different β .
Table 2. α min and P max under different β .
β P max α min
0.2051.56300.1712
0.1560.39230.1634
0.1076.63820.1492
0.05116.89980.1189
0.01176.23260.0865
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, N.; Qiu, C.; Wu, X.; Zhao, Y. Can the Collateral Value of a Data Asset Be Increased by Insurance? Mathematics 2025, 13, 3596. https://doi.org/10.3390/math13223596

AMA Style

Zhang N, Qiu C, Wu X, Zhao Y. Can the Collateral Value of a Data Asset Be Increased by Insurance? Mathematics. 2025; 13(22):3596. https://doi.org/10.3390/math13223596

Chicago/Turabian Style

Zhang, Nan, Chunjuan Qiu, Xianyi Wu, and Yongchao Zhao. 2025. "Can the Collateral Value of a Data Asset Be Increased by Insurance?" Mathematics 13, no. 22: 3596. https://doi.org/10.3390/math13223596

APA Style

Zhang, N., Qiu, C., Wu, X., & Zhao, Y. (2025). Can the Collateral Value of a Data Asset Be Increased by Insurance? Mathematics, 13(22), 3596. https://doi.org/10.3390/math13223596

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop