Next Article in Journal
Commodity Market Risk: Examining Price Co-Movements in the Pakistan Mercantile Exchange
Previous Article in Journal
Multi-Timescale Recurrent Neural Networks Beat Rough Volatility for Intraday Volatility Prediction
Previous Article in Special Issue
The Regime-Switching Structural Default Risk Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Use of Prediction Bias in Active Learning and Its Application to Large Variable Annuity Portfolios

1
Department of Statistical and Actuarial Sciences, Western University, London, ON N6A 3K7, Canada
2
Department of Statistical Science, Duke University, Durham, NC 27708, USA
*
Author to whom correspondence should be addressed.
Risks 2024, 12(6), 85; https://doi.org/10.3390/risks12060085
Submission received: 1 April 2024 / Revised: 2 May 2024 / Accepted: 12 May 2024 / Published: 22 May 2024
(This article belongs to the Special Issue Risks Journal: A Decade of Advancing Knowledge and Shaping the Future)

Abstract

:
Given the computational challenges associated with valuing large variable annuity (VA) portfolios, a variety of data mining frameworks, including metamodeling and active learning, have been proposed in recent years. Active learning, a promising alternative to metamodeling, enhances the efficiency of VA portfolio assessments by adaptively improving a predictive regression model. This is achieved by augmenting data for model training with strategically selected informative samples. Successful application of active learning requires an effective metric in order to gauge the informativeness of data. Current sampling methods, which focus on prediction error-based informativeness, typically rely solely on prediction variance and assume an unbiased predictive model. In this paper, we address the fact that prediction bias can be nonnegligible in large VA portfolio valuation and investigate the impact of prediction bias in both the modeling and sampling stages of active learning. Our experimental results suggest that bias-based sampling can rival the efficacy of traditional ambiguity-based sampling, with its success contingent upon the extent of bias present in the predictive model.

1. Introduction

Variable annuities (VAs), also known as segregated fund contracts in Canada, are a popular type of equity-linked insurance product often utilized as investment vehicles within North American retirement savings plans. VAs typically offer a range of guaranteed benefits aimed at providing protection, such as the Guaranteed Minimum Death Benefit (GMDB), Guaranteed Minimum Maturity Benefit (GMMB), Guaranteed Minimum Accumulation Benefit (GMAB), and Guaranteed Minimum Withdrawal Benefit (GMWB) (see, e.g., Hardy (2003); Ledlie et al. (2008)). A substantial body of literature examines the stochastic modeling of embedded guarantees and their valuation in a fair market (risk-neutral) setting. However, the valuation of these embedded guarantees is complex due to the inherent complexity of diverse benefit designs. For example, GMDB guarantees a minimum amount to the policyholder upon death, exposing the insurer to various risks including investment, interest rate, mortality, and surrender risks. For a comprehensive overview of pricing, valuation, and risk management of individual VA contracts, we refer readers to existing literature, including Feng et al. (2022) and references therein. This paper primarily focuses on the valuation of large portfolios of VAs using statistical tools.
Due to the complex nature of the guarantees’ payoffs, closed-form solutions are often not feasible. Consequently, insurance companies extensively utilize Monte Carlo (MC) simulations for valuation and dynamic hedging purposes. However, valuing a large portfolio of Variable Annuity (VA) contracts using Monte Carlo simulation is notably time-intensive, as it requires projecting each contract across numerous scenarios over an extended time period (Gan and Valdez 2018). To efficiently assess large VA portfolios within constrained timeframes or resources, numerous data mining techniques have been introduced. The predominant framework, metamodeling, integrates both supervised and unsupervised machine learning methods through several sequential modeling stages. It employs unsupervised learning to generate a small, representative dataset, followed by supervised learning to fit a predictive model to these data after valuation is completed. Literature reviews reveal various unsupervised learning algorithms such as the truncated fuzzy c-means algorithm (Gan and Huang 2017), conditional Latin hypercube sampling (Gan and Valdez 2018; Minasny and McBratney 2006), and hierarchical k-means clustering (Gan and Valdez 2019). On the other hand, supervised learning approaches for VA portfolio valuation include kriging (Gan 2013), GB2 (Gan and Valdez 2018), group LASSO (Gan 2018), and several tree-based methods (Gweon and Li 2023; Gweon et al. 2020; Quan et al. 2021).
An alternative data mining approach is active learning (Gweon and Li 2021; Settles 2012), which involves iterative and adaptive data sampling and model training. Within this framework, the data sampling stage identifies a batch of informative policies, integrating these selected contracts into the existing representative data set. This augmented data set is then used for predictive model training. This cyclical process of sampling and updating continues until the allocated computational resources for portfolio assessment are exhausted. A prevalent strategy in active learning is uncertainty or ambiguity sampling (Burbidge et al. 2007; Settles and Craven 2008), which targets contracts where the current model shows significant predictive uncertainty.
Active learning methodologies have predominantly focused on classification problems, where the response variable is categorical. When the output variable is continuous, uncertainty can be quantified by prediction error, which can be decomposed into variance and squared bias. Current approaches to informativeness-based sampling for regression problems typically rely on prediction variance, under the assumption that the predictive model is approximately unbiased (Burbidge et al. 2007; Krogh and Vedelsby 1995; Kumar and Gupta 2020). However, as noted by Gweon et al. (2020), there are practical instances where prediction bias is significant and cannot be overlooked. This recognition that prediction bias is a critical component of prediction error has spurred interest in using prediction bias as a measure of uncertainty in active learning. This paper explores the potential of bias-based sampling techniques in light of these findings.
To the best of our knowledge, this study represents pioneering research that explores the application of bias-based sampling in active learning for regression problems. We introduce a metric for bias-based sampling and investigate its utility under conditions where the predictive model exhibits significant or minimal bias. Empirical evaluations on a large synthetic VA portfolio with six response variables demonstrate the effectiveness of bias-based sampling using random forest (RF, Breiman (2001)), particularly compared to variance-based sampling. However, the effectiveness diminishes when bias reduction techniques for RF are implemented.
The remainder of the paper is structured as follows: Section 2 reviews the active learning framework and discusses the estimation and application of prediction bias within this context. Section 2.4 applies the proposed methods to the efficient valuation of a large synthetic VA dataset. Section 3 concludes the paper.

2. Main Methodology

In this section, we propose several strategies to incorporate prediction bias into the active learning framework. We begin with a brief overview of the active learning framework and explore how prediction uncertainty is measured using random forest. Subsequently, we detail the integration of prediction bias into the modeling and sampling stages of active learning.

2.1. The Active Learning Framework for Large VA Valuation

Supervised statistical learning typically involves a one-time model training session using a complete set of labeled data. When large unlabeled data are available during the model training phase, active learning that uses both labeled and unlabeled data offers a viable alternative to traditional supervised learning. The primary objective of active learning is to iteratively enhance the predictive model by incorporating additional, informative samples selected from the pool of unlabeled data. This approach is particularly advantageous for rapidly valuing large VA portfolios because there are a large number of unlabeled contracts available during the model training process. As described by Gweon and Li (2021), the active learning process comprises the following steps:
(1)
Formulate the initial representative data set by selecting a small subset from the entire portfolio, which can be chosen randomly.
(2)
Determine the values of the response variable for the selected contracts using Monte Carlo simulation, thereby generating the labeled data necessary for training the predictive model.
(3)
(Modeling stage) Train the predictive model using the labeled representative data.
(4)
(Sampling stage) Apply the predictive model to the remaining set of unlabeled VA contracts to identify a subset of contracts whose information could significantly improve the model.
(5)
Use a Monte Carlo simulation to label the newly selected contracts and integrate them into the existing set of labeled data, resulting in an expanded labeled dataset.
(6)
Repeat steps (3) through (5) until specific stopping criteria are met, typically when the allocated time for valuation expires.
(7)
Finalize the assessment by applying the updated predictive model to estimate the response values for the remaining unlabeled contracts.
The key part of an active learning algorithm is the selection of informative samples from the unlabeled data at each sampling iteration. Such informative samples lead to a great improvement in the current predictive model. Most existing approaches for informative sampling have been developed for classification problems and thus cannot be directly adapted to the VA valuation in which the target variables are continuous. Also, sampling a batch, instead of a single contract, is desirable for an efficient use of resources such as model runtime and parallel computing systems. Recently, Gweon and Li (2021) proposed batch-mode active learning approaches using random forest (Breiman 2001) for an efficient valuation of large VA portfolios. To quantify the informativeness of each VA contract, they adopted the principle of Query-By-Committee (Abe and Mamitsuka 1998; Burbidge et al. 2007; Freund et al. 1997) and used prediction ambiguity that is equivalent to the sample variance of bootstrap regression trees. They found that active learning can be highly effective when the informative samples are chosen by a weighted random sampling for which the sampling weights are proportional to prediction ambiguities.
A critical component of the active learning algorithm involves selecting informative samples from the unlabeled data during each iteration. These samples substantially enhance the accuracy and robustness of the existing predictive model. Traditional approaches for informative sampling, primarily developed for classification problems, are not directly applicable to VA valuation, where the target variables are continuous. Moreover, batch sampling—as opposed to selecting individual contracts—is desirable for an efficient use of resources, such as model runtime and parallel computing systems. Recently, Gweon and Li (2021) introduced batch-mode active learning strategies using random forest for the efficient valuation of large VA portfolios. To assess the informativeness of each contract, they implemented the Query-By-Committee principle (Breiman 2001) and measured prediction ambiguity, akin to the sample variance of bootstrap regression trees. Their findings indicate that active learning can be exceptionally effective when informative samples are selected through weighted random sampling, where the weights are proportional to the prediction ambiguities.

2.2. Random Forests and Quantifying Prediction Bias

Suppose that a VA portfolio contains a total of N contracts, X = { x 1 , , x N } where each x i R p represents a vector of feature variables associated with a contact. Random forest predicts the continuous response value Y i for each contract x i using the regression form:
Y i = f ( x i ) + ϵ i ,
where f ( · ) represents the underlying model, and ϵ i denotes random error. This modeling approach employs bootstrap regression trees (Breiman 1984) to estimate the regression function, with foundational concepts detailed in Gweon et al. (2020) and Quan et al. (2021).
Let L represent the labeled representative data containing n VA contracts. Random forest consisting of M regression trees constructs each regression tree f ^ L m ( x ) using the m-th bootstrap sample L m ( m = 1 , , M ). The predicted response value for an unlabeled contract x is then derived by averaging across all M bootstrap regression trees:
f ^ ( x ) = 1 M m = 1 M f ^ L m ( x ) .
The effectiveness of random forest in VA portfolio valuation has been demonstrated in prior studies (Gweon et al. 2020; Quan et al. 2021).
Given the necessity of quantifying prediction uncertainty in active learning, we focus on the mean square error (MSE), defined as:
MSE ( f ^ ( x ) ) = E ( f ^ ( x ) f ( x ) ) 2 .
MSE is decomposed into the prediction (squared) bias and variance:
MSE ( f ^ ( x ) ) = E ( f ^ ( x ) E ( f ^ ( x ) ) + E ( f ^ ( x ) ) f ( x ) ) 2 = E ( f ^ ( x ) ) f ( x ) 2 + E ( f ^ ( x ) E ( f ^ ( x ) ) ) 2 ,
where the first term represents the squared bias and the second term, the variance.
Although the averaging technique in random forest greatly improves the predictive performance of regression trees by reducing the prediction variance, it does not influence the prediction bias. With the response values in a VA portfolio potentially exhibiting a highly skewed distribution, Gweon et al. (2020) identified that the prediction bias of random forest is not negligible. Gweon et al. (2020) showed a bias-correction technique (Breiman 1999; Zhang and Lu 2012) can effectively estimate and correct prediction bias. Specifically, prediction bias can be assessed by fitting an additional random forest model to the out-of-bag (OOB) errors of the trained model. The OOB prediction for a data vector x i in the labeled set is defined as:
f ^ O O B ( x i ) = 1 M i m = 1 M f ^ L m ( x i ) I ( ( x i , y i ) L m ) ,
where I ( · ) denotes the indicator function, and M i is the number of bootstrap samples not containing ( x i , y i ) (i.e., M i = m = 1 M I ( ( x i , y i ) L m ) ). A subsequent model g ( · ) is then trained where the response variable is the bias B ( x ) = f ^ O O B ( x ) Y , allowing the estimation of the prediction bias:
B ^ ( x ) = g ^ ( x ) .
Combining the two random forest models, the bias-corrected random forest (BC-RF) predicts the response variable as the sum of f ^ ( x ) and g ^ ( x ) . Gweon et al. (2020) utilized the quantity B ^ ( x ) exclusively for mitigating prediction bias within the RF model as applied in the metamodeling framework. In the following section, we discuss how this bias is leveraged within the active learning process.

2.3. Using Prediction Bias in Active Learning Process

Our focus is to explore the integration of prediction bias, B ^ ( x ) , into the sampling stage of the active learning process. Instead of developing a new sampling methodology, we adopt the weighted random sampling (WRS) approach from Gweon and Li (2021), which is already established as effective. Utilizing prediction bias within this framework is straightforward. Under WRS with prediction bias, the active learning algorithm selects a batch of unlabeled contracts of size d, where the sampling probability for a contract x i within the unlabeled data set U is defined as:
w i = B i 2 h B h 2 ,
where B i = B ^ ( x i ) for i = 1 , , | U | . We employ the squared bias terms to ensure non-negative sampling probabilities for all contracts and to facilitate a natural combination with the ambiguity component since the mean square error of the regression trees is the sum of variance and squared bias. We denote this approach as WRS-B to differentiate it from the ambiguity-based WRS (WRS-A), which calculates the weight as:
w i = A i h A h ,
where the ambiguity score A i = 1 M 1 m = 1 M f ^ L m ( x i ) f ^ ( x i ) 2 represents the sample variance of the bootstrap regression trees, as detailed in Gweon and Li (2021).
When is WRS-B particularly effective in active learning? We posit that its effectiveness is contingent on the severity of prediction biases. If the predicted biases of the target response variable are significant, the weight w i in WRS-B acts as a measure of relative prediction uncertainty for the i-th contract. Conversely, if predictions are nearly unbiased, the bias estimates ( B i ) may merely introduce noise and offer no substantial information, rendering the use of these scores a less effective uncertainty measure.
This analysis suggests that WRS-B could serve as a promising uncertainty measure for RF in the valuation of large VA portfolios in the presence of prediction biases, as demonstrated by Gweon et al. (2020). A deeper analysis is required for BC-RF, which predicts the response variable using two sequential RF models. In the j-th iteration of active learning, the prediction by BC-RF is formulated as:
Y ^ = f ^ j ( x ) + g ^ j ( x ) ,
where f ^ j ( x ) is the response prediction by an RF model, and g ^ j ( x ) is the predicted bias by a subsequent RF model. If the primary contribution of WRS-B to the enhancement of f ^ j + 1 ( x ) in the ( j + 1 ) -th iteration is bias reduction, the updated model f ^ j + 1 ( x ) will yield less biased predictions, potentially reducing the impact of g ^ j + 1 ( x ) .
Furthermore, we need to examine the relative effectiveness of WRS-B compared to WRS-A. If prediction ambiguity and bias are highly correlated, both sampling approaches may yield similar outcomes. However, if they are not, the samples selected by WRS-B and WRS-A are likely to differ. Moreover, these two uncertainty components—ambiguity and bias—can be combined in various ways. Since their sum corresponds to the mean square error of the individual trees, another variant, WRS based on mean square error (WRS-MSE), uses the weight:
w i = A i + B i 2 h ( A h + B h 2 ) .
In Section 2.4, we further explore the application of prediction bias in active learning across various settings.

2.4. Application in Large Variable Annuity Portfolio Valuation

2.5. Description of Data

The data utilized in this study comprises a synthetic dataset from Gan and Valdez (2017), containing 190,000 variable annuity (VA) contracts. These contracts are characterized by 16 explanatory variables (14 continuous and 2 categorical), as summarized in Table 1. Each of the ten investment funds within the dataset is linked to one or more market indices, with the specific mappings to the five market indices detailed in Table 2.
For the response variables, we analyze six variables: the fair market values (FMV) and five deltas. The FMV of a VA contract is defined as the difference between the guaranteed payoff and the guaranteed cost. The delta measures the rate of change of the theoretical option value relative to changes in the underlying asset’s price for a specific market index. The partial dollar deltas for a VA contract are calculated as follows:
Delta i = F M V ( P A 1 , , 1.01 P A i , , P A 5 ) 0.02 F M V ( P A 1 , , 0.99 P A i , , P A 5 ) 0.02 ,
where P A i denotes the partial account value linked to the i-th market index and FMV is the fair market value as a function of partial account values. The partial dollar delta indicates the sensitivity of the guarantee value to fluctuations in the index, which is critical for determining appropriate hedge positions.
Figure 1 illustrates the distribution of the six response variables. Observations indicate that the distribution of FMV is significantly skewed to the right, while the distributions of the deltas are predominantly skewed to the left.

2.6. Experimental Setting

The active learning framework was initiated with a set of 100 representative contracts, selected through simple random sampling without replacement, to fit an initial regression model. For the response variable and the bias estimation, we employed two RF models, each using 300 regression trees to ensure stable performance, as supported by Gweon et al. (2020); Quan et al. (2021). Furthermore, consistent with findings from Quan et al. (2021), all 16 features were utilized in each binary split during tree construction, which has been shown to optimize performance for this dataset.
During each iteration of active learning, each sampling approach selected 50 contracts based on sampling probabilities. These contracts were then incorporated into the representative data set, and the RF models were subsequently updated with this augmented data. The iterative process was terminated once the representative data set expanded to include 800 contracts. To mitigate the impact of random variability, each active learning method was executed 20 times using different seeds. This experimental setup was consistently applied across all six response variables.
For each of the five partial deltas, it was noted that some contracts exhibited zero delta values, indicating no association with the relevant market index. Given that the link between a contract and its market index is predetermined, contracts with zero delta values were excluded from our prediction assessments to maintain the integrity and relevance of the analysis.

2.7. Evaluation Measures

To assess predictive performance, we employ several metrics: R 2 , mean absolute error ( MAE ), and absolute percentage error ( APE ):
R 2 = 1 i ( f ^ ( x i ) f ( x i ) ) 2 i ( f ( x i ) f ¯ ( x ) ) 2 ,
MAE = 1 N i | f ^ ( x i ) f ( x i ) | ,
and
APE = | i f ( x i ) i f ^ ( x i ) | i f ( x i ) ,
where f ¯ ( x ) = N 1 i f ( x i ) . The R 2 and MAE measure accuracy at the individual contract level, while APE assesses portfolio-level accuracy, allowing for the offsetting of positive and negative prediction errors at the contract level. Higher values of R 2 indicate better performance, whereas lower values of MAE and APE are preferable.

2.8. Results

Initial investigations focused on identifying prediction bias across the six response variables. Figure 2 presents the out-of-bag (OOB) error of RF as a function of the true response variable. An observed increasing or decreasing pattern suggests a likely severe bias in RF predictions of the remaining data, underscoring the potential of prediction bias as a valuable uncertainty measure in active learning.
Figure 3 compares the performance of RF when using simple random sampling (RS), WRS-A and WRS-B. Both WRS-A and WRS-B consistently outperformed RS in terms of R 2 and MAE across all response variables as the active learning process progressed. However, in terms of APE , WRS-B tended to show more biased performance than RS. No clear advantage was observed between WRS-A and WRS-B; for instance, WRS-B yielded better R 2 results than WRS-A for FMV, while WRS-A excelled over WRS-B for delta 5. The two sampling methods performed comparably for other response variables.
The observed correlations between ambiguity and squared bias ranged from 0.4 to 0.5, indicating that WRS-A and WRS-B selected somewhat different contracts during the sampling process. Figure 4 illustrates the average prediction standard errors and absolute biases under WRS-A and WRS-B, noting a gradual decrease across multiple iterations of active learning. Interestingly, WRS-A was more effective in reducing prediction standard errors than WRS-B, which excelled in diminishing prediction bias. These findings demonstrate distinct contributions of WRS-A and WRS-B to model improvement.
Subsequently, we explore the utilization of prediction bias in the modeling and sampling stages. Table 3 compares the performance of RF with BC-RF under each sampling approach. The introduction of bias correction in the modeling stage significantly enhanced predictive performance across all metrics. Notably, the large APE previously observed in RF under both WRS-A and WRS-B were considerably reduced in BC-RF. Given these comparative results, BC-RF is recommended for use in active learning regardless of the sampling strategy.
We further examine the performance of BC-RF using WRS-A, WRS-B, and WRS-MSE in Figure 5. Contrary to the inconclusive results observed with RF where no distinct advantage was noted between WRS-A and WRS-B, the implementation of BC-RF in the modeling stage generally showed WRS-A to be more effective than WRS-B in most situations. This observation aligns with earlier findings indicating that WRS-B substantially contributes to bias reduction in RF (refer to Figure 4). Given that RF under WRS-B is relatively less biased compared to its counterpart under WRS-A, the subsequent application of BC-RF yields smaller gains from further bias reduction. This outcome is evidenced in Table 4 and Table 5, where the transition from RF to BC-RF exhibits more pronounced improvements under WRS-A.
Table 6 presents the runtime for each method until the representative VA contracts reached 800. When utilizing RF, WRS-A required approximately half the time compared to the other methods, as it did not involve an additional RF model for estimating prediction bias. In contrast, when BC-RF was employed, all sampling methods showed similar runtimes.
Interestingly, WRS-MSE often matched or exceeded the performance of WRS-A, as noted in the cases of Delta 1 and Delta 4. Considering that the computational times for WRS-MSE and WRS-A are nearly identical, these findings support a preference for WRS-MSE. Based on our analysis, we recommend employing BC-RF in the modeling stage alongside WRS-A or WRS-MSE in the sampling stage as an effective active learning strategy.

3. Concluding Remarks

In this study, we explored the integration of prediction bias within the active learning framework, specifically for the valuation of large variable annuity portfolios. Using a synthetic portfolio with six distinct response variables, we assessed the utility of prediction bias in both the modeling and sampling stages of active learning. Our experimental findings indicate that (a) employing prediction bias for uncertainty sampling (WRS-B) enhances performance of RF in terms of R 2 and MAE , proving more effective than random sampling and as competitive as ambiguity-based uncertainty sampling (WRS-A); (b) uncertainty sampling based on prediction bias significantly contributes to the reduction of bias in RF; (c) prediction bias proves beneficial in the modeling stage across different sampling methods; and (d) WRS-B tends to be less effective than WRS-A when bias reduction is actively pursued during the modeling phase.
The batch-mode active learning framework necessitates an initial set of labeled representative data and a defined batch size. Although we employed simple random sampling to generate the initial datasets, alternative strategies involving unsupervised learning algorithms might be viable. Our tests with conditional Latin hypercube sampling indicated that the unsupervised learning method did not significantly influence the outcomes within the active learning framework. Furthermore, the batch size inversely affects the runtime of the active learning process; a larger batch size decreases the number of iterations needed to achieve the desired dataset size, as more data are labeled and incorporated at each step. Our analysis, initially based on a batch size of 50, showed that increasing this to 100 maintained similar performance across different sampling methods.
In conclusion, this paper highlights the potential of prediction bias as an effective uncertainty measure within active learning for actuarial applications. While our approach utilized RF, other statistical learning algorithms could be considered if prediction bias can be appropriately quantified. A comprehensive examination of bias-based active learning employing various statistical methods merits further investigation.

Author Contributions

Conceptualization, H.G. and S.L.; methodology, H.G.; software, Y.X.; validation, H.G., S.L. and Y.X.; formal analysis, S.L. and Y.X.; investigation, Y.X.; resources, H.G.; data curation, Y.X.; writing—original draft preparation, H.G. and Y.X.; writing—review and editing, H.G. and S.L.; visualization, Y.X.; supervision, H.G. and S.L.; project administration, H.G. and S.L.; funding acquisition, H.G. and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the Natural Sciences and Engineering Research Council of Canada (NSERC, grant numbers 06219 (PI: Shu Li) and 04698 (PI: Hyukjun Gweon)).

Data Availability Statement

The synthetic dataset used in this article is available at https://www2.math.uconn.edu/~gan/software.html (accessed on 12 May 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Abe, Naoki, and Hiroshi Mamitsuka. 1998. Query learning strategies using boosting and bagging. Paper presented at the 15th International Conference on Machine Learning (ICML), Madison, WI, USA, July 24–27; pp. 1–9. [Google Scholar]
  2. Breiman, Leo. 1984. Classification and Regression Trees. Boca Raton: Taylor & Francis, LLC. [Google Scholar]
  3. Breiman, Leo. 1999. Using Adaptive Bagging to Debias Regressions. Technical Report. Berkeley: University of California at Berkeley. [Google Scholar]
  4. Breiman, Leo. 2001. Random forests. Machine Learning 45: 5–32. [Google Scholar] [CrossRef]
  5. Burbidge, Robert, Jem J. Rowland, and Ross D. King. 2007. Active learning for regression based on query by committee. In Intelligent Data Engineering and Automated Learning-IDEAL 2007. Edited by H. Yin, P. Tino, E. Corchado, W. Byrne and X. Yao. Berlin and Heidelberg: Springer, pp. 209–18. [Google Scholar]
  6. Feng, Runhuan, Guojun Gan, and Ning Zhang. 2022. Variable annuity pricing, valuation, and risk management: A survey. Scandinavian Actuarial Journal 2022: 1–34. [Google Scholar] [CrossRef]
  7. Freund, Yoav, H. Sebastian Seung, Eli Shamir, and Naftali Tishby. 1997. Selective sampling using the query by committee algorithm. Machine Learning 28: 133–68. [Google Scholar] [CrossRef]
  8. Gan, Guojun. 2013. Application of data clustering and machine learning in variable annuity valuation. Insurance: Mathematics and Economics 53: 795–801. [Google Scholar]
  9. Gan, Guojun. 2018. Valuation of large variable annuity portfolios using linear models with interactions. Risks 6: 71. [Google Scholar] [CrossRef]
  10. Gan, Guojun, and Jimmy Xiangji Huang. 2017. A data mining framework for valuing large portfolios of variable annuities. Paper presented at the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, August 13–17; pp. 1467–75. [Google Scholar]
  11. Gan, Guojun, and Emiliano A. Valdez. 2017. Valuation of large variable annuity portfolios: Monte carlo simulation and synthetic datasets. Dependence Modeling 5: 354–74. [Google Scholar] [CrossRef]
  12. Gan, Guojun, and Emiliano A. Valdez. 2018. Regression modeling for the valuation of large variable annuity portfolios. North American Actuarial Journal 22: 40–54. [Google Scholar] [CrossRef]
  13. Gan, Guojun, and Emiliano A. Valdez. 2019. Data clustering with actuarial applications. North American Actuarial Journal, 1–19. [Google Scholar] [CrossRef]
  14. Gweon, Hyukjun, and Shu Li. 2021. Batch mode active learning framework and its application on valuing large variable annuity portfolios. Insurance: Mathematics and Economics 99: 105–15. [Google Scholar] [CrossRef]
  15. Gweon, Hyukjun, and Shu Li. 2023. A hybrid data mining framework for variable annuity portfolio valuation. ASTIN Bulletin: The Journal of the IAA 53: 580–95. [Google Scholar]
  16. Gweon, Hyukjun, Shu Li, and Rogemar Mamon. 2020. An effective bias-corrected bagging method for the valuation of large variable annuity portfolios. ASTIN Bulletin: The Journal of the IAA 50: 853–71. [Google Scholar] [CrossRef]
  17. Hardy, Mary. 2003. Investment Guarantees: Modelling and Risk Management for Equity-Linked Life Insurance. Hoboken: John Wiley & Sons, Inc. [Google Scholar]
  18. Krogh, Anders, and Jesper Vedelsby. 1995. Neural network ensembles, cross validation, and active learning. In Advances in Neural Information Processing Systems. Cambridge: MIT Press, pp. 231–38. [Google Scholar]
  19. Kumar, Punit, and Atul Gupta. 2020. Active learning query strategies for classification, regression, and clustering: A survey. Journal of Computer Science and Technology 35: 913–45. [Google Scholar] [CrossRef]
  20. Ledlie, M. Colin, Dermot P. Corry, Gary S. Finkelstein, Alan J. Ritchie, Ken Su, and D. C. E. Wilson. 2008. Variable annuities. British Actuarial Journal 14: 327–89. [Google Scholar] [CrossRef]
  21. Minasny, Budiman, and Alex B. McBratney. 2006. A conditioned latin hypercube method for sampling in the presence of ancillary information. Computers & Geosciences 32: 1378–88. [Google Scholar]
  22. Quan, Zhiyu, Guojun Gan, and Emiliano Valdez. 2021. Tree-based models for variable annuity valuation: Parameter tuning and empirical analysis. Annals of Actuarial Science 16: 95–118. [Google Scholar] [CrossRef]
  23. Settles, Burr. 2012. Active Learning. San Rafael: Morgan & Claypool. [Google Scholar]
  24. Settles, Burr, and Mark Craven. 2008. An analysis of active learning strategies for sequence labeling tasks. Paper presented at the Conference on Empirical Methods in Natural Language Processing, Honolulu, HI, USA, October 25–27; pp. 1070–79. [Google Scholar]
  25. Zhang, Guoyi, and Yan Lu. 2012. Bias-corrected random forests in regression. Journal of Applied Statistics 39: 151–60. [Google Scholar] [CrossRef]
Figure 1. Distributions of the six response variables.
Figure 1. Distributions of the six response variables.
Risks 12 00085 g001
Figure 2. Out-of-bag prediction error results of RF on the initial representative data. Each panel corresponds to one of the six response variables.
Figure 2. Out-of-bag prediction error results of RF on the initial representative data. Each panel corresponds to one of the six response variables.
Risks 12 00085 g002
Figure 3. Predictive performance of RF under simple random sampling (solid), WRS-A (dotted), and WRS-B (dashed).
Figure 3. Predictive performance of RF under simple random sampling (solid), WRS-A (dotted), and WRS-B (dashed).
Risks 12 00085 g003
Figure 4. Averaged uncertainty of RF predictions for unlabeled data under WRS-A (dashed) and WRS-B (solid).
Figure 4. Averaged uncertainty of RF predictions for unlabeled data under WRS-A (dashed) and WRS-B (solid).
Risks 12 00085 g004
Figure 5. Boxplot results of BC-RF under different active learning sampling approaches.
Figure 5. Boxplot results of BC-RF under different active learning sampling approaches.
Risks 12 00085 g005
Table 1. Summary statistics of the predictor attributes in the VA dataset.
Table 1. Summary statistics of the predictor attributes in the VA dataset.
CategoryDescription Count
GenderM (Male Policyholder) 113,993
F (Female policyholder) 76,007
Product typeABRP (GMAB with return of premium) 10,000
ABRU (GMAB with annual roll-up) 10,000
ABSU (GMAB with annual ratchet) 10,000
DBAB (GMDB + GMAB with annual ratchet) 10,000
DBIB (GMDB + GMIB with annual ratchet) 10,000
DBMB (GMDB + GMMB with annual ratchet) 10,000
DBRP (GMDB with return of premium) 10,000
DBRU (GMDB with annual roll-up) 10,000
DBSU (GMDB with annual ratchet) 10,000
DBWB (GMDB + GMWB with annual ratchet) 10,000
IBRP (GMIB with return of premium) 10,000
IBRU (GMIB with annual roll-up) 10,000
IBSU (GMIB with annual ratchet) 10,000
MBRP (GMMB with return of premium) 10,000
MBRU (GMMB with annual roll-up) 10,000
MBSU (GMMB with annual ratchet) 10,000
WBRP (GMWB with return of premium) 10,000
WBRU (GMWB with annual roll-up) 10,000
WBSU (GMWB with annual ratchet) 10,000
ContinuousDescriptionMinimumMeanMaximum
gmwbBalanceGMWB balance035,611.54499,708.73
gbAmtGuaranteed benefit amount0326,834.591,105,731.57
FundValue1Account value of the 1st investment fund033,433.871,099,204.71
FundValue2Account value of the 2nd investment fund038,542.811,136,895.87
FundValue3Account value of the 3rd investment fund026,740.18752,945.34
FundValue4Account value of the 4th investment fund026,141.80610,579.68
FundValue5Account value of the 5th investment fund023,026.50498,479.36
FundValue6Account value of the 6th investment fund035,575.671,091,155.87
FundValue7Account value of the 7th investment fund029,973.25834,253.63
FundValue8Account value of the 8th investment fund030,212.11725,744.64
FundValue9Account value of the 9th investment fund029,958.29927,513.49
FundValue10Account value of the 10th investment fund029,862.24785,978.60
ageAge of the policyholder34.5249.4964.46
ttmTime to maturity in years0.5914.5428.52
Table 2. Ten investment funds. Each row is a mapping from an investment fund to a combination of five indices.
Table 2. Ten investment funds. Each row is a mapping from an investment fund to a combination of five indices.
FundUS LargeUS SmallIntl EquityFixed IncomeMoney Market
110000
201000
300100
400010
500001
60.60.4000
70.500.500
80.5000.50
900.30.700
100.20.20.20.20.2
Table 3. Predictive performance of RF and BC-RF under each sampling approach (N = 800).
Table 3. Predictive performance of RF and BC-RF under each sampling approach (N = 800).
R 2 MAEAPE
Response Sampling RF BC-RF RF BC-RF RF BC-RF
FMVRS0.91920.948025821.80721142.4560.02530.0176
WRS-A0.95370.969023813.41119586.2610.04090.0104
WRS-B0.95780.965022804.38720953.2400.02790.0119
Delta 1RS0.69620.73969078.1347812.3800.01970.0143
WRS-A0.77290.81148761.1087469.9060.04250.0139
WRS-B0.77610.80628847.6887877.7130.04120.0135
Delta 2RS0.68210.67356464.2445831.8080.02340.0171
WRS-A0.79030.82786282.4345448.8180.04930.0194
WRS-B0.79490.81736326.9275742.3360.05960.0185
Delta 3RS0.64830.69266899.1865992.5170.03120.0175
WRS-A0.73880.79286603.9415544.8160.04900.0132
WRS-B0.74430.78026708.5725925.1640.07620.0310
Delta 4RS0.70590.74704673.5813973.6290.04620.0194
WRS-A0.81260.86474486.3393573.0410.03950.0124
WRS-B0.82700.85744413.9753905.5420.06290.0304
Delta 5RS0.77420.77664524.4833960.9670.04250.0159
WRS-A0.91450.93114072.9073438.1340.03950.0079
WRS-B0.90380.91114091.3173716.2840.03730.0142
Table 4. Relative improvement (in %) obtained by switching RF to BC-RF under WRS-A and WRS-B (N = 500).
Table 4. Relative improvement (in %) obtained by switching RF to BC-RF under WRS-A and WRS-B (N = 500).
R 2 MAEAPE
WRS-A WRS-B WRS-A WRS-B WRS-A WRS-B
FMV1.800.8715.548.9660.5432.84
Delta 17.463.1515.429.1846.3160.88
Delta 25.183.6311.889.2954.4767.90
Delta 38.986.5014.8611.4247.0554.50
Delta 48.044.3718.3210.0844.9841.40
Delta 52.071.4915.229.6063.5562.05
Table 5. Relative improvement (in %) obtained by switching RF to BC-RF under WRS-A and WRS-B (N = 800).
Table 5. Relative improvement (in %) obtained by switching RF to BC-RF under WRS-A and WRS-B (N = 800).
R 2 MAEAPE
WRS-A WRS-B WRS-A WRS-B WRS-A WRS-B
FMV1.600.7517.758.1274.4757.29
Delta 14.993.8814.7410.9667.1967.15
Delta 24.752.8213.279.2460.6568.93
Delta 37.314.8416.0411.6872.9959.30
Delta 46.413.6820.3611.5268.4851.72
Delta 51.820.8115.599.1779.9661.90
Table 6. Runtime (in seconds) of each active learning approach. Each active learning process was terminated at N = 800 .
Table 6. Runtime (in seconds) of each active learning approach. Each active learning process was terminated at N = 800 .
RFBC-RF
WRS-A WRS-B WRS-MSE WRS-A WRS-B WRS-MSE
FMV61.63128.38128.37130.74128.52130.94
Delta 164.00130.33133.53133.32135.07132.94
Delta 257.48119.00122.13122.50120.47119.65
Delta 359.28124.65125.05119.17122.96122.27
Delta 452.71107.24108.37109.23109.72109.74
Delta 552.82101.41102.77104.10103.87104.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gweon, H.; Li, S.; Xu, Y. Use of Prediction Bias in Active Learning and Its Application to Large Variable Annuity Portfolios. Risks 2024, 12, 85. https://doi.org/10.3390/risks12060085

AMA Style

Gweon H, Li S, Xu Y. Use of Prediction Bias in Active Learning and Its Application to Large Variable Annuity Portfolios. Risks. 2024; 12(6):85. https://doi.org/10.3390/risks12060085

Chicago/Turabian Style

Gweon, Hyukjun, Shu Li, and Yangxuan Xu. 2024. "Use of Prediction Bias in Active Learning and Its Application to Large Variable Annuity Portfolios" Risks 12, no. 6: 85. https://doi.org/10.3390/risks12060085

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop