Risks
http://www.mdpi.com/journal/risks
Latest open access articles published in Risks at http://www.mdpi.com/journal/risks<![CDATA[Risks, Vol. 4, Pages 35: A Note on the Impact of Parameter Uncertainty on Barrier Derivatives]]>
http://www.mdpi.com/2227-9091/4/4/35
This paper presents a comprehensive extension of pricing two-dimensional derivatives depending on two barrier constraints. We assume randomness on the covariance matrix as a way of generalizing. We analyse common barrier derivatives, enabling us to study parameter uncertainty and the risk related to the estimation procedure (estimation risk). In particular, we use the distribution of empirical parameters from IBM and EURO STOXX50. The evidence suggests that estimation risk should not be neglected in the context of multidimensional barrier derivatives, as it could cause price differences of up to 70%.Risks2016-09-2944Article10.3390/risks4040035352227-90912016-09-29doi: 10.3390/risks4040035Marcos EscobarSven Panz<![CDATA[Risks, Vol. 4, Pages 34: Sharp Convex Bounds on the Aggregate Sums–An Alternative Proof]]>
http://www.mdpi.com/2227-9091/4/4/34
It is well known that a random vector with given marginals is comonotonic if and only if it has the largest convex sum, and that a random vector with given marginals (under an additional condition) is mutually exclusive if and only if it has the minimal convex sum. This paper provides an alternative proof of these two results using the theories of distortion risk measure and expected utility.Risks2016-09-2944Article10.3390/risks4040034342227-90912016-09-29doi: 10.3390/risks4040034Chuancun YinDan Zhu<![CDATA[Risks, Vol. 4, Pages 33: Multivariate TVaR-Based Risk Decomposition for Vector-Valued Portfolios]]>
http://www.mdpi.com/2227-9091/4/4/33
In order to protect stakeholders of insurance companies and financial institutions against adverse outcomes of risky businesses, regulators and senior management use capital allocation techniques. For enterprise-wide risk management, it has become important to calculate the contribution of each risk within a portfolio. For that purpose, bivariate lower and upper orthant tail value-at-risk can be used for capital allocation. In this paper, we present multivariate value-at-risk and tail-value-at-risk for d ≥ 2 , and we focus on three different methods to calculate optimal values for the contribution of each risk within the sums of random vectors to the overall portfolio, which could particularly apply to insurance and financial portfolios.Risks2016-09-2344Article10.3390/risks4040033332227-90912016-09-23doi: 10.3390/risks4040033Mélina MailhotMhamed Mesfioui<![CDATA[Risks, Vol. 4, Pages 32: The Wasserstein Metric and Robustness in Risk Management]]>
http://www.mdpi.com/2227-9091/4/3/32
In the aftermath of the financial crisis, it was realized that the mathematical models used for the valuation of financial instruments and the quantification of risk inherent in portfolios consisting of these financial instruments exhibit a substantial model risk. Consequently, regulators and other stakeholders have started to require that the internal models used by financial institutions are robust. We present an approach to consistently incorporate the robustness requirements into the quantitative risk management process of a financial institution, with a special focus on insurance. We advocate the Wasserstein metric as the canonical metric for approximations in robust risk management and present supporting arguments. Representing risk measures as statistical functionals, we relate risk measures with the concept of robustness and hence continuity with respect to the Wasserstein metric. This allows us to use results from robust statistics concerning continuity and differentiability of functionals. Finally, we illustrate our approach via practical applications.Risks2016-08-3143Article10.3390/risks4030032322227-90912016-08-31doi: 10.3390/risks4030032Rüdiger KieselRobin RühlickeGerhard StahlJinsong Zheng<![CDATA[Risks, Vol. 4, Pages 31: Choosing Markovian Credit Migration Matrices by Nonlinear Optimization]]>
http://www.mdpi.com/2227-9091/4/3/31
Transition matrices, containing credit risk information in the form of ratings based on discrete observations, are published annually by rating agencies. A substantial issue arises, as for higher rating classes practically no defaults are observed yielding default probabilities of zero. This does not always reflect reality. To circumvent this shortcoming, estimation techniques in continuous-time can be applied. However, raw default data may not be available at all or not in the desired granularity, leaving the practitioner to rely on given one-year transition matrices. Then, it becomes necessary to transform the one-year transition matrix to a generator matrix. This is known as the embedding problem and can be formulated as a nonlinear optimization problem, minimizing the distance between the exponential of a potential generator matrix and the annual transition matrix. So far, in credit risk-related literature, solving this problem directly has been avoided, but approximations have been preferred instead. In this paper, we show that this problem can be solved numerically with sufficient accuracy, thus rendering approximations unnecessary. Our direct approach via nonlinear optimization allows one to consider further credit risk-relevant constraints. We demonstrate that it is thus possible to choose a proper generator matrix with additional structural properties.Risks2016-08-3043Article10.3390/risks4030031312227-90912016-08-30doi: 10.3390/risks4030031Maximilian HughesRalf Werner<![CDATA[Risks, Vol. 4, Pages 30: On the Capital Allocation Problem for a New Coherent Risk Measure in Collective Risk Theory]]>
http://www.mdpi.com/2227-9091/4/3/30
In this paper we introduce a new coherent cumulative risk measure on a subclass in the space of càdlàg processes. This new coherent risk measure turns out to be tractable enough within a class of models where the aggregate claims is driven by a spectrally positive Lévy process. We focus our motivation and discussion on the problem of capital allocation. Indeed, this risk measure is well-suited to address the problem of capital allocation in an insurance context. We show that the capital allocation problem for this risk measure has a unique solution determined by the Euler allocation method. Some examples and connections with existing results as well as practical implications are also discussed.Risks2016-08-1643Article10.3390/risks4030030302227-90912016-08-16doi: 10.3390/risks4030030Hirbod AssaManuel MoralesHassan Omidi Firouzi<![CDATA[Risks, Vol. 4, Pages 29: Optimal Insurance with Heterogeneous Beliefs and Disagreement about Zero-Probability Events]]>
http://www.mdpi.com/2227-9091/4/3/29
In problems of optimal insurance design, Arrow’s classical result on the optimality of the deductible indemnity schedule holds in a situation where the insurer is a risk-neutral Expected-Utility (EU) maximizer, the insured is a risk-averse EU-maximizer, and the two parties share the same probabilistic beliefs about the realizations of the underlying insurable loss. Recently, Ghossoub re-examined Arrow’s problem in a setting where the two parties have different subjective beliefs about the realizations of the insurable random loss, and he showed that if these beliefs satisfy a certain compatibility condition that is weaker than the Monotone Likelihood Ratio (MLR) condition, then optimal indemnity schedules exist and are nondecreasing in the loss. However, Ghossoub only gave a characterization of these optimal indemnity schedules in the special case of an MLR. In this paper, we consider the general case, allowing for disagreement about zero-probability events. We fully characterize the class of all optimal indemnity schedules that are nondecreasing in the loss, in terms of their distribution under the insured’s probability measure, and we obtain Arrow’s classical result, as well as one of the results of Ghossoub as corollaries. Finally, we formalize Marshall’s argument that, in a setting of belief heterogeneity, an optimal indemnity schedule may take “any”shape.Risks2016-08-0543Article10.3390/risks4030029292227-90912016-08-05doi: 10.3390/risks4030029Mario Ghossoub<![CDATA[Risks, Vol. 4, Pages 28: Using Climate and Weather Data to Support Regional Vulnerability Screening Assessments of Transportation Infrastructure]]>
http://www.mdpi.com/2227-9091/4/3/28
Extreme weather and climate change can have a significant impact on all types of infrastructure and assets, regardless of location, with the potential for human casualties, physical damage to assets, disruption of operations, economic and community distress, and environmental degradation. This paper describes a methodology for using extreme weather and climate data to identify climate-related risks and to quantify the potential impact of extreme weather events on certain types of transportation infrastructure as part of a vulnerability screening assessment. This screening assessment can be especially useful when a large number of assets or large geographical areas are being studied, with the results enabling planners and asset managers to undertake a more detailed assessment of vulnerability on a more targeted number of assets or locations. The methodology combines climate, weather, and impact data to identify vulnerabilities to a range of weather and climate related risks over a multi-decadal planning period. The paper applies the methodology to perform an extreme weather and climate change vulnerability screening assessment on transportation infrastructure assets for the State of Tennessee. This paper represents the results of one of the first efforts at spatial vulnerability assessments of transportation infrastructure and provides important insights for any organization considering the impact of climate and weather events on transportation or other critical infrastructure systems.Risks2016-08-0343Article10.3390/risks4030028282227-90912016-08-03doi: 10.3390/risks4030028Leah DundonKatherine NelsonJaney CampMark AbkowitzAlan Jones<![CDATA[Risks, Vol. 4, Pages 25: Understanding Reporting Delay in General Insurance]]>
http://www.mdpi.com/2227-9091/4/3/25
The aim of this paper is to understand and to model claims arrival and reporting delay in general insurance. We calibrate two real individual claims data sets to the statistical model of Jewell and Norberg. One data set considers property insurance and the other one casualty insurance. For our analysis we slightly relax the model assumptions of Jewell allowing for non-stationarity so that the model is able to cope with trends and with seasonal patterns. The performance of our individual claims data prediction is compared to the prediction based on aggregate data using the Poisson chain-ladder method.Risks2016-07-0843Article10.3390/risks4030025252227-90912016-07-08doi: 10.3390/risks4030025Richard VerrallMario Wüthrich<![CDATA[Risks, Vol. 4, Pages 27: Lead–Lag Relationship Using a Stop-and-Reverse-MinMax Process]]>
http://www.mdpi.com/2227-9091/4/3/27
The intermarket analysis, in particular the lead–lag relationship, plays an important role within financial markets. Therefore, a mathematical approach to be able to find interrelations between the price development of two different financial instruments is developed in this paper. Computing the differences of the relative positions of relevant local extrema of two charts, i.e., the local phase shifts of these price developments, gives us an empirical distribution on the unit circle. With the aid of directional statistics, such angular distributions are studied for many pairs of markets. It is shown that there are several very strongly correlated financial instruments in the field of foreign exchange, commodities and indexes. In some cases, one of the two markets is significantly ahead with respect to the relevant local extrema, i.e., there is a phase shift unequal to zero between them.Risks2016-07-0743Article10.3390/risks4030027272227-90912016-07-07doi: 10.3390/risks4030027Stanislaus Maier-PaapeAndreas Platen<![CDATA[Risks, Vol. 4, Pages 26: Optimal Reinsurance with Heterogeneous Reference Probabilities]]>
http://www.mdpi.com/2227-9091/4/3/26
This paper studies the problem of optimal reinsurance contract design. We let the insurer use dual utility, and the premium is an extended Wang’s premium principle. The novel contribution is that we allow for heterogeneity in the beliefs regarding the underlying probability distribution. We characterize layer-reinsurance as an optimal reinsurance contract. Moreover, we characterize layer-reinsurance as optimal contracts when the insurer faces costs of holding regulatory capital. We illustrate this in cases where both firms use the Value-at-Risk or the conditional Value-at-Risk.Risks2016-07-0743Article10.3390/risks4030026262227-90912016-07-07doi: 10.3390/risks4030026Tim Boonen<![CDATA[Risks, Vol. 4, Pages 23: Risk Minimization for Insurance Products via F-Doubly Stochastic Markov Chains]]>
http://www.mdpi.com/2227-9091/4/3/23
We study risk-minimization for a large class of insurance contracts. Given that the individual progress in time of visiting an insurance policy’s states follows an F -doubly stochastic Markov chain, we describe different state-dependent types of insurance benefits. These cover single payments at maturity, annuity-type payments and payments at the time of a transition. Based on the intensity of the F -doubly stochastic Markov chain, we provide the Galtchouk-Kunita-Watanabe decomposition for a general insurance contract and specify risk-minimizing strategies in a Brownian financial market setting. The results are further illustrated explicitly within an affine structure for the intensity.Risks2016-07-0743Article10.3390/risks4030023232227-90912016-07-07doi: 10.3390/risks4030023Francesca BiaginiAndreas GrollJan Widenmann<![CDATA[Risks, Vol. 4, Pages 24: Superforecasting: The Art and Science of Prediction. By Philip Tetlock and Dan Gardner]]>
http://www.mdpi.com/2227-9091/4/3/24
Let me say from the outset that this is an excellent book to read. It is not only informative, as it should be for a book on forecasting, but it is highly entertaining.[...]Risks2016-07-0543Book Review10.3390/risks4030024242227-90912016-07-05doi: 10.3390/risks4030024Daniel Buncic<![CDATA[Risks, Vol. 4, Pages 22: A Unified Pricing of Variable Annuity Guarantees under the Optimal Stochastic Control Framework]]>
http://www.mdpi.com/2227-9091/4/3/22
In this paper, we review pricing of the variable annuity living and death guarantees offered to retail investors in many countries. Investors purchase these products to take advantage of market growth and protect savings. We present pricing of these products via an optimal stochastic control framework and review the existing numerical methods. We also discuss pricing under the complete/incomplete financial market models, stochastic mortality and optimal/sub-optimal policyholder behavior, and in the presence of taxes. For numerical valuation of these contracts in the case of simple risky asset process, we develop a direct integration method based on the Gauss-Hermite quadratures with a one-dimensional cubic spline for calculation of the expected contract value, and a bi-cubic spline interpolation for applying the jump conditions across the contract cashflow event times. This method is easier to implement and faster when compared to the partial differential equation methods if the transition density (or its moments) of the risky asset underlying the contract is known in closed form between the event times. We present accurate numerical results for pricing of a Guaranteed Minimum Accumulation Benefit (GMAB) guarantee available on the market that can serve as a numerical benchmark for practitioners and researchers developing pricing of variable annuity guarantees to assess the accuracy of their numerical implementation.Risks2016-07-0543Article10.3390/risks4030022222227-90912016-07-05doi: 10.3390/risks4030022Pavel ShevchenkoXiaolin Luo<![CDATA[Risks, Vol. 4, Pages 21: The Myth of Methuselah and the Uncertainty of Death: The Mortality Fan Charts]]>
http://www.mdpi.com/2227-9091/4/3/21
This paper uses mortality fan charts to illustrate prospective future male mortality. These fan charts show both the most likely path of male mortality and the bands of uncertainty surrounding that path. The fan charts are based on a model of male mortality that is known to provide a good fit to UK mortality data. The fan charts suggest that there are clear limits to longevity—that future mortality rates are very uncertain and tend to become more uncertain the further ahead the forecast—and that forecasts of future mortality uncertainty must also take account of uncertainty in the parameters of the underlying mortality model.Risks2016-07-0443Article10.3390/risks4030021212227-90912016-07-04doi: 10.3390/risks4030021Kevin DowdDavid BlakeAndrew Cairns<![CDATA[Risks, Vol. 4, Pages 20: Survey on Log-Normally Distributed Market-Technical Trend Data]]>
http://www.mdpi.com/2227-9091/4/3/20
In this survey, a short introduction of the recent discovery of log-normally-distributed market-technical trend data will be given. The results of the statistical evaluation of typical market-technical trend variables will be presented. It will be shown that the log-normal assumption fits better to empirical trend data than to daily returns of stock prices. This enables one to mathematically evaluate trading systems depending on such variables. In this manner, a basic approach to an anti-cyclic trading system will be given as an example.Risks2016-07-0443Article10.3390/risks4030020202227-90912016-07-04doi: 10.3390/risks4030020René BrennerStanislaus Maier-Paape<![CDATA[Risks, Vol. 4, Pages 19: An Optimal Turkish Private Pension Plan with a Guarantee Feature]]>
http://www.mdpi.com/2227-9091/4/3/19
The Turkish Private Pension System is an investment system which aims to generate income for future consumption. This is a volunteer system, and the contributions are held in individual portfolios. Therefore, management of the funds is an important issue for both the participants and the insurance company. In this study, we propose an optimal private pension plan with a guarantee feature that is based on Constant Proportion Portfolio Insurance (CPPI). We derive a closed form formula for the optimal strategy with the help of dynamic programming. Moreover, our model is evaluated with numerical examples, and we compare its performance by implementing a sensitivity analysis.Risks2016-06-2743Article10.3390/risks4030019192227-90912016-06-27doi: 10.3390/risks4030019Ayşegül İşcanog̃lu-Çekiç<![CDATA[Risks, Vol. 4, Pages 18: Consistent Re-Calibration of the Discrete-Time Multifactor Vasiček Model]]>
http://www.mdpi.com/2227-9091/4/3/18
The discrete-time multifactor Vasiček model is a tractable Gaussian spot rate model. Typically, two- or three-factor versions allow one to capture the dependence structure between yields with different times to maturity in an appropriate way. In practice, re-calibration of the model to the prevailing market conditions leads to model parameters that change over time. Therefore, the model parameters should be understood as being time-dependent or even stochastic. Following the consistent re-calibration (CRC) approach, we construct models as concatenations of yield curve increments of Hull–White extended multifactor Vasiček models with different parameters. The CRC approach provides attractive tractable models that preserve the no-arbitrage premise. As a numerical example, we fit Swiss interest rates using CRC multifactor Vasiček models.Risks2016-06-2343Article10.3390/risks4030018182227-90912016-06-23doi: 10.3390/risks4030018Philipp HarmsDavid StefanovitsJosef TeichmannMario Wüthrich<![CDATA[Risks, Vol. 4, Pages 17: Ruin Probabilities with Dependence on the Number of Claims within a Fixed Time Window]]>
http://www.mdpi.com/2227-9091/4/2/17
We analyse the ruin probabilities for a renewal insurance risk process with inter-arrival times depending on the claims that arrive within a fixed (past) time window. This dependence could be explained through a regenerative structure. The main inspiration of the model comes from the bonus-malus (BM) feature of pricing car insurance. We discuss first the asymptotic results of ruin probabilities for different regimes of claim distributions. For numerical results, we recognise an embedded Markov additive process, and via an appropriate change of measure, ruin probabilities could be computed to a closed-form formulae. Additionally, we employ the importance sampling simulations to derive ruin probabilities, which further permit an in-depth analysis of a few concrete cases.Risks2016-06-1542Article10.3390/risks4020017172227-90912016-06-15doi: 10.3390/risks4020017Corina ConstantinescuSuhang DaiWeihong NiZbigniew Palmowski<![CDATA[Risks, Vol. 4, Pages 16: Spouses’ Dependence across Generations and Pricing Impact on Reversionary Annuities]]>
http://www.mdpi.com/2227-9091/4/2/16
This paper studies the dependence between coupled lives, i.e., the spouses’ dependence, across different generations, and its effects on prices of reversionary annuities in the presence of longevity risk. Longevity risk is represented via a stochastic mortality intensity. We find that a generation-based model is important, since spouses’ dependence decreases when passing from older generations to younger generations. The independence assumption produces quantifiable mispricing of reversionary annuities, with different effects on different generations. The research is conducted using a well-known dataset of double life contracts.Risks2016-05-2542Article10.3390/risks4020016162227-90912016-05-25doi: 10.3390/risks4020016Elisa LucianoJaap SpreeuwElena Vigna<![CDATA[Risks, Vol. 4, Pages 15: Improving Convergence of Binomial Schemes and the Edgeworth Expansion]]>
http://www.mdpi.com/2227-9091/4/2/15
Binomial trees are very popular in both theory and applications of option pricing. As they often suffer from an irregular convergence behavior, improving this is an important task. We build upon a new version of the Edgeworth expansion for lattice models to construct new and quickly converging binomial schemes with a particular application to barrier options.Risks2016-05-2342Article10.3390/risks4020015152227-90912016-05-23doi: 10.3390/risks4020015Alona BockRalf Korn<![CDATA[Risks, Vol. 4, Pages 14: Estimating Quantile Families of Loss Distributions for Non-Life Insurance Modelling via L-Moments]]>
http://www.mdpi.com/2227-9091/4/2/14
This paper discusses different classes of loss models in non-life insurance settings. It then overviews the class of Tukey transform loss models that have not yet been widely considered in non-life insurance modelling, but offer opportunities to produce flexible skewness and kurtosis features often required in loss modelling. In addition, these loss models admit explicit quantile specifications which make them directly relevant for quantile based risk measure calculations. We detail various parameterisations and sub-families of the Tukey transform based models, such as the g-and-h, g-and-k and g-and-j models, including their properties of relevance to loss modelling. One of the challenges that are amenable to practitioners when fitting such models is to perform robust estimation of the model parameters. In this paper we develop a novel, efficient, and robust procedure for estimating the parameters of this family of Tukey transform models, based on L-moments. It is shown to be more efficient than the current state of the art estimation methods for such families of loss models while being simple to implement for practical purposes.Risks2016-05-2042Article10.3390/risks4020014142227-90912016-05-20doi: 10.3390/risks4020014Gareth PetersWilson ChenRichard Gerlach<![CDATA[Risks, Vol. 4, Pages 12: Macro vs. Micro Methods in Non-Life Claims Reserving (an Econometric Perspective)]]>
http://www.mdpi.com/2227-9091/4/2/12
Traditionally, actuaries have used run-off triangles to estimate reserve (“macro” models, on aggregated data). However, it is possible to model payments related to individual claims. If those models provide similar estimations, we investigate uncertainty related to reserves with “macro” and “micro” models. We study theoretical properties of econometric models (Gaussian, Poisson and quasi-Poisson) on individual data, and clustered data. Finally, applications in claims reserving are considered.Risks2016-05-1442Article10.3390/risks4020012122227-90912016-05-14doi: 10.3390/risks4020012Arthur CharpentierMathieu Pigeon<![CDATA[Risks, Vol. 4, Pages 13: Community Analysis of Global Financial Markets]]>
http://www.mdpi.com/2227-9091/4/2/13
We analyze the daily returns of stock market indices and currencies of 56 countries over the period of 2002–2012. We build a network model consisting of two layers, one being the stock market indices and the other the foreign exchange markets. Synchronous and lagged correlations are used as measures of connectivity and causality among different parts of the global economic system for two different time intervals: non-crisis (2002–2006) and crisis (2007–2012) periods. We study community formations within the network to understand the influences and vulnerabilities of specific countries or groups of countries. We observe different behavior of the cross correlations and communities for crisis vs. non-crisis periods. For example, the overall correlation of stock markets increases during crisis while the overall correlation in the foreign exchange market and the correlation between stock and foreign exchange markets decrease, which leads to different community structures. We observe that the euro, while being central during the relatively calm period, loses its dominant role during crisis. Furthermore we discover that the troubled Eurozone countries, Portugal, Italy, Greece and Spain, form their own cluster during the crisis period.Risks2016-05-1342Article10.3390/risks4020013132227-90912016-05-13doi: 10.3390/risks4020013Irena VodenskaAlexander BeckerDi ZhouDror KenettH. StanleyShlomo Havlin<![CDATA[Risks, Vol. 4, Pages 11: Participating Life Insurance Products with Alternative Guarantees: Reconciling Policyholders’ and Insurers’ Interests]]>
http://www.mdpi.com/2227-9091/4/2/11
Traditional participating life insurance contracts with year-to-year (cliquet-style) guarantees have come under pressure in the current situation of low interest rates and volatile capital markets, in particular when priced in a market-consistent valuation framework. In addition, such guarantees lead to rather high capital requirements under risk-based solvency frameworks such as Solvency II or the Swiss Solvency Test (SST). Therefore, insurers in several countries have developed new forms of participating products with alternative (typically weaker and/or lower) guarantees that are less risky for the insurer. In a previous paper, it has been shown that such alternative product designs can lead to higher capital efficiency, i.e., higher and more stable profits and reduced capital requirements. As a result, the financial risk for the insurer is significantly reduced while the main guarantee features perceived and requested by the policyholder are preserved. Based on these findings, this paper now combines the insurer’s and the policyholder’s perspective by analyzing product versions that compensate policyholders for the less valuable guarantees. We particularly identify combinations of asset allocation and profit participation rate for the different product designs that lead to an identical expected profit for the insurer (and identical risk-neutral value for the policyholder), but differ with respect to the insurer’s risk and solvency capital requirements as well as with respect to the real-world return distribution for the policyholder. We show that alternative products can be designed in a way that the insurer’s expected profitability remains unchanged, the insurer’s risk and hence capital requirement is substantially reduced and the policyholder’s expected return is increased. This illustrates that such products might be able to reconcile insurers’ and policyholders’ interests and serve as an alternative to the rather risky cliquet-style products.Risks2016-05-0542Article10.3390/risks4020011112227-90912016-05-05doi: 10.3390/risks4020011Andreas ReußJochen RußJochen Wieland<![CDATA[Risks, Vol. 4, Pages 10: Telematics and Gender Discrimination: Some Usage-Based Evidence on Whether Men’s Risk of Accidents Differs from Women’s]]>
http://www.mdpi.com/2227-9091/4/2/10
Pay-as-you-drive (PAYD), or usage-based automobile insurance (UBI), is a policy agreement tied to vehicle usage. In this paper we analyze the effect of the distance traveled on the risk of accidents among young drivers with a PAYD policy. We use regression models for survival data to estimate how long it takes them to have their first accident at fault during the coverage period. Our empirical application with real data is presented and shows that gender differences are mainly attributable to the intensity of use. Indeed, although gender has a significant effect in explaining the time to the first crash, this effect is no longer significant when the average distance traveled per day is introduced in the model. This suggests that gender differences in the risk of accidents are, to a large extent, attributable to the fact that men drive more often than women. Estimates of the time to the first accident for different driver risk types are presented. We conclude that no gender discrimination is necessary if telematics provides enough information on driving habits.Risks2016-04-0842Article10.3390/risks4020010102227-90912016-04-08doi: 10.3390/risks4020010Mercedes AyusoMontserrat GuillenAna Pérez-Marín<![CDATA[Risks, Vol. 4, Pages 9: Inflation Protected Investment Strategies]]>
http://www.mdpi.com/2227-9091/4/2/9
In this paper, a dynamic inflation-protected investment strategy is presented, which is based on traditional asset classes and Markov-switching models. Different stock market, as well as inflation regimes are identified, and within those regimes, the inflation hedging potential of stocks, bonds, real estate, commodities and gold are investigated. Within each regime, we determine optimal investment portfolios driven by the investment idea of protection from losses due to changing inflation if inflation is rising or high, but decoupling the performance from inflation if inflation is low. The results clearly indicate that these asset classes behave differently in different stock market and inflation regimes. Whereas in the long-run, we agree with the general opinion in the literature that stocks and bonds are a suitable hedge against inflation, we observe for short time horizons that the hedging potential of each asset class, especially of real estate and commodities, depend strongly on the state of the current market environment. Thus, our approach provides a possible explanation for different statements in the literature regarding the inflation hedging properties of these asset classes. A dynamic inflation-protected investment strategy is developed, which combines inflation protection and upside potential. This strategy outperforms standard buy-and-hold strategies, as well as the well-known 1 N -portfolio.Risks2016-03-2842Article10.3390/risks402000992227-90912016-03-28doi: 10.3390/risks4020009Mirco MahlstedtRudi Zagst<![CDATA[Risks, Vol. 4, Pages 8: Optimal Insurance for a Minimal Expected Retention: The Case of an Ambiguity-Seeking Insurer]]>
http://www.mdpi.com/2227-9091/4/1/8
In the classical expected utility framework, a problem of optimal insurance design with a premium constraint is equivalent to a problem of optimal insurance design with a minimum expected retention constraint. When the insurer has ambiguous beliefs represented by a non-additive probability measure, as in Schmeidler, this equivalence no longer holds. Recently, Amarante, Ghossoub and Phelps examined the problem of optimal insurance design with a premium constraint when the insurer has ambiguous beliefs. In particular, they showed that when the insurer is ambiguity-seeking, with a concave distortion of the insured’s probability measure, then the optimal indemnity schedule is a state-contingent deductible schedule, in which the deductible depends on the state of the world only through the insurer’s distortion function. In this paper, we examine the problem of optimal insurance design with a minimum expected retention constraint, in the case where the insurer is ambiguity-seeking. We obtain the aforementioned result of Amarante, Ghossoub and Phelps and the classical result of Arrow as special cases.Risks2016-03-2141Article10.3390/risks401000882227-90912016-03-21doi: 10.3390/risks4010008Massimiliano AmaranteMario Ghossoub<![CDATA[Risks, Vol. 4, Pages 7: Nonlinear Time Series and Neural-Network Models of Exchange Rates between the US Dollar and Major Currencies]]>
http://www.mdpi.com/2227-9091/4/1/7
This paper features an analysis of major currency exchange rate movements in relation to the US dollar, as constituted in US dollar terms. Euro, British pound, Chinese yuan, and Japanese yen are modelled using a variety of non-linear models, including smooth transition regression models, logistic smooth transition regressions models, threshold autoregressive models, nonlinear autoregressive models, and additive nonlinear autoregressive models, plus Neural Network models. The models are evaluated on the basis of error metrics for twenty day out-of-sample forecasts using the mean average percentage errors (MAPE). The results suggest that there is no dominating class of time series models, and the different currency pairs relationships with the US dollar are captured best by neural net regression models, over the ten year sample of daily exchange rate returns data, from August 2005 to August 2015.Risks2016-03-1641Article10.3390/risks401000772227-90912016-03-16doi: 10.3390/risks4010007David AllenMichael McAleerShelton PeirisAbhay Singh<![CDATA[Risks, Vol. 4, Pages 6: Analysis of Insurance Claim Settlement Process with Markovian Arrival Processes]]>
http://www.mdpi.com/2227-9091/4/1/6
This paper proposes a model for the claim occurrence, reporting, and handling process of insurance companies. It is assumed that insurance claims occur according to a Markovian arrival process. An incurred claim goes through some stages of a claim reporting and handling process, such as Incurred But Not Reported (IBNR), Reported But Not Settled (RBNS) and Settled (S). We derive formulas for the joint distribution and the joint moments for the amount of INBR, RBNS and Settled claims. This model generalizes previous ones in the literature, which generally assume Poisson claim arrivals. Due to the flexibility of the Markovian arrival process, the model can be used to evaluate how the claim occurring, reporting, and handling mechanisms may affect the volatilities of the amount of IBNR, RBNS and Settled claims, and the interdependencies among them.Risks2016-03-1141Article10.3390/risks401000662227-90912016-03-11doi: 10.3390/risks4010006Jiandong Ren<![CDATA[Risks, Vol. 4, Pages 5: High-Frequency Financial Econometrics]]>
http://www.mdpi.com/2227-9091/4/1/5
This book is fundamentally about the estimation of risk.[...]Risks2016-02-2641Book Review10.3390/risks401000552227-90912016-02-26doi: 10.3390/risks4010005Harley Thompson<![CDATA[Risks, Vol. 4, Pages 4: Multivariate Frequency-Severity Regression Models in Insurance]]>
http://www.mdpi.com/2227-9091/4/1/4
In insurance and related industries including healthcare, it is common to have several outcome measures that the analyst wishes to understand using explanatory variables. For example, in automobile insurance, an accident may result in payments for damage to one’s own vehicle, damage to another party’s vehicle, or personal injury. It is also common to be interested in the frequency of accidents in addition to the severity of the claim amounts. This paper synthesizes and extends the literature on multivariate frequency-severity regression modeling with a focus on insurance industry applications. Regression models for understanding the distribution of each outcome continue to be developed yet there now exists a solid body of literature for the marginal outcomes. This paper contributes to this body of literature by focusing on the use of a copula for modeling the dependence among these outcomes; a major advantage of this tool is that it preserves the body of work established for marginal models. We illustrate this approach using data from the Wisconsin Local Government Property Insurance Fund. This fund offers insurance protection for (i) property; (ii) motor vehicle; and (iii) contractors’ equipment claims. In addition to several claim types and frequency-severity components, outcomes can be further categorized by time and space, requiring complex dependency modeling. We find significant dependencies for these data; specifically, we find that dependencies among lines are stronger than the dependencies between the frequency and average severity within each line.Risks2016-02-2541Article10.3390/risks401000442227-90912016-02-25doi: 10.3390/risks4010004Edward FreesGee LeeLu Yang<![CDATA[Risks, Vol. 4, Pages 3: Premiums for Long-Term Care Insurance Packages: Sensitivity with Respect to Biometric Assumptions]]>
http://www.mdpi.com/2227-9091/4/1/3
Long-term care insurance (LTCI) covers are rather recent products, in the framework of health insurance. It follows that specific biometric data are scanty; pricing and reserving problems then arise because of difficulties in the choice of appropriate technical bases. Different benefit structures imply different sensitivity degrees with respect to changes in biometric assumptions. Hence, an accurate sensitivity analysis can help in designing LTCI products and, in particular, in comparing stand-alone products to combined products, i.e., packages including LTCI benefits and other lifetime-related benefits. Numerical examples show, in particular, that the stand-alone cover is much riskier than all of the LTCI combined products that we have considered. As a consequence, the LTCI stand-alone cover is a highly “absorbing” product as regards capital requirements for solvency purposes.Risks2016-02-2241Article10.3390/risks401000332227-90912016-02-22doi: 10.3390/risks4010003Ermanno Pitacco<![CDATA[Risks, Vol. 4, Pages 2: Ruin Analysis of a Discrete-Time Dependent Sparre Andersen Model with External Financial Activities and Randomized Dividends]]>
http://www.mdpi.com/2227-9091/4/1/2
We consider a discrete-time dependent Sparre Andersen risk model which incorporates multiple threshold levels characterizing an insurer’s minimal capital requirement, dividend paying situations, and external financial activities. We focus on the development of a recursive computational procedure to calculate the finite-time ruin probabilities and expected total discounted dividends paid prior to ruin associated with this model. We investigate several numerical examples and make some observations concerning the impact our threshold levels have on the finite-time ruin probabilities and expected total discounted dividends paid prior to ruin.Risks2016-02-0341Article10.3390/risks401000222227-90912016-02-03doi: 10.3390/risks4010002Sung KimSteve Drekic<![CDATA[Risks, Vol. 4, Pages 1: Acknowledgement to Reviewers of Risks in 2015]]>
http://www.mdpi.com/2227-9091/4/1/1
The editors of Risks would like to express their sincere gratitude to the following reviewers for assessing manuscripts in 2015. [...]Risks2016-01-2141Editorial10.3390/risks401000112227-90912016-01-21doi: 10.3390/risks4010001 Risks Editorial Office<![CDATA[Risks, Vol. 3, Pages 624-646: Modified Munich Chain-Ladder Method]]>
http://www.mdpi.com/2227-9091/3/4/624
The Munich chain-ladder method for claims reserving was introduced by Quarg and Mack on an axiomatic basis. We analyze these axioms, and we define a modified Munich chain-ladder method which is based on an explicit stochastic model. This stochastic model then allows us to consider claims prediction and prediction uncertainty for the Munich chain-ladder method in a consistent way.Risks2015-12-2134Article10.3390/risks30406246246462227-90912015-12-21doi: 10.3390/risks3040624Michael MerzMario Wüthrich<![CDATA[Risks, Vol. 3, Pages 599-623: Dependence Uncertainty Bounds for the Expectile of a Portfolio]]>
http://www.mdpi.com/2227-9091/3/4/599
We study upper and lower bounds on the expectile risk measure of risky portfolios when the joint distribution of the risky components is not fully specified. First, we summarize methods for obtaining bounds when only the marginal distributions of the components are known, but not their interdependence (unconstrained bounds). In particular, we provide the best-possible upper bound and the best-possible lower bound (under some conditions), as well as numerical procedures to compute them. We also derive simple analytic bounds that appear adequate in various situations of interest. Second, we study bounds when some information on interdependence is available (constrained bounds). When the variance of the portfolio is known, a simple-to-compute upper bound is provided, and we illustrate that it may significantly improve the unconstrained upper bound. We also show that the unconstrained lower bound cannot be readily improved using variance information. Next, we derive improved bounds when the bivariate distributions of each of the risky components and a risk factor are known. When the factor induces a positive dependence among the components, it is typically possible to improve the unconstrained lower bound. Finally, the unconstrained dependence uncertainty spreads of expected shortfall, value-at-risk and the expectile are compared.Risks2015-12-1034Article10.3390/risks30405995996232227-90912015-12-10doi: 10.3390/risks3040599Edgars JakobsonsSteven Vanduffel<![CDATA[Risks, Vol. 3, Pages 573-598: Information-Based Trade in German Real Estate and Equity Markets]]>
http://www.mdpi.com/2227-9091/3/4/573
This paper employs four established market microstructure measures on information-based trade in financial markets. A set of German mid and small caps is used to analyze potential differential information content in real estate stocks compared to other asset classes. After linking substantially lower amounts of information-based trade in real estate stocks to higher liquidity premia, it is found that the evolution of the information content in real estate and other assets follows similar trends. Consequently, interdependence is tested for rolling time windows, revealing strong informational links between real estate and other assets. Particularly, small caps, financials, as well as companies offering consumer goods and services show a close relationship to real estate. Depending on the choice of the measure of information-based trade, up to 75% of the variation in the information content in real estate shares is related to other asset classes, pointing to the notion of high dependence.Risks2015-12-0734Article10.3390/risks30405735735982227-90912015-12-07doi: 10.3390/risks3040573Marco Wölfle<![CDATA[Risks, Vol. 3, Pages 553-572: Stochastic Optimal Control for Online Seller under Reputational Mechanisms]]>
http://www.mdpi.com/2227-9091/3/4/553
In this work we propose and analyze a model which addresses the pulsing behavior of sellers in an online auction (store). This pulsing behavior is observed when sellers switch between advertising and processing states. We assert that a seller switches her state in order to maximize her profit, and further that this switch can be identified through the seller’s reputation. We show that for each seller there is an optimal reputation, i.e., the reputation at which the seller should switch her state in order to maximize her total profit. We design a stochastic behavioral model for an online seller, which incorporates the dynamics of resource allocation and reputation. The design of the model is optimized by using a stochastic advertising model from [1] and used effectively in the Stochastic Optimal Control of Advertising [2]. This model of reputation is combined with the effect of online reputation on sales price empirically verified in [3]. We derive the Hamilton-Jacobi-Bellman (HJB) differential equation, whose solution relates optimal wealth level to a seller’s reputation. We formulate both a full model, as well as a reduced model with fewer parameters, both of which have the same qualitative description of the optimal seller behavior. Coincidentally, the reduced model has a closed form analytical solution that we construct.Risks2015-12-0434Article10.3390/risks30405535535722227-90912015-12-04doi: 10.3390/risks3040553Milan BradonjićMatthew CausleyAlbert Cohen<![CDATA[Risks, Vol. 3, Pages 543-552: Production Flexibility and Hedging]]>
http://www.mdpi.com/2227-9091/3/4/543
We extend the analysis on hedging with price and output uncertainty by endogenizing the output decision. Specifically, we consider the joint determination of output and hedging in the case of flexibility in production. We show that the risk-averse firm always maintains a short position in the futures market when the futures price is actuarially fair. Moreover, in the context of an example, we show that the presence of production flexibility reduces the incentive to hedge for all risk averse agents.Risks2015-12-0434Article10.3390/risks30405435435522227-90912015-12-04doi: 10.3390/risks3040543Georges DionneMarc Santugini<![CDATA[Risks, Vol. 3, Pages 515-542: The Impact of Guarantees on the Performance of Pension Saving Schemes: Insights from the Literature]]>
http://www.mdpi.com/2227-9091/3/4/515
Guarantees are often seen as the key characteristics of pension saving products, but securing them can become costly and is of central relevance especially in the course of the current low interest rate environment. In this article, we deal with the question of how costly the typical types of guarantees are, in the sense that they reduce a pension saving scheme’s financial performance over time. In this context, we aim to provide a presentation of insights from selected literature studying the impact of point-to-point guarantees and cliquet-style interest rate guarantees on the performance of pension contracts. The comparative analysis emphasizes that, in most cases, guarantee costs are not negligible with regard to a contract’s financial performance, especially compared to benchmarks, and that customers knowingly opt for such guarantees (or not) is, thus, indispensable. To further investigate the willingness-to-pay for guarantees in life insurance is an area for future research, in particular for innovative contract design.Risks2015-11-2034Article10.3390/risks30405155155422227-90912015-11-20doi: 10.3390/risks3040515Alexander Bohnert<![CDATA[Risks, Vol. 3, Pages 491-514: On the Joint Analysis of the Total Discounted Payments to Policyholders and Shareholders: Dividend Barrier Strategy]]>
http://www.mdpi.com/2227-9091/3/4/491
In the compound Poisson insurance risk model under a dividend barrier strategy, this paper aims to analyze jointly the aggregate discounted claim amounts until ruin and the total discounted dividends until ruin, which represent the insurer’s payments to its policyholders and shareholders, respectively. To this end, we introduce a Gerber–Shiu-type function, which further incorporates the higher moments of these two quantities. This not only unifies the individual study of various ruin-related quantities, but also allows for new measures concerning covariances to be calculated. The integro-differential equation satisfied by the generalized Gerber–Shiu function and the boundary condition are derived. In particular, when the claim severity is distributed as a combination of exponentials, explicit expressions for this Gerber–Shiu function in some special cases are given. Numerical examples involving the covariances between any two of (i) the aggregate discounted claims until ruin, (ii) the discounted dividend payments until ruin and (iii) the time of ruin are presented along with some interpretations.Risks2015-11-1034Article10.3390/risks30404914915142227-90912015-11-10doi: 10.3390/risks3040491Eric CheungHaibo LiuJae-Kyung Woo<![CDATA[Risks, Vol. 3, Pages 474-490: Combining Alphas via Bounded Regression]]>
http://www.mdpi.com/2227-9091/3/4/474
We give an explicit algorithm and source code for combining alpha streams via bounded regression. In practical applications, typically, there is insufficient history to compute a sample covariance matrix (SCM) for a large number of alphas. To compute alpha allocation weights, one then resorts to (weighted) regression over SCM principal components. Regression often produces alpha weights with insufficient diversification and/or skewed distribution against, e.g., turnover. This can be rectified by imposing bounds on alpha weights within the regression procedure. Bounded regression can also be applied to stock and other asset portfolio construction. We discuss illustrative examples.Risks2015-11-0434Article10.3390/risks30404744744902227-90912015-11-04doi: 10.3390/risks3040474Zura Kakushadze<![CDATA[Risks, Vol. 3, Pages 455-473: Hidden Markov Model for Stock Selection]]>
http://www.mdpi.com/2227-9091/3/4/455
The hidden Markov model (HMM) is typically used to predict the hidden regimes of observation data. Therefore, this model finds applications in many different areas, such as speech recognition systems, computational molecular biology and financial market predictions. In this paper, we use HMM for stock selection. We first use HMM to make monthly regime predictions for the four macroeconomic variables: inflation (consumer price index (CPI)), industrial production index (INDPRO), stock market index (S&amp;P 500) and market volatility (VIX). At the end of each month, we calibrate HMM’s parameters for each of these economic variables and predict its regimes for the next month. We then look back into historical data to find the time periods for which the four variables had similar regimes with the forecasted regimes. Within those similar periods, we analyze all of the S&amp;P 500 stocks to identify which stock characteristics have been well rewarded during the time periods and assign scores and corresponding weights for each of the stock characteristics. A composite score of each stock is calculated based on the scores and weights of its features. Based on this algorithm, we choose the 50 top ranking stocks to buy. We compare the performances of the portfolio with the benchmark index, S&amp;P 500. With an initial investment of $100 in December 1999, over 15 years, in December 2014, our portfolio had an average gain per annum of 14.9% versus 2.3% for the S&amp;P 500.Risks2015-10-2934Article10.3390/risks30404554554732227-90912015-10-29doi: 10.3390/risks3040455Nguyet NguyenDung Nguyen<![CDATA[Risks, Vol. 3, Pages 445-454: Risk Classification Efficiency and the Insurance Market Regulation]]>
http://www.mdpi.com/2227-9091/3/4/445
Given that the insurance market is characterized by asymmetric information, its efficiency has traditionally been based to a large extent on risk classification. In certain regulations, however, we can find restrictions on these differentiations, primarily the ban on those considered to be “discriminatory”. In 2011, following the European Union Directive 2004/113/EC, the European Court of Justice concluded that any gender-based discrimination was prohibited, meaning that gender equality in the European Union had to be ensured from 21 December 2012. Another restriction was imposed by EU and national competition regulation on the exchange of information considered as anti-competitive behavior. This paper aims to contribute to the recent policy debate in the EU, evaluating the negative economic consequences of these regulatory restrictions in terms of market efficiency.Risks2015-09-2534Article10.3390/risks30404454454542227-90912015-09-25doi: 10.3390/risks3040445Donatella Porrini<![CDATA[Risks, Vol. 3, Pages 420-444: The Financial Stress Index: Identification of Systemic Risk Conditions]]>
http://www.mdpi.com/2227-9091/3/3/420
This paper develops a financial stress measure for the United States, the Cleveland Financial Stress Index (CFSI). The index is based on publicly available data describing a six-market partition of the financial system comprising credit, funding, real estate, securitization, foreign exchange, and equity markets. This paper improves upon existing stress measures by objectively selecting between several index weighting methodologies across a variety of monitoring frequencies through comparison against a volatility-based benchmark series. The resulting measure facilitates the decomposition of stress to identify disruptions in specific markets and provides insight into historical stress regimes.Risks2015-09-1633Article10.3390/risks30304204204442227-90912015-09-16doi: 10.3390/risks3030420Mikhail OetJohn DooleyStephen Ong<![CDATA[Risks, Vol. 3, Pages 390-419: Multi-Objective Stochastic Optimization Programs for a Non-Life Insurance Company under Solvency Constraints]]>
http://www.mdpi.com/2227-9091/3/3/390
In the paper, we introduce a multi-objective scenario-based optimization approach for chance-constrained portfolio selection problems. More specifically, a modified version of the normal constraint method is implemented with a global solver in order to generate a dotted approximation of the Pareto frontier for bi- and tri-objective programming problems. Numerical experiments are carried out on a set of portfolios to be optimized for an EU-based non-life insurance company. Both performance indicators and risk measures are managed as objectives. Results show that this procedure is effective and readily applicable to achieve suitable risk-reward tradeoff analysis.Risks2015-09-1533Article10.3390/risks30303903904192227-90912015-09-15doi: 10.3390/risks3030390Massimiliano KaucicRoberto Daris<![CDATA[Risks, Vol. 3, Pages 365-389: Supervising System Stress in Multiple Markets]]>
http://www.mdpi.com/2227-9091/3/3/365
This paper develops an extended financial stress measure that considers the supervisory objective of identifying risks to the stability of the financial system. The measure provides a continuous and bounded signal of financial stress using daily public market data. Broad coverage of material financial system markets over time is achieved by leveraging dynamic credit weights. We consider how this measure can be used to monitor, analyze, and alert financial system stress.Risks2015-09-1433Article10.3390/risks30303653653892227-90912015-09-14doi: 10.3390/risks3030365Mikhail OetJohn DooleyAmanda JanoskoDieter GramlichStephen Ong<![CDATA[Risks, Vol. 3, Pages 338-364: Valuation of Index-Linked Cash Flows in a Heath–Jarrow–Morton Framework]]>
http://www.mdpi.com/2227-9091/3/3/338
In this paper, we study the valuation of stochastic cash flows that exhibit dependence on interest rates. We focus on insurance liability cash flows linked to an index, such as a consumer price index or wage index, where changes in the index value can be partially understood in terms of changes in the term structure of interest rates. Insurance liability cash flows that are not explicitly linked to an index may still be valued in our framework by interpreting index returns as so-called claims inflation, i.e., an increase in claims cost per sold insurance contract. We focus primarily on the case when a deep and liquid market for index-linked contracts is absent or when the market price data are unreliable. Firstly, we present an approach for assigning a monetary value to a stochastic cash flow that does not require full knowledge of the joint dynamics of the cash flow and the term structure of interest rates. Secondly, we investigate in detail model selection, estimation and validation in a Heath–Jarrow–Morton framework. Finally, we analyze the effects of model uncertainty on the valuation of the cash flows and how forecasts of cash flows and interest rates translate into model parameters and affect the valuation.Risks2015-09-1033Article10.3390/risks30303383383642227-90912015-09-10doi: 10.3390/risks3030338Jonas AlmFilip Lindskog<![CDATA[Risks, Vol. 3, Pages 318-337: Delivering Left-Skewed Portfolio Payoff Distributions in the Presence of Transaction Costs]]>
http://www.mdpi.com/2227-9091/3/3/318
For pension-savers, a low payoff is a financial disaster. Such investors will most likely prefer left-skewed payoff distributions over right-skewed payoff distributions. We explore how such distributions can be delivered. Cautious-relaxed utility measures are cautious in ensuring that payoffs don’t fall much below a reference value, but relaxed about exceeding it. We find that the payoff distribution delivered by a cautious-relaxed utility measure has appealing features which payoff distributions delivered by traditional utility functions don’t. In particular, cautious-relaxed distributions can have the mass concentrated on the left, hence be left-skewed. However, cautious-relaxed strategies prescribe frequent portfolio adjustments which may be expensive if transaction costs are charged. In contrast, more traditional strategies can be time-invariant. Thus we investigate the impact of transaction costs on the appeal of cautious-relaxed strategies. We find that relatively high transaction fees are required for the cautious-relaxed strategy to lose its appeal. This paper contributes to the literature which compares utility measures by the payoff distributions they produce and finds that a cautious-relaxed utility measure will deliver payoffs that many investors will prefer.Risks2015-08-2133Article10.3390/risks30303183183372227-90912015-08-21doi: 10.3390/risks3030318Jacek Krawczyk<![CDATA[Risks, Vol. 3, Pages 290-317: Life Insurance Cash Flows with Policyholder Behavior]]>
http://www.mdpi.com/2227-9091/3/3/290
The problem of the valuation of life insurance payments with policyholder behavior is studied. First, a simple survival model is considered, and it is shown how cash flows without policyholder behavior can be modified to include surrender and free policy behavior by calculation of simple integrals. In the second part, a more general disability model with recovery is studied. Here, cash flows are determined by solving a modified Kolmogorov forward differential equation. We conclude the paper with numerical examples illustrating the methods proposed and the impact of policyholder behavior.Risks2015-07-2433Article10.3390/risks30302902903172227-90912015-07-24doi: 10.3390/risks3030290Kristian BuchardtThomas Møller<![CDATA[Risks, Vol. 3, Pages 277-289: Monopolistic Insurance and the Value of Information]]>
http://www.mdpi.com/2227-9091/3/3/277
The value of information regarding risk class for a monopoly insurer and its customers is examined in both symmetric and asymmetric information environments. A monopolist always prefers contracting with uninformed customers as this maximizes the rent extracted under symmetric information while also avoiding the cost of adverse selection when information is held asymmetrically. Although customers are indifferent to symmetric information when they are initially uninformed, they prefer contracting with hidden knowledge rather than symmetric information since the monopoly responds to adverse selection by sharing gains from trade with high-risk customers when low risks are predominant in the insurance pool. However, utilitarian social welfare is highest when customers are uninformed, and is higher when information is symmetric rather than asymmetric.Risks2015-07-2433Article10.3390/risks30302772772892227-90912015-07-24doi: 10.3390/risks3030277Arthur Snow<![CDATA[Risks, Vol. 3, Pages 250-276: Best-Estimates in Bond Markets with Reinvestment Risk]]>
http://www.mdpi.com/2227-9091/3/3/250
The concept of best-estimate, prescribed by regulators to value insurance liabilities for accounting and solvency purposes, has recently been discussed extensively in the industry and related academic literature. To differentiate hedgeable and non-hedgeable risks in a general case, recent literature defines best-estimates using orthogonal projections of a claim on the space of replicable payoffs. In this paper, we apply this concept of best-estimate to long-maturity claims in a market with reinvestment risk, since in this case the total liability cannot easily be separated into hedgeable and non-hedgeable parts. We assume that a limited number of short-maturity bonds are traded, and derive the best-estimate price of bonds with longer maturities, thus obtaining a best-estimate yield curve. We therefore use the multifactor Vasiˇcek model and derive within this framework closed-form expressions for the best-estimate prices of long-term bonds.Risks2015-07-1633Article10.3390/risks30302502502762227-90912015-07-16doi: 10.3390/risks3030250Anne MacKayMario Wüthrich<![CDATA[Risks, Vol. 3, Pages 234-249: Options with Extreme Strikes]]>
http://www.mdpi.com/2227-9091/3/3/234
In this short paper, we study the asymptotics for the price of call options for very large strikes and put options for very small strikes. The stock price is assumed to follow the Black–Scholes models. We analyze European, Asian, American, Parisian and perpetual options and conclude that the tail asymptotics for these option types fall into four scenarios.Risks2015-07-0833Article10.3390/risks30302342342492227-90912015-07-08doi: 10.3390/risks3030234Lingjiong Zhu<![CDATA[Risks, Vol. 3, Pages 219-233: Multiscale Analysis of the Predictability of Stock Returns]]>
http://www.mdpi.com/2227-9091/3/2/219
Due to the strong complexity of financial markets, economics does not have a unified theory of price formation in financial markets. The most common assumption is the Efficient-Market Hypothesis, which has been attacked by a number of researchers, using different tools. There were varying degrees to which these tools complied with the formal definitions of efficiency and predictability. In our earlier work, we analysed the predictability of stock returns at two time scales using the entropy rate, which can be directly linked to the mathematical definition of predictability. Nonetheless, none of the above-mentioned studies allow any general understanding of how the financial markets work, beyond disproving the Efficient-Market Hypothesis. In our previous study, we proposed the Maximum Entropy Production Principle, which uses the entropy rate to create a general principle underlying the price formation processes. Both of these studies show that the predictability of price changes is higher at the transaction level intraday scale than the scale of daily returns, but ignore all scales in between. In this study we extend these ideas using the multiscale entropy analysis framework to enhance our understanding of the predictability of price formation processes at various time scales.Risks2015-06-0832Article10.3390/risks30202192192332227-90912015-06-08doi: 10.3390/risks3020219Paweł Fiedor<![CDATA[Risks, Vol. 3, Pages 183-218: A Two-Account Life Insurance Model for Scenario-Based Valuation Including Event Risk]]>
http://www.mdpi.com/2227-9091/3/2/183
Using a two-account model with event risk, we model life insurance contracts taking into account both guaranteed and non-guaranteed payments in participating life insurance as well as in unit-linked insurance. Here, event risk is used as a generic term for life insurance events, such as death, disability, etc. In our treatment of participating life insurance, we have special focus on the bonus schemes “consolidation” and “additional benefits”, and one goal is to formalize how these work and interact. Another goal is to describe similarities and differences between participating life insurance and unit-linked insurance. By use of a two-account model, we are able to illustrate general concepts without making the model too abstract. To allow for complicated financial markets without dramatically increasing the mathematical complexity, we focus on economic scenarios. We illustrate the use of our model by conducting scenario analysis based on Monte Carlo simulation, but the model applies to scenarios in general and to worst-case and best-estimate scenarios in particular. In addition to easy computations, our model offers a common framework for the valuation of life insurance payments across product types. This enables comparison of participating life insurance products and unit-linked insurance products, thus building a bridge between the two different ways of formalizing life insurance products. Finally, our model distinguishes itself from the existing literature by taking into account the Markov model for the state of the policyholder and, hereby, facilitating event risk.Risks2015-06-0432Article10.3390/risks30201831832182227-90912015-06-04doi: 10.3390/risks3020183Ninna JensenKristian Schomacker<![CDATA[Risks, Vol. 3, Pages 164-182: The Impact of Reinsurance Strategies on Capital Requirements for Premium Risk in Insurance]]>
http://www.mdpi.com/2227-9091/3/2/164
New risk-based solvency requirements for insurance companies across European markets have been introduced by Solvency II and will come in force from 1 January 2016. These requirements, derived by a Standard Formula or an Internal Model, will be by far more risk-sensitive than the required solvency margin provided by the current legislation. In this regard, a Partial Internal Model for Premium Risk is developed here for a multi-line Non-Life insurer. We follow a classical approach based on a Collective Risk Model properly extended in order to consider not only the volatility of aggregate claim amounts but also expense volatility. To measure the effect of risk mitigation, suitable reinsurance strategies are pursued. We analyze how naïve coverage as conventional Quota Share and Excess of Loss reinsurance may modify the exact moments of the distribution of technical results. Furthermore, we investigate how alternative choices of commission rates in proportional treaties may affect the variability of distribution. Numerical results are also figured out in the last part of the paper with evidence of different effects for small and large companies. The main reasons for these differences are pointed out.Risks2015-06-0332Article10.3390/risks30201641641822227-90912015-06-03doi: 10.3390/risks3020164Gian ClementeNino SavelliDiego Zappa<![CDATA[Risks, Vol. 3, Pages 139-163: Interconnectedness of Financial Conglomerates]]>
http://www.mdpi.com/2227-9091/3/2/139
Being active in both the insurance sector and the banking sector, financial conglomerates intrinsically increase the interconnections between the banking sector and the insurance sector. We address two main concerns about financial conglomerates using a unique database on bilateral exposures between 21 French financial institutions. First, we investigate to what extent to which the insurers that are part of financial conglomerates differ from pure insurers. Second, we show that in the presence of sovereign risk, the components of a financial conglomerate are better off than if they were distinct entities. Our empirical findings bring a new perspective to the previous results of the literature based on using different types of data.Risks2015-05-2132Article10.3390/risks30201391391632227-90912015-05-21doi: 10.3390/risks3020139Gaël HautonJean-Cyprien Héam<![CDATA[Risks, Vol. 3, Pages 112-138: Custom v. Standardized Risk Models]]>
http://www.mdpi.com/2227-9091/3/2/112
We discuss when and why custom multi-factor risk models are warranted and give source code for computing some risk factors. Pension/mutual funds do not require customization but standardization. However, using standardized risk models in quant trading with much shorter holding horizons is suboptimal: (1) longer horizon risk factors (value, growth, etc.) increase noise trades and trading costs; (2) arbitrary risk factors can neutralize alpha; (3) “standardized” industries are artificial and insufficiently granular; (4) normalization of style risk factors is lost for the trading universe; (5) diversifying risk models lowers P&amp;L correlations, reduces turnover and market impact, and increases capacity. We discuss various aspects of custom risk model building.Risks2015-05-2032Article10.3390/risks30201121121382227-90912015-05-20doi: 10.3390/risks3020112Zura KakushadzeJim Liew<![CDATA[Risks, Vol. 3, Pages 103-111: Rationality Parameter for Exercising American Put]]>
http://www.mdpi.com/2227-9091/3/2/103
In this paper, irrational exercise behavior of the buyer of an American put is characterized by a single parameter. We model irrational exercise rules as the first jump time of a point processes with stochastic intensity. By the rationality parameter, we parameterize a family of stochastic intensities that depends on the value of the put itself. We present a probabilistic proof that the value of the American put using the irrational exercise rule converges to the arbitrage-free price as the rationality parameter converges to infinity. Another application of this result is the penalty method for approximating the price of an American put.Risks2015-05-2032Article10.3390/risks30201031031112227-90912015-05-20doi: 10.3390/risks3020103Kamille GadJesper Pedersen<![CDATA[Risks, Vol. 3, Pages 77-102: Portability, Salary and Asset Price Risk: A Continuous-Time Expected Utility Comparison of DB and DC Pension Plans]]>
http://www.mdpi.com/2227-9091/3/1/77
This paper compares two different types of private retirement plans from the perspective of a representative beneficiary: a defined benefit (DB) and a defined contribution (DC) plan. While salary risk is the main common risk factor in DB and DC pension plans, one of the key differences is that DB plans carry portability risks, whereas DC plans bear asset price risk. We model these tradeoffs explicitly in this paper and compare these two plans in a utility-based framework. Our numerical analysis focuses on answering the question of when the beneficiary is indifferent between the DB and DC plan. Most of our results confirm the findings in the existing literature, among which, e.g., portability losses considerably reduce the relative attractiveness of the DB plan. However, we also find that the attractiveness of the DB plan can decrease in the level of risk aversion, which is inconsistent with the existing literature.Risks2015-03-1331Article10.3390/risks3010077771022227-90912015-03-13doi: 10.3390/risks3010077An ChenFilip Uzelac<![CDATA[Risks, Vol. 3, Pages 61-76: Double Crowding-Out Effects of Means-Tested Public Provision for Long-Term Care]]>
http://www.mdpi.com/2227-9091/3/1/61
Publicly provided long-term care (LTC) insurance with means-tested benefits is suspected to crowd out either private saving or informal care. This contribution predicts crowding-out effects for both private saving and informal care for policy measures designed to relieve the public purse from LTC expenditure such as more stringent means testing and increased taxation of inheritance. These effects result from the interaction of a parent who decides on the amount of saving in retirement and a caregiver who decides on the effort devoted to informal care which lowers the probability of admission to a nursing home. Double crowding-out effects are also found to be the consequence of exogenous influences, notably a higher opportunity cost of caregiving.Risks2015-02-2531Article10.3390/risks301006161762227-90912015-02-25doi: 10.3390/risks3010061Christophe CourbagePeter Zweifel<![CDATA[Risks, Vol. 3, Pages 35-60: Safety Margins for Systematic Biometric and Financial Risk in a Semi-Markov Life Insurance Framework]]>
http://www.mdpi.com/2227-9091/3/1/35
Insurance companies use conservative first order valuation bases to calculate insurance premiums and reserves. These valuation bases have a significant impact on the insurer’s solvency and on the premiums of the insurance products. Safety margins for systematic biometric and financial risk are in practice typically chosen as time-constant percentages on top of the best estimate transition intensities. We develop a risk-oriented method for the allocation of a total safety margin to the single safety margins at each point in time and each state. In a case study, we demonstrate the suitability of the proposed method in different frameworks. The results show that the traditional method yields an unwanted variability of the safety level with respect to time, whereas the variability can be significantly reduced by the new method. Furthermore, the case study supports the German 60 percent rule for the technical interest rate.Risks2015-01-1931Article10.3390/risks301003535602227-90912015-01-19doi: 10.3390/risks3010035Andreas Niemeyer<![CDATA[Risks, Vol. 3, Pages 26-34: Paradox-Proof Utility Functions for Heavy-Tailed Payoffs: Two Instructive Two-Envelope Problems]]>
http://www.mdpi.com/2227-9091/3/1/26
We identify restrictions on a decision maker’s utility function that are both necessary and sufficient to preserve dominance reasoning in each of two versions of the Two-Envelope Paradox (TEP). For the classical TEP, the utility function must satisfy a certain recurrence inequality. For the St. Petersburg TEP, the utility function must be bounded above asymptotically by a power function, which can be tightened to a constant. By determining the weakest conditions for dominance reasoning to hold, the article settles an open question in the research literature. Remarkably, neither constant-bounded utility nor finite expected utility is necessary for resolving the classical TEP; instead, finite expected utility is both necessary and sufficient for resolving the St. Petersburg TEP.Risks2015-01-1931Article10.3390/risks301002626342227-90912015-01-19doi: 10.3390/risks3010026Michael Powers<![CDATA[Risks, Vol. 3, Pages 24-25: Acknowledgement to Reviewers of Risks in 2014]]>
http://www.mdpi.com/2227-9091/3/1/24
The editors of Risks would like to express their sincere gratitude to the following reviewers for assessing manuscripts in 2014:[...]Risks2015-01-0831Editorial10.3390/risks301002424252227-90912015-01-08doi: 10.3390/risks3010024 Risks Editorial Office<![CDATA[Risks, Vol. 3, Pages 1-23: Inhomogeneous Long-Range Percolation for Real-Life Network Modeling]]>
http://www.mdpi.com/2227-9091/3/1/1
The study of random graphs has become very popular for real-life network modeling, such as social networks or financial networks. Inhomogeneous long-range percolation (or scale-free percolation) on the lattice Zd, d ≥ 1, is a particular attractive example of a random graph model because it fulfills several stylized facts of real-life networks. For this model, various geometric properties, such as the percolation behavior, the degree distribution and graph distances, have been analyzed. In the present paper, we complement the picture of graph distances and we prove continuity of the percolation probability in the phase transition point. We also provide an illustration of the model connected to financial networks.Risks2015-01-0631Article10.3390/risks30100011232227-90912015-01-06doi: 10.3390/risks3010001Philippe DeprezRajat HazraMario Wüthrich<![CDATA[Risks, Vol. 2, Pages 469-488: Worst-Case Portfolio Optimization under Stochastic Interest Rate Risk]]>
http://www.mdpi.com/2227-9091/2/4/469
We investigate a portfolio optimization problem under the threat of a market crash, where the interest rate of the bond is modeled as a Vasicek process, which is correlated with the stock price process. We adopt a non-probabilistic worst-case approach for the height and time of the market crash. On a given time horizon [0; T], we then maximize the investor’s expected utility of terminal wealth in the worst-case crash scenario. Our main result is an explicit characterization of the worst-case optimal portfolio strategy for the class of HARA (hyperbolic absolute risk aversion) utility functions.Risks2014-12-0124Article10.3390/risks20404694694882227-90912014-12-01doi: 10.3390/risks2040469Tina EnglerRalf Korn<![CDATA[Risks, Vol. 2, Pages 467-468: Special Issue on Risk Management Techniques for Catastrophic and Heavy-Tailed Risks]]>
http://www.mdpi.com/2227-9091/2/4/467
The publication of several special issues was part of the initiatives taken in 2013 to launch Risks as a new online journal. It seemed natural to devote one to this important, concrete and complex problem of managing catastrophic and heavy tailed risks. We received an enthusiastic response last spring to the call for invited and contributed research papers and are proud of the special issue now being published. The emphasis was put on quality rather than quantity; this special issue contains three invited and two contributed research papers.Risks2014-11-1424Editorial10.3390/risks20404674674682227-90912014-11-14doi: 10.3390/risks2040467Alejandro BalbásJosé Garrido<![CDATA[Risks, Vol. 2, Pages 456-466: A Duality Result for the Generalized Erlang Risk Model]]>
http://www.mdpi.com/2227-9091/2/4/456
In this article, we consider the generalized Erlang risk model and its dual model. By using a conditional measure-preserving correspondence between the two models, we derive an identity for two interesting conditional probabilities. Applications to the discounted joint density of the surplus prior to ruin and the deficit at ruin are also discussed.Risks2014-11-0624Article10.3390/risks20404564564662227-90912014-11-06doi: 10.3390/risks2040456Lanpeng JiChunsheng Zhang<![CDATA[Risks, Vol. 2, Pages 434-455: A Markov Chain Model for Contagion]]>
http://www.mdpi.com/2227-9091/2/4/434
We introduce a bivariate Markov chain counting process with contagion for modelling the clustering arrival of loss claims with delayed settlement for an insurance company. It is a general continuous-time model framework that also has the potential to be applicable to modelling the clustering arrival of events, such as jumps, bankruptcies, crises and catastrophes in finance, insurance and economics with both internal contagion risk and external common risk. Key distributional properties, such as the moments and probability generating functions, for this process are derived. Some special cases with explicit results and numerical examples and the motivation for further actuarial applications are also discussed. The model can be considered a generalisation of the dynamic contagion process introduced by Dassios and Zhao (2011).Risks2014-11-0524Article10.3390/risks20404344344552227-90912014-11-05doi: 10.3390/risks2040434Angelos DassiosHongbiao Zhao<![CDATA[Risks, Vol. 2, Pages 425-433: A Note on the Fundamental Theorem of Asset Pricing under Model Uncertainty]]>
http://www.mdpi.com/2227-9091/2/4/425
We show that the recent results on the Fundamental Theorem of Asset Pricing and the super-hedging theorem in the context of model uncertainty can be extended to the case in which the options available for static hedging (hedging options) are quoted with bid-ask spreads. In this set-up, we need to work with the notion of robust no-arbitrage which turns out to be equivalent to no-arbitrage under the additional assumption that hedging options with non-zero spread are non-redundant. A key result is the closedness of the set of attainable claims, which requires a new proof in our setting.Risks2014-10-1024Article10.3390/risks20404254254332227-90912014-10-10doi: 10.3390/risks2040425Erhan BayraktarYuchong ZhangZhou Zhou<![CDATA[Risks, Vol. 2, Pages 411-424: Measuring Risk When Expected Losses Are Unbounded]]>
http://www.mdpi.com/2227-9091/2/4/411
This paper proposes a new method to introduce coherent risk measures for risks with infinite expectation, such as those characterized by some Pareto distributions. Extensions of the conditional value at risk, the weighted conditional value at risk and other examples are given. Actuarial applications are analyzed, such as extensions of the expected value premium principle when expected losses are unbounded.Risks2014-09-3024Article10.3390/risks20404114114242227-90912014-09-30doi: 10.3390/risks2040411Alejandro BalbásIván BlancoJosé Garrido<![CDATA[Risks, Vol. 2, Pages 393-410: Tail Risk in Commercial Property Insurance]]>
http://www.mdpi.com/2227-9091/2/4/393
We present some new evidence on the tail distribution of commercial property losses based on a recently constructed dataset on large commercial risks. The dataset is based on contributions from Lloyd’s of London syndicates, and provides information on over three thousand claims occurred during the period 2000–2012, including detailed information on exposures. We use occupancy characteristics to compare the tail risk profiles of different commercial property exposures, and find evidence of substantial heterogeneity in tail behavior. The results demonstrate the benefits of aggregating granular information on both claims and exposures from different data sources, and provide warning against the use of reserving and capital modeling approaches that are not robust to heavy tails.Risks2014-09-2924Article10.3390/risks20403933934102227-90912014-09-29doi: 10.3390/risks2040393Enrico BiffisErik Chavez<![CDATA[Risks, Vol. 2, Pages 349-392: An Optimal Three-Way Stable and Monotonic Spectrum of Bounds on Quantiles: A Spectrum of Coherent Measures of Financial Risk and Economic Inequality]]>
http://www.mdpi.com/2227-9091/2/3/349
A spectrum of upper bounds (Qα(X ; p) αε[0,∞] on the (largest) (1-p)-quantile Q(X;p) of an arbitrary random variable X is introduced and shown to be stable and monotonic in α, p, and X , with Q0(X ;p) = Q(X;p). If p is small enough and the distribution of X is regular enough, then Qα(X ; p) is rather close to Q(X ; p). Moreover, these quantile bounds are coherent measures of risk. Furthermore, Qα(X ; p) is the optimal value in a certain minimization problem, the minimizers in which are described in detail. This allows of a comparatively easy incorporation of these bounds into more specialized optimization problems. In finance, Q0(X;p) and Q1(X ; p) are known as the value at risk (VaR) and the conditional value at risk (CVaR). The bounds Qα(X ; p) can also be used as measures of economic inequality. The spectrum parameter α plays the role of an index of sensitivity to risk. The problems of the effective computation of the bounds are considered. Various other related results are obtained.Risks2014-09-2323Article10.3390/risks20303493493922227-90912014-09-23doi: 10.3390/risks2030349Iosif Pinelis<![CDATA[Risks, Vol. 2, Pages 315-348: Model Risk in Portfolio Optimization]]>
http://www.mdpi.com/2227-9091/2/3/315
We consider a one-period portfolio optimization problem under model uncertainty. For this purpose, we introduce a measure of model risk. We derive analytical results for this measure of model risk in the mean-variance problem assuming we have observations drawn from a normal variance mixture model. This model allows for heavy tails, tail dependence and leptokurtosis of marginals. The results show that mean-variance optimization is seriously compromised by model uncertainty, in particular, for non-Gaussian data and small sample sizes. To mitigate these shortcomings, we propose a method to adjust the sample covariance matrix in order to reduce model risk.Risks2014-08-0623Article10.3390/risks20303153153482227-90912014-08-06doi: 10.3390/risks2030315David StefanovitsUrs SchubigerMario Wüthrich<![CDATA[Risks, Vol. 2, Pages 289-314: Joint Asymptotic Distributions of Smallest and Largest Insurance Claims]]>
http://www.mdpi.com/2227-9091/2/3/289
Assume that claims in a portfolio of insurance contracts are described by independent and identically distributed random variables with regularly varying tails and occur according to a near mixed Poisson process. We provide a collection of results pertaining to the joint asymptotic Laplace transforms of the normalised sums of the smallest and largest claims, when the length of the considered time interval tends to infinity. The results crucially depend on the value of the tail index of the claim distribution, as well as on the number of largest claims under consideration.Risks2014-07-3123Article10.3390/risks20302892893142227-90912014-07-31doi: 10.3390/risks2030289Hansjörg AlbrecherChristian RobertJef Teugels<![CDATA[Risks, Vol. 2, Pages 277-288: Random Shifting and Scaling of Insurance Risks]]>
http://www.mdpi.com/2227-9091/2/3/277
Random shifting typically appears in credibility models whereas random scaling is often encountered in stochastic models for claim sizes reflecting the time-value property of money. In this article we discuss some aspects of random shifting and random scaling of insurance risks focusing in particular on credibility models, dependence structure of claim sizes in collective risk models, and extreme value models for the joint dependence of large losses. We show that specifying certain actuarial models using random shifting or scaling has some advantages for both theoretical treatments and practical applications.Risks2014-07-2223Article10.3390/risks20302772772882227-90912014-07-22doi: 10.3390/risks2030277Enkelejd HashorvaLanpeng Ji<![CDATA[Risks, Vol. 2, Pages 260-276: The Impact of Systemic Risk on the Diversification Benefits of a Risk Portfolio]]>
http://www.mdpi.com/2227-9091/2/3/260
Risk diversification is the basis of insurance and investment. It is thus crucial to study the effects that could limit it. One of them is the existence of systemic risk that affects all of the policies at the same time. We introduce here a probabilistic approach to examine the consequences of its presence on the risk loading of the premium of a portfolio of insurance policies. This approach could be easily generalized for investment risk. We see that, even with a small probability of occurrence, systemic risk can reduce dramatically the diversification benefits. It is clearly revealed via a non-diversifiable term that appears in the analytical expression of the variance of our models. We propose two ways of introducing it and discuss their advantages and limitations. By using both VaR and TVaR to compute the loading, we see that only the latter captures the full effect of systemic risk when its probability to occur is low.Risks2014-07-0923Article10.3390/risks20302602602762227-90912014-07-09doi: 10.3390/risks2030260Marc BusseMichel DacorognaMarie Kratz<![CDATA[Risks, Vol. 2, Pages 249-259: Elementary Bounds on the Ruin Capital in a Diffusion Model of Risk]]>
http://www.mdpi.com/2227-9091/2/3/249
In a diffusion model of risk, we focus on the initial capital needed to make the probability of ruin within finite time equal to a prescribed value. It is defined as a solution of a nonlinear equation. The endeavor to write down and to investigate analytically this solution as a function of the premium rate seems not technically feasible. Instead, we obtain informative bounds for this capital in terms of elementary functions.Risks2014-07-0823Article10.3390/risks20302492492592227-90912014-07-08doi: 10.3390/risks2030249Vsevolod Malinovskii<![CDATA[Risks, Vol. 2, Pages 226-248: Demand of Insurance under the Cost-of-Capital Premium Calculation Principle]]>
http://www.mdpi.com/2227-9091/2/2/226
We study the optimal insurance design problem. This is a risk sharing problem between an insured and an insurer. The main novelty in this paper is that we study this optimization problem under a risk-adjusted premium calculation principle for the insurance cover. This risk-adjusted premium calculation principle uses the cost-of-capital approach as it is suggested (and used) by the regulator and the insurance industry.Risks2014-06-1722Article10.3390/risks20202262262482227-90912014-06-17doi: 10.3390/risks2020226Michael MerzMario Wüthrich<![CDATA[Risks, Vol. 2, Pages 211-225: When the U.S. Stock Market Becomes Extreme?]]>
http://www.mdpi.com/2227-9091/2/2/211
Over the last three decades, the world economy has been facing stock market crashes, currency crisis, the dot-com and real estate bubble burst, credit crunch and banking panics. As a response, extreme value theory (EVT) provides a set of ready-made approaches to risk management analysis. However, EVT is usually applied to standardized returns to offer more reliable results, but remains difficult to interpret in the real world. This paper proposes a quantile regression to transform standardized returns into theoretical raw returns making them economically interpretable. An empirical test is carried out on the S&amp;P500 stock index from 1950 to 2013. The main results indicate that the U.S stock market becomes extreme from a price variation of ±1.5% and the largest one-day decline of the 2007–2008 period is likely, on average, to be exceeded one every 27 years.Risks2014-05-2822Article10.3390/risks20202112112252227-90912014-05-28doi: 10.3390/risks2020211Sofiane Aboura<![CDATA[Risks, Vol. 2, Pages 195-210: Neumann Series on the Recursive Moments of Copula-Dependent Aggregate Discounted Claims]]>
http://www.mdpi.com/2227-9091/2/2/195
We study the recursive moments of aggregate discounted claims, where the dependence between the inter-claim time and the subsequent claim size is considered. Using the general expression for the m-th order moment proposed by Léveillé and Garrido (Scand. Actuar. J. 2001, 2, 98–110), which takes the form of the Volterra integral equation (VIE), we used the method of successive approximation to derive the Neumann series of the recursive moments. We then compute the first two moments of aggregate discounted claims, i.e., its mean and variance, based on the Neumann series expression, where the dependence structure is captured by a Farlie–Gumbel–Morgenstern (FGM) copula, a Gaussian copula and a Gumbel copula with exponential marginal distributions. Insurance premium calculations with their figures are also illustrated.Risks2014-05-2722Article10.3390/risks20201951952102227-90912014-05-27doi: 10.3390/risks2020195Siti Mohd RamliJiwook Jang<![CDATA[Risks, Vol. 2, Pages 171-194: Optimal Consumption and Investment with Labor Income and European/American Capital Guarantee]]>
http://www.mdpi.com/2227-9091/2/2/171
We present the optimal consumption and investment strategy for an investor, endowed with labor income, searching to maximize utility from consumption and terminal wealth when facing a binding capital constraint of a European (constraint on terminal wealth) or an American (constraint on the wealth process) type. In both cases, the optimal strategy is proven to be of the option-based portfolio insurance type. The optimal strategy combines a long position in the optimal unrestricted allocation with a put option. In the American case, where the investor is restricted to fulfill a capital guarantee at every intermediate time point over the interval of optimization, we prove that the investor optimally changes his budget constraint for the unrestricted allocation whenever the constraint is active. The strategy is explained in a step-by-step manner, and numerical illustrations are presented in order to support intuition and to compare the restricted optimal strategy with the unrestricted optimal counterpart.Risks2014-05-1622Article10.3390/risks20201711711942227-90912014-05-16doi: 10.3390/risks2020171Morten Kronborg<![CDATA[Risks, Vol. 2, Pages 146-170: Attracting Health Insurance Buyers through Selective Contracting: Results of a Discrete-Choice Experiment among Users of Hospital Services in the Netherlands]]>
http://www.mdpi.com/2227-9091/2/2/146
In 2006, the Netherlands commenced market based reforms in its health care system. The reforms included selective contracting of health care providers by health insurers. This paper focuses on how health insurers may increase their market share on the health insurance market through selective contracting of health care providers. Selective contracting is studied by eliciting the preferences of health care consumers for attributes of health care services that an insurer could negotiate on behalf of its clients with health care providers. Selective contracting may provide incentives for health care providers to deliver the quality that consumers need and demand. Selective contracting also enables health insurers to steer individual patients towards selected health care providers. We used a stated preference technique known as a discrete choice experiment to collect and analyze the data. Results indicate that consumers care about both costs and quality of care, with healthy consumers placing greater emphasis on costs and consumers with poorer health placing greater emphasis on quality of care. It is possible for an insurer to satisfy both of these criteria by selective contracting health care providers who consequently purchase health care that is both efficient and of good quality.Risks2014-04-1522Article10.3390/risks20201461461702227-90912014-04-15doi: 10.3390/risks2020146Evelien BergrathMilena PavlovaWim Groot<![CDATA[Risks, Vol. 2, Pages 132-145: Effectively Tackling Reinsurance Problems by Using Evolutionary and Swarm Intelligence Algorithms]]>
http://www.mdpi.com/2227-9091/2/2/132
This paper is focused on solving different hard optimization problems that arise in the field of insurance and, more specifically, in reinsurance problems. In this area, the complexity of the models and assumptions considered in the definition of the reinsurance rules and conditions produces hard black-box optimization problems (problems in which the objective function does not have an algebraic expression, but it is the output of a system (usually a computer program)), which must be solved in order to obtain the optimal output of the reinsurance. The application of traditional optimization approaches is not possible in this kind of mathematical problem, so new computational paradigms must be applied to solve these problems. In this paper, we show the performance of two evolutionary and swarm intelligence techniques (evolutionary programming and particle swarm optimization). We provide an analysis in three black-box optimization problems in reinsurance, where the proposed approaches exhibit an excellent behavior, finding the optimal solution within a fraction of the computational cost used by inspection or enumeration methods.Risks2014-04-0122Article10.3390/risks20201321321452227-90912014-04-01doi: 10.3390/risks2020132Sancho Salcedo-SanzLeo Carro-CalvoMercè ClaramuntAna CastañerMaite Mármol<![CDATA[Risks, Vol. 2, Pages 103-131: 1980–2008: The Illusion of the Perpetual Money Machine and What It Bodes for the Future]]>
http://www.mdpi.com/2227-9091/2/2/103
We argue that the present crisis and stalling economy that have been ongoing since 2007 are rooted in the delusionary belief in policies based on a “perpetual money machine” type of thinking. We document strong evidence that, since the early 1980s, consumption has been increasingly funded by smaller savings, booming financial profits, wealth extracted from house price appreciation and explosive debt. This is in stark contrast with the productivity-fueled growth that was seen in the 1950s and 1960s. We describe the transition, in gestation in the 1970s, towards the regime of the “illusion of the perpetual money machine”, which started at full speed in the early 1980s and developed until 2008. This regime was further supported by a climate of deregulation and a massive growth in financial derivatives designed to spread and diversify the risks globally. The result has been a succession of bubbles and crashes, including the worldwide stock market bubble and great crash of October 1987, the savings and loans crisis of the 1980s, the burst in 1991 of the enormous Japanese real estate and stock market bubbles, the emerging markets bubbles and crashes in 1994 and 1997, the Long-Term Capital Management (LTCM) crisis of 1998, the dotcom bubble bursting in 2000, the recent house price bubbles, the financialization bubble via special investment vehicles, the stock market bubble, the commodity and oil bubbles and the current debt bubble, all developing jointly and feeding on each other until 2008. This situation may be further aggravated in the next decade by an increase in financialization, through exchange-traded-funds (ETFs), speed and automation, through algorithmic trading and public debt, and through growing unfunded liabilities. We conclude that, to get out of this catch 22 situation, we should better manage and understand the incentive structures in our society, we need to focus our efforts on our real economy and we have to respect and master the art of planning and prediction. Only gradual change, with a clear long term planning, can steer our financial and economic system from the turbulence associated with the perpetual money machine to calmer and more sustainable waters. Risks2014-04-0122Article10.3390/risks20201031031312227-90912014-04-01doi: 10.3390/risks2020103Didier SornettePeter Cauwels<![CDATA[Risks, Vol. 2, Pages 89-102: Initial Investigations of Intra-Day News Flow of S&P500 Constituents]]>
http://www.mdpi.com/2227-9091/2/2/89
In this work, we examine Thomas Reuters News Analytics (TRNA) data. We found several fascinating discoveries. First, we document the phenomenon that we label “Jam-the-Close”: The last half hour of trading (15:30 to 16:00 EST) contains a substantial and statistically significant amount of news sentiment releases. This finding is robust across years and months of the year. Next, upon further investigations we found that the “novelty” score is on average 0.67 in this period vs. 2.09 prior to midday. This indicates that “new” news is flowing at a rapid pace prior to the close. Finally, we discuss the implication of such phenomena in the context of existing financial literature.Risks2014-04-0122Article10.3390/risks2020089891022227-90912014-04-01doi: 10.3390/risks2020089Jim LiewZhechao Zhou<![CDATA[Risks, Vol. 2, Pages 74-88: Modeling Cycle Dependence in Credit Insurance]]>
http://www.mdpi.com/2227-9091/2/1/74
Business and credit cycles have an impact on credit insurance, as they do on other businesses. Nevertheless, in credit insurance, the impact of the systemic risk is even more important and can lead to major losses during a crisis. Because of this, the insurer surveils and manages policies almost continuously. The management actions it takes limit the consequences of a downturning cycle. However, the traditional modeling of economic capital does not take into account this important feature of credit insurance. This paper proposes a model aiming to estimate future losses of a credit insurance portfolio, while taking into account the insurer’s management actions. The model considers the capacity of the credit insurer to take on less risk in the case of a cycle downturn, but also the inverse, in the case of a cycle upturn; so, losses are predicted with a more dynamic perspective. According to our results, the economic capital is over-estimated when not considering the management actions of the insurer.Risks2014-03-1421Article10.3390/risks201007474882227-90912014-03-14doi: 10.3390/risks2010074Anisa CajaFrédéric Planchet<![CDATA[Risks, Vol. 2, Pages 49-73: Modeling and Performance of Bonus-Malus Systems: Stationarity versus Age-Correction]]>
http://www.mdpi.com/2227-9091/2/1/49
In a bonus-malus system in car insurance, the bonus class of a customer is updated from one year to the next as a function of the current class and the number of claims in the year (assumed Poisson). Thus the sequence of classes of a customer in consecutive years forms a Markov chain, and most of the literature measures performance of the system in terms of the stationary characteristics of this Markov chain. However, the rate of convergence to stationarity may be slow in comparison to the typical sojourn time of a customer in the portfolio. We suggest an age-correction to the stationary distribution and present an extensive numerical study of its effects. An important feature of the modeling is a Bayesian view, where the Poisson rate according to which claims are generated for a customer is the outcome of a random variable specific to the customer.Risks2014-03-1121Article10.3390/risks201004949732227-90912014-03-11doi: 10.3390/risks2010049Søren Asmussen<![CDATA[Risks, Vol. 2, Pages 25-48: An Academic Response to Basel 3.5]]>
http://www.mdpi.com/2227-9091/2/1/25
Recent crises in the financial industry have shown weaknesses in the modeling of Risk-Weighted Assets (RWAs). Relatively minor model changes may lead to substantial changes in the RWA numbers. Similar problems are encountered in the Value-at-Risk (VaR)-aggregation of risks. In this article, we highlight some of the underlying issues, both methodologically, as well as through examples. In particular, we frame this discussion in the context of two recent regulatory documents we refer to as Basel 3.5.Risks2014-02-2721Article10.3390/risks201002525482227-90912014-02-27doi: 10.3390/risks2010025Paul EmbrechtsGiovanni PuccettiLudger RüschendorfRuodu WangAntonela Beleraj<![CDATA[Risks, Vol. 2, Pages 3-24: Catastrophe Insurance Modeled by Shot-Noise Processes]]>
http://www.mdpi.com/2227-9091/2/1/3
Shot-noise processes generalize compound Poisson processes in the following way: a jump (the shot) is followed by a decline (noise). This constitutes a useful model for insurance claims in many circumstances; claims due to natural disasters or self-exciting processes exhibit similar features. We give a general account of shot-noise processes with time-inhomogeneous drivers inspired by recent results in credit risk. Moreover, we derive a number of useful results for modeling and pricing with shot-noise processes. Besides this, we obtain some highly tractable examples and constitute a useful modeling tool for dynamic claims processes. The results can in particular be used for pricing Catastrophe Bonds (CAT bonds), a traded risk-linked security. Additionally, current results regarding the estimation of shot-noise processes are reviewed.Risks2014-02-2121Article10.3390/risks20100033242227-90912014-02-21doi: 10.3390/risks2010003Thorsten Schmidt<![CDATA[Risks, Vol. 2, Pages 1-2: Publishing Risks]]>
http://www.mdpi.com/2227-9091/2/1/1
“What is complicated is not necessarily insightful and what is insightful is not necessarily complicated: Risks welcomes simple manuscripts that contribute with insight, outlook, understanding and overview”—a quote from the first editorial of this journal [1]. Good articles are not characterized by their level of complication but by their level of imagination, innovation, and power of penetration. Creativity sessions and innovative tasks are most elegant and powerful when they are delicately simple. This is why the articles you most remember are not the complicated ones that you struggled to digest, but the simpler ones you enjoyed swallowing. [...]Risks2014-02-2121Editorial10.3390/risks2010001122227-90912014-02-21doi: 10.3390/risks2010001Mogens Steffensen<![CDATA[Risks, Vol. 1, Pages 192-212: Ruin Time and Severity for a Lévy Subordinator Claim Process: A Simple Approach]]>
http://www.mdpi.com/2227-9091/1/3/192
This paper is concerned with an insurance risk model whose claim process is described by a Lévy subordinator process. Lévy-type risk models have been the object of much research in recent years. Our purpose is to present, in the case of a subordinator, a simple and direct method for determining the finite time (and ultimate) ruin probabilities, the distribution of the ruin severity, the reserves prior to ruin, and the Laplace transform of the ruin time. Interestingly, the usual net profit condition will be essentially relaxed. Most results generalize those known for the compound Poisson claim process.Risks2013-12-1313Article10.3390/risks10301921922122227-90912013-12-13doi: 10.3390/risks1030192Claude LefèvrePhilippe Picard<![CDATA[Risks, Vol. 1, Pages 176-191: Impact of Climate Change on Heat Wave Risk]]>
http://www.mdpi.com/2227-9091/1/3/176
We study a new risk measure inspired from risk theory with a heat wave risk analysis motivation. We show that this risk measure and its sensitivities can be computed in practice for relevant temperature stochastic processes. This is in particular useful for measuring the potential impact of climate change on heat wave risk. Numerical illustrations are given.Risks2013-12-1213Article10.3390/risks10301761761912227-90912013-12-12doi: 10.3390/risks1030176Romain BiardChristophette Blanchet-ScallietAnne Eyraud-LoiselStéphane Loisel<![CDATA[Risks, Vol. 1, Pages 162-175: U.S. Equity Mean-Reversion Examined]]>
http://www.mdpi.com/2227-9091/1/3/162
In this paper we introduce an intra-sector dynamic trading strategy that captures mean-reversion opportunities across liquid U.S. stocks. Our strategy combines the Avellaneda and Lee methodology (AL; Quant. Financ. 2010, 10, 761–782) within the Black and Litterman framework (BL; J. Fixed Income, 1991, 1, 7–18; Financ. Anal. J. 1992, 48, 28–43). In particular, we incorporate the s-scores and the conditional mean returns from the Orstein and Ulhembeck (Phys. Rev. 1930, 36, 823–841) process into BL. We find that our combined strategy ALBL has generated a 45% increase in Sharpe Ratio when compared to the uncombined AL strategy over the period from January 2, 2001 to May 27, 2010. These new indices, built to capture dynamic trading strategies, will definitely be an interesting addition to the growing hedge fund index offerings. This paper introduces our first “focused-core” strategy, namely, U.S. Equity Mean-Reversion.Risks2013-12-0413Article10.3390/risks10301621621752227-90912013-12-04doi: 10.3390/risks1030162Jim LiewRyan Roberts<![CDATA[Risks, Vol. 1, Pages 148-161: A Risk Model with an Observer in a Markov Environment]]>
http://www.mdpi.com/2227-9091/1/3/148
We consider a spectrally-negative Markov additive process as a model of a risk process in a random environment. Following recent interest in alternative ruin concepts, we assume that ruin occurs when an independent Poissonian observer sees the process as negative, where the observation rate may depend on the state of the environment. Using an approximation argument and spectral theory, we establish an explicit formula for the resulting survival probabilities in this general setting. We also discuss an efficient evaluation of the involved quantities and provide a numerical illustration.Risks2013-11-1113Article10.3390/risks10301481481612227-90912013-11-11doi: 10.3390/risks1030148Hansjörg AlbrecherJevgenijs Ivanovs<![CDATA[Risks, Vol. 1, Pages 119-147: Optimal Dynamic Portfolio with Mean-CVaR Criterion]]>
http://www.mdpi.com/2227-9091/1/3/119
Value-at-risk (VaR) and conditional value-at-risk (CVaR) are popular risk measures from academic, industrial and regulatory perspectives. The problem of minimizing CVaR is theoretically known to be of a Neyman–Pearson type binary solution. We add a constraint on expected return to investigate the mean-CVaR portfolio selection problem in a dynamic setting: the investor is faced with a Markowitz type of risk reward problem at the final horizon, where variance as a measure of risk is replaced by CVaR. Based on the complete market assumption, we give an analytical solution in general. The novelty of our solution is that it is no longer the Neyman–Pearson type, in which the final optimal portfolio takes only two values. Instead, in the case in which the portfolio value is required to be bounded from above, the optimal solution takes three values; while in the case in which there is no upper bound, the optimal investment portfolio does not exist, though a three-level portfolio still provides a sub-optimal solution.Risks2013-11-1113Article10.3390/risks10301191191472227-90912013-11-11doi: 10.3390/risks1030119Jing LiMingxin Xu<![CDATA[Risks, Vol. 1, Pages 101-118: Optimal Deterministic Investment Strategies for Insurers]]>
http://www.mdpi.com/2227-9091/1/3/101
We consider an insurance company whose risk reserve is given by a Brownian motion with drift and which is able to invest the money into a Black–Scholes financial market. As optimization criteria, we treat mean-variance problems, problems with other risk measures, exponential utility and the probability of ruin. Following recent research, we assume that investment strategies have to be deterministic. This leads to deterministic control problems, which are quite easy to solve. Moreover, it turns out that there are some interesting links between the optimal investment strategies of these problems. Finally, we also show that this approach works in the Lévy process framework.Risks2013-11-0713Article10.3390/risks10301011011182227-90912013-11-07doi: 10.3390/risks1030101Nicole BäuerleUlrich Rieder<![CDATA[Risks, Vol. 1, Pages 81-100: Gaussian and Affine Approximation of Stochastic Diffusion Models for Interest and Mortality Rates]]>
http://www.mdpi.com/2227-9091/1/3/81
In the actuarial literature, it has become common practice to model future capital returns and mortality rates stochastically in order to capture market risk and forecasting risk. Although interest rates often should and mortality rates always have to be non-negative, many authors use stochastic diffusion models with an affine drift term and additive noise. As a result, the diffusion process is Gaussian and, thus, analytically tractable, but negative values occur with positive probability. The argument is that the class of Gaussian diffusions would be a good approximation of the real future development. We challenge that reasoning and study the asymptotics of diffusion processes with affine drift and a general noise term with corresponding diffusion processes with an affine drift term and an affine noise term or additive noise. Our study helps to quantify the error that is made by approximating diffusive interest and mortality rate models with Gaussian diffusions and affine diffusions. In particular, we discuss forward interest and forward mortality rates and the error that approximations cause on the valuation of life insurance claims.Risks2013-10-2513Article10.3390/risks1030081811002227-90912013-10-25doi: 10.3390/risks1030081Marcus Christiansen<![CDATA[Risks, Vol. 1, Pages 57-80: A Welfare Analysis of Capital Insurance]]>
http://www.mdpi.com/2227-9091/1/2/57
This paper presents a welfare analysis of several capital insurance programs in a rational expectation equilibrium setting. We first explicitly characterize the equilibrium of each capital insurance program. Then, we demonstrate that a capital insurance program based on aggregate loss is better than classical insurance, when big financial institutions have similar expected loss exposures. By contrast, classical insurance is more desirable when the bank’s individual risk is consistent with the expected loss in a precise way. Our analysis shows that a capital insurance program is a useful tool to hedge systemic risk from the regulatory perspective.Risks2013-09-1712Article10.3390/risks102005757802227-90912013-09-17doi: 10.3390/risks1020057Ekaterina PanttserWeidong Tian