Next Article in Journal
Why to Buy Insurance? An Explainable Artificial Intelligence Approach
Next Article in Special Issue
The Weak Convergence Rate of Two Semi-Exact Discretization Schemes for the Heston Model
Previous Article in Journal
How Efficient Are Indian Banks in Managing the Risk-Return Trade-Off? An Empirical Analysis
Previous Article in Special Issue
Least-Squares Monte Carlo for Proxy Modeling in Life Insurance: Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Neural Network Algorithm for Semilinear Elliptic PDEs with Applications in Insurance Mathematics

by
Stefan Kremsner
1,*,
Alexander Steinicke
2 and
Michaela Szölgyenyi
3
1
Department of Mathematics, University of Graz, Heinrichstraße 36, 8010 Graz, Austria
2
Department of Mathematics and Information Technology, Montanuniversitaet Leoben, Peter Tunner-Straße 25/I, 8700 Leoben, Austria
3
Department of Statistics, University of Klagenfurt, Universitätsstraße 65-67, 9020 Klagenfurt, Austria
*
Author to whom correspondence should be addressed.
Risks 2020, 8(4), 136; https://doi.org/10.3390/risks8040136
Submission received: 29 October 2020 / Revised: 1 December 2020 / Accepted: 2 December 2020 / Published: 9 December 2020
(This article belongs to the Special Issue Computational Finance and Risk Analysis in Insurance)

Abstract

:
In insurance mathematics, optimal control problems over an infinite time horizon arise when computing risk measures. An example of such a risk measure is the expected discounted future dividend payments. In models which take multiple economic factors into account, this problem is high-dimensional. The solutions to such control problems correspond to solutions of deterministic semilinear (degenerate) elliptic partial differential equations. In the present paper we propose a novel deep neural network algorithm for solving such partial differential equations in high dimensions in order to be able to compute the proposed risk measure in a complex high-dimensional economic environment. The method is based on the correspondence of elliptic partial differential equations to backward stochastic differential equations with unbounded random terminal time. In particular, backward stochastic differential equations—which can be identified with solutions of elliptic partial differential equations—are approximated by means of deep neural networks.

1. Introduction

Classical optimal control problems in insurance mathematics include finding risk measures like the probability of ruin or the expected discounted future dividend payments. Mathematically, these are problems over a potentially infinite time horizon, ending at an unbounded random terminal time—the time of ruin of the insurance company. In recent models which take multiple economic factors into account, the problems are high dimensional. For computing these risk measures, optimal control problems need to be solved numerically. A standard method for solving control problems is to derive the associated Hamilton–Jacobi–Bellman (HJB) equation—a semilinear (sometimes integro) partial differential equation (PDE)—and show that its (numerical) solution also solves the original control problem. In the case of infinite time horizon problems, these HJB equations are (degenerate) elliptic. In this paper we propose a novel deep neural network algorithm for semilinear (degenerate) elliptic PDEs associated to infinite time horizon control problems in high dimensions.
We apply this method to solve the dividend maximization problem in insurance mathematics. This problem originates in the seminal work by De Finetti (1957), who introduced expected discounted future dividend payments as a valuation principle for a homogeneous insurance portfolio. This constitutes an alternative risk measure to the (even more) classical probability of ruin. Classical results on the dividend maximization problem are Asmussen and Taksar (1997); Jeanblanc-Picqué and Shiryaev (1995); Radner and Shepp (1996); Shreve et al. (1984). Overviews can be found in Albrecher and Thonhauser (2009); Avanzi (2009); for an introduction to optimization problems in insurance we refer to Azcue and Muler (2014); Schmidli (2008). Recent models for the surplus of an insurance company allow for changes in the underlying economy. Such models have been studied, e.g., in Jiang and Pistorius (2012); Leobacher et al. (2014); Sotomayor and Cadenillas (2011); Zhu and Chen (2013); Szölgyenyi (2013, 2016); Reppen et al. (2020). In Leobacher et al. (2014); Szölgyenyi (2013, 2016) hidden Markov models for the underlying economic environment were proposed that allow for taking (multiple) exogenous, even not directly observable, economic factors into account. While these authors study the dividend maximization problem from a theoretical perspective, we are interested in computing the risk measure. However, classical numerical methods fail when the problem becomes high-dimensional, that is for example when exogenous economic factors are taken into account. In this paper we propose a novel deep neural network algorithm to solve high-dimensional problems. As an application we use it to solve the dividend maximization problem in the model from Szölgyenyi (2016) in high dimensions numerically.
Classical algorithms for solving semilinear (degenerate) elliptic PDEs like finite difference or finite element methods suffer from the so-called curse of dimensionality—the computational complexity for solving the discretized equation grows exponentially in the dimension. In high-dimensions (say > 10 ) one has to resort to costly quadrature methods such as multilevel-Monte Carlo or the quasi-Monte Carlo-based method presented in Kritzer et al. (2019). In recent years, deep neural network (DNN) algorithms for high-dimensional PDEs have been studied extensively. Prominent examples are Han et al. (2017, 2018), where semilinear parabolic PDEs are associated with backward stochastic differential equations (BSDEs) through the (nonlinear) Feynman-Kac formula and a DNN algorithm is proposed that solves these PDEs by solving the associated BSDEs. In the literature there exists a variety of DNN approaches for solving PDEs, in particular (degenerate) parabolic ones. Great literature overviews are given, e.g., in Beck et al. (2020); Grohs et al. (2019a), out of which we list some contributions here: (Beck et al. 2018, 2019b, 2019c; Becker et al. 2019a, 2019b; Berg and Nyström 2018; Chan-Wai-Nam et al. 2019; Chen and Wan 2020; Dockhorn 2019; E and Yu 2018; Farahmand et al. 2017; Fujii et al. 2019; Goudenѐge et al. 2019; Han and Long 2020; Han et al. 2020; Henry-Labordѐre 2017; Huré et al. 2019; Jacquier and Oumgari 2019; Long et al. 2018; Lu et al. 2019; Lye et al. 2020; Magill et al. 2018; Pham et al. 2019; Raissi 2018; Sirignano and Spiliopoulos 2018).
While mathematical finance control problems (e.g., investment problems) are studied over relatively short time horizons, leading to (degenerate) parabolic PDEs, in insurance mathematics they are often considered over the whole lifetime of the insurance company, leading to (degenerate) elliptic PDEs. For elliptic PDEs, a multilevel Picard iteration algorithm is studied in Beck et al. (2020), a derivative-free method using Brownian walkers without explicit calculation of the derivatives of the neural network is studied in Han et al. (2020), and a walk-on-the-sphere algorithm is introduced in Grohs and Herrmann (2020) for the Poisson equation, where the existence of DNNs that are able to approximate the solution to certain elliptic PDEs is shown.
In the present article we propose a novel DNN algorithm for a large class of semilinear (degenerate) elliptic PDEs. For this, we adopt the approach from Han et al. (2018) for (degenerate) parabolic PDEs. The difference here is that we use the correspondence between the PDEs we seek to solve and BSDEs with random terminal time. This correspondence was first presented in Pardoux (1998), and elaborated afterwards, e.g., in Briand and Hu (1998); Briand and Confortola (2008); Darling and Pardoux (1997); Pardoux (1999); Royer (2004).
As these results are not as standard as the BSDE correspondence to parabolic PDEs, we summarize the theory in Section 2 for the convenience of the reader. In Section 3 we present the DNN algorithm, and test it in Section 4.1 and Section 4.2. In Section 4.3 we present the model from Szölgyenyi (2016) in which we seek to solve the dividend maximization problem and hence to compute the risk measure. That this method works also in high dimensions is demonstrated at the end of Section 4.3, where numerical results are presented.
The method presented here can be applied to many other high-dimensional semilinear (degenerate) elliptic PDE problems in insurance mathematics, such as the calculation of ruin probabilities, but we emphasize that its application possibilities are not limited to insurance problems.

2. BSDEs Associated with Elliptic PDEs

This section contains a short survey on scalar backward stochastic differential equations with random terminal times and on how they are related to a certain type of semilinear elliptic partial differential equations.

2.1. BSDEs with Random Terminal Times

Let ( Ω , F , P , ( F t ) t [ 0 , ) ) be a filtered probability space satisfying the usual conditions and let W = ( W t ) t [ 0 , ) be a d-dimensional standard Brownian motion on it. We assume that ( F t ) t [ 0 , ) is equal to the augmented natural filtration generated by W. For all real valued row or column vectors x, let | x | denote their Euclidean norm. We need the following notations and definitions for BSDEs.
Definition 1.
A BSDE with random terminal time is a triple ( τ , ξ , f), where
  • the terminal time τ : Ω [ 0 , ] is an ( F t ) t [ 0 , ) -stopping time,
  • the generator f: Ω × [ 0 , ) × R × R 1 × d R is a process which satisfies that for all ( y , z ) R × R 1 × d , the process t f ( t , y , z ) is progressively measurable,
  • the terminal condition ξ : Ω R is an F τ -measurable random variable with ξ = 0 on { τ = } .
Definition 2.
A solution to the BSDE ( τ , ξ , f) is a pair of progressively measurable processes ( Y , Z ) = ( Y t ) t [ 0 , ) , ( Z t ) t [ 0 , ) with values in R × R 1 × d , where
  • Y is continuous P -a.s. and for all T ( 0 , ) , the trajectories t Z t belong to L 2 ( [ 0 , T ] , R 1 × d ) , and  t f ( t , Y t , Z t ) is in L 1 ( [ 0 , T ] ) ,
  • for all T ( 0 , ) and all t [ 0 , T ] it holds a.s. that
    Y t = Y T + t τ T τ f ( s , Y s , Z s ) d s t τ T τ Z s d W s ,
  • Y t = ξ and Z t = 0 on { t τ } .
Results on existence of solutions of BSDEs with random terminal time can be found in Pardoux’ seminal article Pardoux (1998) (see (Pardoux 1998, Theorem 3.2)), in Briand and Confortola (2008) for generators with quadratic growth (see (Briand and Confortola 2008, Theorem 3.3)), and, e.g., in Briand and Hu (1998); Briand et al. (2003); Darling and Pardoux (1997); Pardoux (1999); Royer (2004); many of them cover multidimensional state spaces for the Y-process.
Optimal control problems which can be treated using a BSDE setting have for example been studied in (Briand and Confortola 2008, Section 6). For this they consider generators of the forward-backward form
f ( t , y , z ) = F ( X t , y , z ) = inf { g ( X t , u ) + z r ( X t , u ) : u U } λ y ,
where X is a forward diffusion (see also the notation in the following subsection), U is a Banach space, r is a Hilbert space-valued function (in their setting z takes values in the according dual space) with linear growth, g a real valued function with quadratic growth in u, and λ ( 0 , ) . In the sequel we focus on generators of forward-backward form.

2.2. Semilinear Elliptic PDEs and BSDEs with Random Terminal Time

In this subsection we recall the connection between semilinear elliptic PDEs and BSDEs with random (and possibly infinite) terminal time. The relationship between the theories is based on a nonlinear extension of the Feynman-Kac formula, see (Pardoux 1998, Section 4).
We define the forward process X as the stochastic process satisfying a.s.,
X t = x + 0 t μ ( X s ) d s + 0 t σ ( X s ) d W s , t [ 0 , ) ,
where x R d and μ : R d R d and σ : R d R d × d are globally Lipschitz functions.
In this paper we consider the following class of PDEs.
Definition 3.
  • A semilinear (degenerate) elliptic PDE on the whole R d is of the form
    L u + F ( · , u , ( u ) σ ) = 0 ,
    where the differential operator L acting on C 2 ( R d ) is given by
    L : = 1 2 i , j = 1 d ( σ σ ) i , j ( x ) 2 x i x j + i = 1 d μ i ( x ) x i ,
    and F : R d × R × R 1 × d R is such that the process ( t , y , z ) F ( X t , y , z ) is a generator of a BSDE in the sense of Definition 1.
  • We say that a function u satisfies Equation (4) with Dirichlet boundary conditions on the open, bounded domain G R d , if
    L u + F ( · , u , ( u ) σ ) = 0 , x G , u ( x ) = g ( x ) , x G ,
    where g : R d R is a bounded, continuous function.
Definition 4.
1. 
A BSDE associated to the PDE (4) on the whole R d is given by the triplet ( τ , ξ , f), where  τ , ξ = 0 , f( t , y , z ) = F ( X t , y , z ) , X is as in (3), and the solution satisfies a.s. for all T ( 0 , ) that
Y t = Y T + t T F ( X s , Y s , Z s ) d s t T Z s d W s , t [ 0 , T ] .
2. 
A BSDE associated to the PDE (6) with Dirichlet boundary conditions is given by the triplet ( τ , g ( X τ ) , f), where τ = inf { t [ 0 , ) : X t G ¯ } , f( t , y , z ) = F ( X t , y , z ) , X is as in (3), and the solution satisfies a.s. for all T ( 0 , ) that
Y t = Y T + t τ T τ F ( X s , Y s , Z s ) d s t τ T τ Z s d W s , t [ 0 , T ] , Y t = g ( X τ ) , Z t = 0 , t τ .
In order to keep the notation simple, we do not highlight the dependence of X , Y , Z on x.
For later use we also introduce the following notion of solutions of PDEs, which we will use later.
Definition 5.
  • A function u C ( R d ) is called viscosity subsolution of (4), if for all φ C 2 ( R d ) and all points x R d where u φ has a local maximum,
    L φ ( x ) + F ( x , u ( x ) , ( φ ( x ) ) σ ( x ) ) 0 .
  • A function u C ( R d ) is called viscosity supersolution of (4), if for all φ C 2 ( R d ) and all points x R d where u φ has a local minimum,
    L φ ( x ) + F ( x , u ( x ) , ( φ ( x ) ) σ ( x ) ) 0 .
  • A function u C ( R d ) is called viscosity solution of (4), if it is a viscosity sub- and supersolution.
A similar definition of viscosity solutions can be given for the case of Dirichlet boundary conditions (6), see Pardoux (1998).
For later use, note that (8) can be rewritten in forward form as
Y t = Y 0 0 t F ( X s , Y s , Z s ) d s + 0 t Z s d W s , t [ 0 , τ ) , Y t = g ( X τ ) , Z t = 0 , t τ .
The following theorems link the semilinear elliptic PDEs (4) and (6) to the associated BSDEs.
Theorem 1
((Pardoux 1998, Theorem 4.1)). Let ( t , y , z ) F ( X t , y , z ) meet the assumptions of (Pardoux 1998, Theorem 3.2) and let u C 2 ( R d ) satisfy
E 0 e λ t | ( ( u ) σ ) ( X t ) | 2 d t <
with λ as in (Pardoux 1998, Theorem 3.2). If u is a classical solution of the PDE (4), then
Y t = u ( X t ) , Z t = ( ( u ) σ ) ( X t )
solve the BSDE (7). An equivalent statement can be established for the system with boundary conditions (6) and Equation (8), see Pardoux (1998).
Note that for all x R d , Y and Z are stochastic processes adapted to ( F t ) t [ 0 , ) . Therefore, Y 0 , Z 0 are F 0 -measurable and hence a.s. deterministic. For us, the connection between PDEs and BSDEs is of relevance because of the converse result, where x Y 0 delivers a solution to the respective PDE.
Theorem 2
((Pardoux 1998, Theorem 4.3)). Assume that for some K , K , p ( 0 , ) , γ ( , 0 ) the function F satisfies for all x , y , y , z , z ,
(i) 
| F ( x , y , z ) | K ( 1 + | x | p + | y | + | z | ) ,
(ii) 
y y , F ( x , y , z ) F ( x , y , z ) γ | y y | 2 ,
(iii) 
| F ( x , y , z ) F ( x , y , z ) | K | z z | .
Then (Pardoux 1998, Theorem 3.2) can be applied to the generator ( t , y , z ) F ( X t , y , z ) , showing that the function u given by u ( x ) = Y 0 is a viscosity solution to (4), where Y is the first component of the unique solution to (7) in the class of solutions from (Pardoux 1998, Theorem 3.2).
The case of the Dirichlet problem requires additional assumptions on the domain G and the exit time τ from (8). We refer to (Pardoux 1998, Theorem 4.3). A corresponding result for BSDEs with quadratic generator is (Briand and Confortola 2008, Theorem 5.2).
To conclude, the correspondence between PDE (6) and BSDE (8) is given by Y t = u ( X t ) , Z t = ( ( u ) σ ) ( X t ) , ξ = g ( X τ ) . For tackling elliptic PDEs which are degenerate (as it is the case for our insurance mathematics example) we need to take the relationship a little further in order to escape the not so convenient structure of the Z-process. We factor Z σ ( X ) = Z for cases where this equation is solvable for Z ( σ needs not necessarily be invertible) and define f ( x , y , ζ ) : = F ( x , y , ζ σ ( x ) ) 1, giving the correspondence Y t = u ( X t ) , Z t = u ( X t ) , ξ = g ( X τ ) . This relationship motivates us to solve semilinear degenerate elliptic PDEs by solving the corresponding BSDEs forward in time (cf. (9))
Y t = Y 0 0 t f ( X s , Y s , Z s ) d s + 0 t Z s σ ( X s ) d W s , t [ 0 , τ )
for Y 0 by approximating the paths of Z = u ( X ) by a DNN, see Section 3. Doing so, we obtain an estimate of a solution value u ( x ) for a given x R d .

3. Algorithm

The idea of the proposed algorithm is inspired by Han et al. (2018), where the authors use the correspondence between BSDEs and semilinear parabolic PDEs to construct a DNN algorithm for solving the latter. In the same spirit, we construct a DNN algorithm based on the correspondence to BSDEs with random terminal time for solving semilinear elliptic PDEs.
The details of the algorithm are described in three steps of increasing specificity. First we explain the DNN algorithm mathematically. This is done below. Second, Algorithm 1 at the end of this section provides a pseudocode. Third, our program code is provided on Github2 under a creative commons license. The algorithm is implemented in a generic manner so that it can be reused for other elliptic PDE problems.
The goal of the algorithm is to calculate solution values u ( x ) of the semilinear (degenerate) elliptic PDE of interest. For the construction of the algorithm we use the correspondence to a BSDE with random terminal time. Recall from Section 2 that such a BSDE is given by a triplet ( τ , g ( X τ ) , f ) that can be determined from the given PDE and by
X t = x + 0 t μ ( X s ) d s + 0 t σ ( X s ) d W s
and
Y t = Y 0 0 t f ( X s , Y s , Z s ) d s + 0 t Z s σ ( X s ) d W s .
Furthermore, recall that we have identified Y t = u ( X t ) , where u is the solution of the PDE we are interested in.
The first step for calculating u is to approximate (10) up to the stopping time τ . To make this computationally feasible, we choose T large and stop at τ T , hence at T at the latest. Now let 0 = t 0 < t 1 < < t N = T , Δ t n = t n + 1 t n . We simulate M paths ω 1 , , ω M of the Brownian motion W. With this we approximate the forward process using the Euler–Maruyama scheme, that is X 0 = x and
X t n + 1 X t n + μ ( X t n ) Δ t n + σ ( X t n ) Δ W n .
In the next step we compute Z . For all t n , Z t n = u ( X t n ) are approximated by DNNs, each mapping G to R d . As noted above, the implementation of this (and all other steps) is provided.
Now, we initialize u ( x ) and use the above approximations to compute the solution to the BSDE forward in time by approximating (11):
u ( X t n + 1 ) u ( X t n ) 𝟙 ( 0 , τ ) ( t n ) f X t n , u ( X t n ) , u ( X t n ) Δ t n + 𝟙 ( 0 , τ ) ( t n ) u ( X t n ) σ ( X t n ) Δ W n .
Note that due to this construction, indirectly u ( X t n + 1 ) is also approximated by a DNN as a combination of DNNs.
For the training of the involved DNNs, we compare u ( X τ T ) with the terminal value ξ . This defines the loss function for the training:
1 M k = 1 M | u ( X τ T ( ω k ) ) ξ ( ω k ) | 2 .
After a certain number of training epochs the loss function is minimized and we obtain an approximate solution value u ( x ) of the PDE.
Remark 1.
Several approximation errors arise in the proposed algorithm:
1. 
the approximation error of the Euler–Maruyama method, which is used for sampling the forward equation,
2. 
the error of approximating the expected loss,
3. 
the error of cutting off the potentially unbounded random terminal time at time T,
4. 
the approximation error of the deep neural network model for approximating Z t n for each t n .
It is well known that for any continuous function we can find DNNs that approximate the function arbitrarily well, see Hornik (1991); Hornik et al. (1989). This is, however, not sufficient to make any statement about the approximation quality. Results on convergence rates are required. Though this question is already studied in the literature (see, e.g., Beck et al. 2019a, 2020; Berner et al. 2020; Elbrächter et al. 2018; Gonon et al. 2019; Grohs et al. 2018, 2019a, 2019b; Hutzenthaler et al. 2018, 2020a, 2020b; Jentzen et al. 2018; Kutyniok et al. 2019; Reisinger and Zhang 2019), results on convergence rates for given constructions are yet scarce and hence many questions remain open while the number of proposed DNN algorithms grows.
We close this section with some comments on the implementation.
Remark 2.
  • All DNNs are initialized with random numbers.
  • For each value of x we average u ( x ) over 5 independent runs. The estimator for u ( x ) is calculated as the mean value of u ( x ) in the last 3 network training epochs of each run, sampled according to the validation size (see below).
  • We choose a nonequidistant time grid in order to get a higher resolution for earlier (and hence probably closer to the stopping time) time points.
  • We use tanh as activation function.
  • We compute u ( x ) simultaneously for 8 values of x by using parallel computing.
Algorithm 1 Elliptic PDE Solver for a BSDE ( f , ξ ) with stopping time τ
Require: number of training epochs E, maximal time T, step-size Δ t , number of timesteps N, number of sample paths M, number of hidden layer neurons dim, initial (random) starting values ( θ 0 ( u ) , θ 0 ( ζ ) )  
    1: function TrainableVariables( dim , θ )       ▹ see Pytorch or Tensorflow 
      return a trainable variable with dimension 1 × dim initialized by θ .  
    2: end function 
    3: function Subnetwork(x)     ▹ allowing x to be a tensor containing M rows (samples)   
     return a trainable DNN, evaluated at x.  
    4: end function 
    5: for i = 0 , , N do  
    6:  t i = timesteps ( i )           ▹ Initialize nonequidistant timesteps   
    7: end for 
    8: for   j = 1 , , M do  
    9: Sample Brownian motion trajectory w t i ( j ) 0 i N  
  10: Sample path from forward process x t i ( j ) 0 i N  
  11: calculate stopping time τ ( j )  
  12: calculate terminal value ξ ( j )  
  13: set x t ( j ) = x τ ( j ) ( j ) for all t > τ ( j )  
  14: end for 
  15: u 0 = TrainableVariables ( 1 , θ 0 ( u ) )              ▹ Initialize u 
  16: u 0 = TrainableVariables ( d , θ 0 ( ζ ) )            ▹ Initialize Z  
  17: for j = 1 , , M do  
  18:   u ( j ) = u 0  
  19:   u ( j ) = u 0  
  20: end for   
  21: for e = 1 , , E do  
  22:  for i = 1 , , N 1 do  
  23:   for j = 1 , , M do    
  24:     u ( j ) = u ( j ) f ( x t i ( j ) , u ( j ) , u ( j ) ) ( t i + 1 t i ) + u ( j ) σ ( x t i ( j ) ) ( w t i + 1 ( j ) w t i ( j ) )  
  25:    if t i + 1 > τ ( j ) then    break  
  26:    end if   
  27:   end for   
  28:          u = Subnetwork ( x t i + 1 )    
  29: end for   
  30: update all trainable variables and the subnetwork’s weights according to the loss function  
1 M j = 1 M ( u ( j ) ξ ( j ) ) 2
 
  31: end for 
   return ( u 0 , u 0 )

4. Examples

In this section we apply the proposed algorithm to three examples. The first one serves as a validity check, the second one as an academic example with a nonlinearity. Finally, we apply the algorithm to solve the dividend maximization problem under incomplete information.

4.1. The Poisson Equation

The first example we study is the Poisson equation—a linear PDE.
Let r ( 0 , ) , G = x R d : | x | < r , G = x R d : | x | = r , b R , and
Δ u ( x ) = b , x G , u ( x ) = 0 , x G .
Solving (14) is equivalent to solving the BSDE with
d X t = 2 d W t , X 0 = x , f ( x , y , ζ ) = b , ξ = 0 ,
up to the stopping time τ = inf { t [ 0 , T ] : | x | > r } .
To obtain a reference solution for this linear BSDE we use an analytic formula for the expectation of τ , see (Øksendal 2003, Example 7.4.2, p. 121). This yields
u ( x ) = b 2 d r 2 | x | 2 .

Numerical Results

We compute u ( x ) on the R 2 and the R 100 for 15 different values of x. Figure 1 shows the approximate solution of u obtained by the DNN algorithm on the diagonal points { ( x , , x ) R d : x [ r , r ] } (in blue) and the analytical reference solution (in green). Table 1 contains the parameters we use.
Note that as the expected value of τ decreases linearly in d, we adapt the cut off time T for d = 100 accordingly.

4.2. Quadratic Gradient

The second example is a semilinear PDE with a quadratic gradient term.
Let r ( 0 , ) , G = x R d : | x | < r , and G = x R d : | x | = r . We consider the PDE
Δ u ( x ) + | u ( x ) | 2 = 2 e u ( x ) , x G , u ( x ) = log r 2 + 1 d , x G .
corresponding to the BSDE
d X t = 2 d W t , X 0 = x ,
f ( x , y , ζ ) = | ζ | 2 2 e y , ξ = log | r | 2 + 1 d .
In addition, for this example we have an analytic reference solution given by
u ( x ) = log | x | 2 + 1 d .

Numerical Results

As in the previous example we compute u ( x ) for 15 different values of x on the R 2 and the R 100 . Figure 2 shows the approximate solution of u obtained by the DNN algorithm on the diagonal points { ( x , , x ) R d : x [ r , r ] } (in blue) and the analytical reference solution (in green). Table 2 contains the parameters we use.
While classical numerical methods for PDEs would be a much better choice in the case d = 2 , their application would not be feasible in the case d = 100 .

4.3. Dividend Maximization

The goal of this paper is to show how to use the proposed DNN algorithm to solve high-dimensional control problems that arise in insurance mathematics. We finally arrived at the point where we are ready to do so.
Our example comes from Szölgyenyi (2016), where the author studies De Finetti’s dividend maximization problem in a setup with incomplete information about the current state of the market. The hidden market-state process determines the trend of the surplus process of the insurance company and is modeled as a d-state Markov chain. Using stochastic filtering, in Szölgyenyi (2016) they achieve to transform the one-dimensional problem under incomplete information to a d-dimensional problem under complete information. The cost is ( d 1 ) additional dimensions in the state space. We state the problem under complete information using different notation than in Szölgyenyi (2016) in order to avoid ambiguities.
The probability that the Markov chain modeling the market-state is in state i { 1 , , d 1 } is given by
π i ( t ) = x i + 0 t q d , i + j = 1 d 1 ( q j , i q d , i ) π j ( s ) d s + 0 t π i ( s ) a i ν s ρ d B s ,
where
ν t = a d + j = 1 d 1 ( a j a d ) π j ( t ) ,
x i ( 0 , 1 ) , B is a one-dimensional Brownian motion, a 1 , , a d R are the values of the surplus trend in the respective market-states of the hidden Markov chain, and ( q i , j ) i , j { 1 , , d } R d × d denotes the intensity matrix of the chain.
Let ( t ) t [ 0 , ) be the dividend rate process. The surplus of the insurance company is given by
X ˜ t d = x d + 0 t ( ν s s ) d s + ρ B t , t [ 0 , ) ,
where x d , ρ ( 0 , ) . For later use we define also the dividend free surplus
X t d = x d + 0 t ν s d s + ρ B t , t [ 0 , ) .
The processes (18) and (20) form the d-dimensional state process underlying the optimal control problem we aim to solve.
The goal of the insurance company is to determine its value by maximizing the discounted dividends payments until the time of ruin η = inf { t ( 0 , ] : X ˜ t d < 0 } , that is it seeks to find
u ( x 1 , , x d ) = sup ( t ) t [ 0 , ) A E x 1 , , x d 0 η e δ t t d t ,
where δ ( 0 , ) is a discount rate, A is the set of admissible controls, and E x 1 , , x d [ · ] denotes the expectation under the initial conditions π i ( 0 ) = x i for i { 1 , , d 1 } and X ˜ 0 d = x d . Admissible controls are ( F t X d ) t 0 -progressively measurable, [ 0 , K ] -valued for K ( 0 , ) and fulfill t 0 for t > η , cf. Szölgyenyi (2016).
In order to tackle the problem, we solve the corresponding Hamilton–Jacobi–Bellmann (HJB) equation4 from Szölgyenyi (2016),
( L δ ) u + sup [ 0 , K ] ( ( 1 u x d ) ) = 0 ,
where L is the second order degenerate elliptic operator
L u = a d u x d + i = 1 d 1 ( a i a d ) x i u x d + q d i + j = 1 d 1 ( q j i q d i ) x i u x i + x i a i ν u x d x i + 1 2 j = 1 d 1 x i a i ν ρ x j a j ν ρ u x i x j + 1 2 ρ 2 u x d x d .
The supremum in (23) is attained at
= K , u x d 1 0 , u x d > 1 .
Plugging this into (23) we end up with a d-dimensional semilinear degenerate elliptic PDE:
( L δ ) u + K ( 1 u x d ) 𝟙 { u x d 1 } = 0 .
The boundary conditions in x d direction are given by
u ( x 1 , , x d ) = K / δ , x d 0 , x d = 0 .
No boundary conditions are required for the other variables, cf. Szölgyenyi (2016).
In (Szölgyenyi 2016, Corollary 3.6) it is proven that the unique viscosity solution to (24) solves the optimal control problem (22). Hence, we can indeed solve the control problem by solving the HJB equation.
For the numerical approximation we cut off x d at r ( 0 , ) . Hence, G = x R d : 0 < x d < r and G = x R d : x d { 0 , r } .
For the convenience of the reader we derive the BSDE corresponding to (24). The forward equation is given by
d X t = ( d π 1 ( t ) , , d π d 1 ( t ) , d X t d ) , X 0 = x ,
that is
d X t = μ ( X t ) d t + σ ( X t ) d W t , X 0 = x ,
where W = ( B , W 2 , , W d ) , x = ( x 1 , , x d ) , and
μ ( x ) = q d , 1 + j = 1 d 1 ( q j , 1 q d , 1 ) x j , , q d , d 1 + j = 1 d 1 ( q j , d 1 q d , d 1 ) x d 1 , a d + j = 1 d 1 ( a j a d ) x j ,
σ ( x ) = x 1 a 1 a d + j = 1 d 1 ( a j a d ) x j ρ 0 0 0 0 x d 1 a d 1 a d + j = 1 d 1 ( a j a d ) x j ρ 0 0 ρ 0 0 .
We claim that the BSDE associated to (24) is given in forward form by
u ( X t ) = u ( x ) 0 t [ K ( 1 u x d ( X s ) ) 𝟙 { u x d ( X s ) 1 } δ u ( X s ) ] d t + 0 t u ( X s ) σ ( X s ) d W s .
Applying Itô’s formula to u ( X ) yields
u ( X t ) = u ( x ) + 0 t L u ( X s ) d t + 0 t u ( X s ) σ ( X s ) d W s .
Combining (25) and (26) gives
u ( x ) 0 t [ K ( 1 u x d ( X s ) ) 𝟙 { u x d ( X s ) 1 } δ u ( X s ) ] d t + 0 t u ( X s ) σ ( X s ) d W s = u ( x ) + 0 t L u ( X s ) d t + 0 t u ( X s ) σ ( X s ) d W s .
Canceling terms verifies (in a heuristic manner) (24).
Hence, the corresponding BSDE has the parameters
f ( x , y , ζ ) = K ( 1 ζ d ) 𝟙 { ζ d 1 } δ y
and
ξ = K / δ , X τ d = r , 0 , X τ d = 0 ,
if τ < .

Numerical Results

As for this example we have no analytic reference solution at hand, we use the solution from Szölgyenyi (2016) for the case d = 2 , which was obtained by a finite difference method and policy iteration. Then we show that the DNN algorithm also provides an approximation in high dimensions in reasonable computation time.
As in the previous examples we compute u ( x ) on the R 2 and on the R 100 for 15 different values of x. Figure 3 shows the approximate solution of the HJB equation and hence the value of the insurance company obtained by the DNN algorithm (in blue) and the reference solution from Szölgyenyi (2016) (in green) for the case d = 2 . For d = 100 we have no reference solution at hand. Figure 4 shows the loss for a fixed value of x in the case d = 100 . Table 3 and Table 4 contain the parameters we use.

5. Conclusions

The goal of this paper was to compute the risk measure given by the expected discounted future dividend payments in a complex high-dimensional economic environment. This demonstrates the effectiveness of using DNN algorithms for solving some high-dimensional PDE problems in insurance mathematics that cannot be solved by classical methods. In the literature the focus so far was on parabolic PDE problems; however, in insurance mathematics we often face problems up to an unbounded random terminal time, e.g., the time of ruin of the insurance company, leading to (degenerate) elliptic problems.
Hence, we have proposed a novel deep neural network algorithm for a large class of semilinear (degenerate) elliptic PDEs associated to infinite time horizon control problems in high dimensions. The method extends the DNN algorithm proposed by Han, Jetzen, and E Han et al. (2018), which was developed for parabolic PDEs, to the case of (degenerate) elliptic semilinear PDEs. We have attacked the problem inspired by a series of results by Pardoux Pardoux (1998).
Of course, in low dimensions one would not use the proposed DNN algorithm—classical methods are more efficient. However, recent models are frequently high dimensional, in which case classical methods fail due to the curse of dimensionality. Then the DNN algorithm presented here can be applied to compute the desired quantity.
We emphasize that the method presented here can also be applied to many other high-dimensional semilinear (degenerate) elliptic PDE problems in insurance mathematics and beyond.
An implementation of the algorithm is provided on Github5 under a creative commons license.

Author Contributions

Writing–original draft, S.K., A.S. and M.S. All authors have read and agreed to the published version of the manuscript.

Funding

Stefan Kremsner was supported by the Austrian Science Fund (FWF): Project F5508-N26, which is part of the Special Research Program “Quasi-Monte Carlo Methods: Theory and Applications”.

Acknowledgments

The authors thank Gunther Leobacher for discussions and suggestions that improved the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Albrecher, Hansjörg, and Stefan Thonhauser. 2009. Optimality results for dividend problems in insurance. RACSAM Revista de la Real Academia de Ciencias Exactas, Fisicas y Naturales. Serie A. Matematicas 103: 295–320. [Google Scholar] [CrossRef] [Green Version]
  2. Asmussen, Søren, and Michael Taksar. 1997. Controlled diffusion models for optimal dividend pay-out. Insurance: Mathematics and Economics 20: 1–15. [Google Scholar] [CrossRef]
  3. Avanzi, Benjamin. 2009. Strategies for dividend distribution: A review. North American Actuarial Journal 13: 217–51. [Google Scholar] [CrossRef]
  4. Azcue, Pablo, and Nora Muler. 2014. Stochastic Optimization in Insurance—A Dynamic Programming Approach. Springer Briefs in Quantitative Finance. New York, Heidelberg, Dordrecht and London: Springer. [Google Scholar]
  5. Beck, Christian, Sebastian Becker, Philipp Grohs, Nor Jaafari, and Arnulf Jentzen. 2018. Solving stochastic differential equations and Kolmogorov equations by means of deep learning. arXiv arXiv:1806.00421. [Google Scholar]
  6. Beck, Christian, Fabian Hornung, Martin Hutzenthaler, Arnulf Jentzen, and Thomas Kruse. 2019a. Overcoming the curse of dimensionality in the numerical approximation of Allen-Cahn partial differential equations via truncated full-history recursive multilevel Picard approximations. arXiv arXiv:1907.06729. [Google Scholar]
  7. Beck, Christian, Sebastian Becker, Patrick Cheridito, Arnulf Jentzen, and Ariel Neufeld. 2019b. Deep splitting method for parabolic PDEs. arXiv arXiv:1907.03452. [Google Scholar]
  8. Beck, Christian, Weinan E, and Arnulf Jentzen. 2019c. Machine learning approximation algorithms for high-dimensional fully nonlinear partial differential equations and second-order backward stochastic differential equations. Journal of Nonlinear Science 29: 1563–619. [Google Scholar] [CrossRef] [Green Version]
  9. Beck, Christian, Lukas Gonon, and Arnulf Jentzen. 2020. Overcoming the curse of dimensionality in the numerical approximation of high-dimensional semilinear elliptic partial differential equations. arXiv arXiv:2003.00596. [Google Scholar]
  10. Becker, Sebastian, Patrick Cheridito, and Arnulf Jentzen. 2019a. Deep optimal stopping. Journal of Machine Learning Research 20: 74. [Google Scholar]
  11. Becker, Sebastian, Patrick Cheridito, Arnulf Jentzen, and Timo Welti. 2019b. Solving high-dimensional optimal stopping problems using deep learning. arXiv arXiv:1908.01602. [Google Scholar]
  12. Berg, Jens, and Kaj Nyström. 2018. A unified deep artificial neural network approach to partial differential equations in complex geometries. Neurocomputing 317: 28–41. [Google Scholar] [CrossRef] [Green Version]
  13. Berner, Julius, Philipp Grohs, and Arnulf Jentzen. 2020. Analysis of the generalization error: Empirical risk minimization over deep artificial neural networks overcomes the curse of dimensionality in the numerical approximation of Black–Scholes partial differential equations. SIAM Journal on Mathematics of Data Science 3: 631–57. [Google Scholar] [CrossRef]
  14. Briand, Philippe, and Ying Hu. 1998. Stability of BSDEs with random terminal time and homogenization of semi-linear elliptic PDEs. Journal of Functional Analysis 155: 455–94. [Google Scholar] [CrossRef] [Green Version]
  15. Briand, Philippe, Bernard Delyon, Ying Hu, Étienne Pardoux, and Lucretiu Stoica. 2003. Lp solutions of backward stochastic differential equations. Stochastic Processes and their Applications 108: 109–29. [Google Scholar] [CrossRef] [Green Version]
  16. Briand, Philippe, and Fulvia Confortola. 2008. Quadratic BSDEs with random terminal time and elliptic PDEs in infinite dimension. Electronic Journal of Probability 13: 1529–61. [Google Scholar]
  17. Chan-Wai-Nam, Quentin, Joseph Mikael, and Xavier Warin. 2019. Machine learning for semi linear PDEs. Journal of Scientific Computing 79: 1667–712. [Google Scholar] [CrossRef] [Green Version]
  18. Chen, Yangang, and Justin W. L. Wan. 2020. Deep neural network framework based on backward stochastic differential equations for pricing and hedging american options in high dimensions. Quantitative Finance, 1–23. [Google Scholar] [CrossRef]
  19. Darling, Richard W. R., and Étienne Pardoux. 1997. Backwards SDE with random terminal time and applications to semilinear elliptic PDE. The Annals of Probability 25: 1135–59. [Google Scholar] [CrossRef]
  20. De Finetti, Bruno. 1957. Su un’impostazione alternativa della teoria collettiva del rischio. Transactions of the XVth International Congress of Actuaries 2: 433–43. [Google Scholar]
  21. Dockhorn, Tim. 2019. A discussion on solving partial differential equations using neural networks. arXiv arXiv:1904.07200. [Google Scholar]
  22. E, Weinan, and Bing Yu. 2018. The deep Ritz method: A deep learning-based numerical algorithm for solving variational problems. Communications in Mathematics and Statistics 6: 1–12. [Google Scholar] [CrossRef] [Green Version]
  23. Elbrächter, Dennis, Philipp Grohs, Arnulf Jentzen, and Christoph Schwab. 2018. DNN expression rate analysis of high-dimensional PDEs: Application to option pricing. arXiv arXiv:1809.07669. [Google Scholar]
  24. Farahmand, Amir-Massoud, Saleh Nabi, and Daniel Nikovski. 2017. Deep reinforcement learning for partial differential equation control. Paper presented at 2017 American Control Conference (ACC), Seattle, WA, USA, May 24–26; pp. 3120–27. [Google Scholar]
  25. Fujii, Masaaki, Akihiko Takahashi, and Masayuki Takahashi. 2019. Asymptotic expansion as prior knowledge in deep learning method for high dimensional BSDEs. Asia-Pacific Financial Markets 26: 391–408. [Google Scholar] [CrossRef] [Green Version]
  26. Gonon, Lukas, Philipp Grohs, Arnulf Jentzen, David Kofler, and David Šiška. 2019. Uniform error estimates for artificial neural network approximations. arXiv arXiv:1911.09647. [Google Scholar]
  27. Goudenège, Ludovic, Andrea Molent, and Antonino Zanette. 2019. Machine learning for pricing American options in high dimension. arXiv arXiv:1903.11275. [Google Scholar]
  28. Grohs, Philipp, Fabian Hornung, Arnulf Jentzen, and Philippe Von Wurstemberger. 2018. A proof that artificial neural networks overcome the curse of dimensionality in the numerical approximation of Black-Scholes partial differential equations. arXiv arXiv:1809.02362. [Google Scholar]
  29. Grohs, Philipp, Arnulf Jentzen, and Diyora Salimova. 2019a. Deep neural network approximations for Monte Carlo algorithms. arXiv arXiv:1908.10828. [Google Scholar]
  30. Grohs, Philipp, Fabian Hornung, Arnulf Jentzen, and Philipp Zimmermann. 2019b. Space-time error estimates for deep neural network approximations for differential equations. arXiv arXiv:1908.03833. [Google Scholar]
  31. Grohs, Philipp, and Lukas Herrmann. 2020. Deep neural network approximation for high-dimensional elliptic PDEs with boundary conditions. arXiv arXiv:2007.05384. [Google Scholar]
  32. Han, Jiequn, Arnulf Jentzen, and Weinan E. 2017. Deep learning-based numerical methods for high-dimensional parabolic partial differential equations and backward stochastic differential equations. Communications in Mathematics and Statistics 5: 349–80. [Google Scholar]
  33. Han, Jiequn, Arnulf Jentzen, and Weinan E. 2018. Solving high-dimensional partial differential equations using deep learning. Proceedings of the National Academy of Sciences 115: 8505–10. [Google Scholar] [CrossRef] [Green Version]
  34. Han, Jiequn, and Jihao Long. 2020. Convergence of the deep BSDE method for coupled FBSDEs. Probability, Uncertainty and Quantitative Risk 5: 1–33. [Google Scholar] [CrossRef]
  35. Han, Jihun, Mihai Nica, and Adam Stinchcombe. 2020. A derivative-free method for solving elliptic partial differential equations with deep neural networks. Journal of Computational Physics 419: 109672. [Google Scholar] [CrossRef]
  36. Henry-Labordère, Pierre. 2017. Deep primal-dual algorithm for BSDEs: Applications of machine learning to CVA and IM. SSRN Electronic Journal. [Google Scholar] [CrossRef]
  37. Hornik, Kurt. 1991. Approximation capabilities of multilayer feedforward networks. Neural Networks 4: 251–57. [Google Scholar] [CrossRef]
  38. Hornik, Kurt, Maxwell Stinchcombe, and Halbert White. 1989. Multilayer feedforward networks are universal approximators. Neural Networks 2: 359–66. [Google Scholar] [CrossRef]
  39. Huré, Côme, Huyên Pham, and Xavier Warin. 2019. Some machine learning schemes for high-dimensional nonlinear PDEs. arXiv arXiv:1902.01599. [Google Scholar]
  40. Hutzenthaler, Martin, Arnulf Jentzen, Thomas Kruse, Tuan Anh Nguyen, and Philippe von Wurstemberger. 2018. Overcoming the curse of dimensionality in the numerical approximation of semilinear parabolic partial differential equations. arXiv arXiv:1807.01212. [Google Scholar]
  41. Hutzenthaler, Martin, Arnulf Jentzen, and Philippe von Wurstemberger. 2020a. Overcoming the curse of dimensionality in the approximative pricing of financial derivatives with default risks. Electronic Journal of Probability 25: 73. [Google Scholar] [CrossRef]
  42. Hutzenthaler, Martin, Arnulf Jentzen, Thomas Kruse, and Tuan Anh Nguyen. 2020b. A proof that rectified deep neural networks overcome the curse of dimensionality in the numerical approximation of semilinear heat equations. SN Partial Differential Equations and Applications 1: 1–34. [Google Scholar] [CrossRef] [Green Version]
  43. Jacquier, Antoine Jack, and Mugad Oumgari. 2019. Deep PPDEs for rough local stochastic volatility. arXiv arXiv:1906.02551. [Google Scholar] [CrossRef] [Green Version]
  44. Jeanblanc-Picqué, Monique, and Albert Nikolaevich Shiryaev. 1995. Optimization of the flow of dividends. Russian Mathematical Surveys 50: 257–77. [Google Scholar] [CrossRef]
  45. Jentzen, Arnulf, Diyora Salimova, and Timo Welti. 2018. A proof that deep artificial neural networks overcome the curse of dimensionality in the numerical approximation of Kolmogorov partial differential equations with constant diffusion and nonlinear drift coefficients. arXiv arXiv:1809.07321. [Google Scholar]
  46. Jiang, Zhengjun, and Martijn Pistorius. 2012. Optimal dividend distribution under markov regime switching. Finance and Stochastics 16: 449–76. [Google Scholar] [CrossRef] [Green Version]
  47. Kritzer, Peter, Gunther Leobacher, Michaela Szölgyenyi, and Stefan Thonhauser. 2019. Approximation methods for piecewise deterministic markov processes and their costs. Scandinavian Actuarial Journal 2019: 308–35. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Kutyniok, Gitta, Philipp Petersen, Mones Raslan, and Reinhold Schneider. 2019. A theoretical analysis of deep neural networks and parametric PDEs. arXiv arXiv:1904.00377. [Google Scholar]
  49. Leobacher, Gunther, Michaela Szölgyenyi, and Stefan Thonhauser. 2014. Bayesian dividend optimization and finite time ruin probabilities. Stochastic Models 30: 216–49. [Google Scholar] [CrossRef] [Green Version]
  50. Long, Zichao, Yiping Lu, Xianzhong Ma, and Bin Dong. 2018. PDE-Net: Learning PDEs from Data. Paper presented at 35th International Conference on Machine Learning, Stockholm, Sweden, July 10–15; pp. 3208–16. [Google Scholar]
  51. Lu, Lu, Xuhui Meng, Zhiping Mao, and George Karniadakis. 2019. DeepXDE: A deep learning library for solving differential equations. arXiv arXiv:1907.04502. [Google Scholar]
  52. Lye, Kjetil O, Siddhartha Mishra, and Deep Ray. 2020. Deep learning observables in computational fluid dynamics. Journal of Computational Physics 410: 109339. [Google Scholar] [CrossRef] [Green Version]
  53. Magill, Martin, Faisal Qureshi, and Hendrick de Haan. 2018. Neural networks trained to solve differential equations learn general representations. Advances in Neural Information Processing Systems 31: 4071–81. [Google Scholar]
  54. Øksendal, Bernt. 2003. Stochastic Differential Equations. Berlin: Springer. [Google Scholar]
  55. Pardoux, Étienne. 1998. Backward stochastic differential equations and viscosity solutions of systems of semilinear parabolic and elliptic PDEs of second order. In Stochastic Analysis and Related Topics VI. Berlin: Springer, pp. 79–127. [Google Scholar]
  56. Pardoux, Étienne. 1999. BSDEs, weak convergence and homogenization of semilinear PDEs. In Nonlinear Analysis, Differential Equations and Control. Berlin: Springer, pp. 503–49. [Google Scholar]
  57. Pham, Huyen, Xavier Warin, and Maximilien Germain. 2019. Neural networks-based backward scheme for fully nonlinear PDEs. arXiv arXiv:1908.00412. [Google Scholar]
  58. Radner, Roy, and Larry Shepp. 1996. Risk vs. profit potential: A model for corporate strategy. Journal of Economic Dynamics and Control 20: 1373–93. [Google Scholar] [CrossRef] [Green Version]
  59. Raissi, Maziar. 2018. Deep hidden physics models: Deep learning of nonlinear partial differential equations. The Journal of Machine Learning Research 19: 932–55. [Google Scholar]
  60. Reisinger, Christoph, and Yufei Zhang. 2019. Rectified deep neural networks overcome the curse of dimensionality for nonsmooth value functions in zero-sum games of nonlinear stiff systems. arXiv arXiv:1903.06652. [Google Scholar]
  61. Reppen, A Max, Jean-Charles Rochet, and H. Mete Soner. 2020. Optimal dividend policies with random profitability. Mathematical Finance 30: 228–59. [Google Scholar] [CrossRef] [Green Version]
  62. Royer, Manuela. 2004. BSDEs with a random terminal time driven by a monotone generator and their links with PDEs. Stochastics and Stochastic Reports 76: 281–307. [Google Scholar] [CrossRef]
  63. Schmidli, Hanspeter. 2008. Stochastic Control in Insurance. Probability and its Applications. London: Springer. [Google Scholar]
  64. Shreve, Steven E., John P. Lehoczky, and Donald P. Gaver. 1984. Optimal consumption for general diffusions with absorbing and reflecting barriers. SIAM Journal on Control and Optimization 22: 55–75. [Google Scholar] [CrossRef]
  65. Sirignano, Justin, and Konstantinos Spiliopoulos. 2018. DGM: A deep learning algorithm for solving partial differential equations. Journal of Computational Physics 375: 1339–64. [Google Scholar] [CrossRef] [Green Version]
  66. Sotomayor, Luz R., and Abel Cadenillas. 2011. Classical and singular stochastic control for the optimal dividend policy when there is regime switching. Insurance: Mathematics and Economics 48: 344–54. [Google Scholar] [CrossRef]
  67. Szölgyenyi, Michaela. 2013. Bayesian dividend maximization: A jump diffusion model. In Handelingen Contactforum Actuarial and Financial Mathematics Conference, Interplay between Finance and Insurance, February 7–8. Brussel: Koninklijke Vlaamse Academie van België voor Wetenschappen en Kunsten, pp. 77–82. [Google Scholar]
  68. Szölgyenyi, Michaela. 2016. Dividend maximization in a hidden Markov switching model. Statistics & Risk Modeling 32: 143–58. [Google Scholar]
  69. Zhu, Jinxia, and Feng Chen. 2013. Dividend optimization for regime-switching general diffusions. Insurance: Mathematics and Economics 53: 439–56. [Google Scholar] [CrossRef]
1
Since Z σ ( X ) = Z is solvable for Z , f is well-defined.
2
3
The numerical examples were run on a Lenovo Thinkpad notebook with an Intel Core i7 processor (2.6 GHz) and 16 GB memory, without CUDA.
4
For abbreviation we use u x d for u x d etc.
5
Figure 1. Approximate solution (blue) and reference solution (green) for the Poisson equation on the R 2 (left) and on the R 100 (right).
Figure 1. Approximate solution (blue) and reference solution (green) for the Poisson equation on the R 2 (left) and on the R 100 (right).
Risks 08 00136 g001
Figure 2. Approximate solution (blue) and reference solution (green) for the equation with quadratic gradient on the R 2 (left) and on the R 100 (right).
Figure 2. Approximate solution (blue) and reference solution (green) for the equation with quadratic gradient on the R 2 (left) and on the R 100 (right).
Risks 08 00136 g002
Figure 3. Approximate solution (blue) and reference solution (green) for the dividend problem on the R 2 for fixed π 1 = π 2 = 0.5 (left) and on the R 100 for fixed π 1 = = π 100 = 0.01 (right, without reference solution).
Figure 3. Approximate solution (blue) and reference solution (green) for the dividend problem on the R 2 for fixed π 1 = π 2 = 0.5 (left) and on the R 100 for fixed π 1 = = π 100 = 0.01 (right, without reference solution).
Risks 08 00136 g003
Figure 4. Interpolated loss for the case d = 100 .
Figure 4. Interpolated loss for the case d = 100 .
Risks 08 00136 g004
Table 1. Parameters for the Poisson equation.
Table 1. Parameters for the Poisson equation.
drbNTEMValidation SizeTime per Eight Points3
20.5 0.75 5000.520064256119.17 s
1000.5 0.75 5000.0120064256613.86 s
Table 2. Parameters for the equation with quadratic gradient.
Table 2. Parameters for the equation with quadratic gradient.
drNTEMValidation SizeTime per Eight Points
21100550064256204.58 s
10011000.150064256321.13 s
Table 3. Parameters for the dividend problem.
Table 3. Parameters for the dividend problem.
drK δ ρ a i NTEMValidation SizeTime per Eight Points
251.80.51 2 i d 100550064256317.42 s
10051.80.51 2 i d 100550064256613.15 s
Table 4. Intensity matrix values for the dividend problem.
Table 4. Intensity matrix values for the dividend problem.
Case i = j Even i = j Odd i = j + 1 Even i = j + 1 3 Odd i = 1 , j = d Otherwise
q i , j 0.5 0.25 0.5 0.250.250
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kremsner, S.; Steinicke, A.; Szölgyenyi, M. A Deep Neural Network Algorithm for Semilinear Elliptic PDEs with Applications in Insurance Mathematics. Risks 2020, 8, 136. https://doi.org/10.3390/risks8040136

AMA Style

Kremsner S, Steinicke A, Szölgyenyi M. A Deep Neural Network Algorithm for Semilinear Elliptic PDEs with Applications in Insurance Mathematics. Risks. 2020; 8(4):136. https://doi.org/10.3390/risks8040136

Chicago/Turabian Style

Kremsner, Stefan, Alexander Steinicke, and Michaela Szölgyenyi. 2020. "A Deep Neural Network Algorithm for Semilinear Elliptic PDEs with Applications in Insurance Mathematics" Risks 8, no. 4: 136. https://doi.org/10.3390/risks8040136

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop