Abstract
This paper investigates algorithms for identifying the optimal policy for pricing American Options. The American Option pricing is reformulated as a Sequential Decision-Making problem with two binary actions (Exercise or Continue), transforming it into an optimal stopping time problem. Both the least square Monte Carlo simulation method (LSM) and Reinforcement Learning (RL)-based methods were utilized to find the optimal policy and, hence, the fair value of the American Put Option. Both Classical Geometric Brownian Motion (GBM) and calibrated Stochastic Volatility models served as the underlying uncertain assets. The novelty of this work lies in two aspects: (1) Applying LSM- and RL-based methods to determine option prices, with a specific focus on analyzing the dynamics of “Decisions” made by each method and comparing final decisions chosen by the LSM and RL methods. (2) Assess how the RL method updates “Decisions” at each batch, revealing the evolution of the decisions during the learning process to achieve optimal policy.
1. Introduction
An American Call/Put option is a contract issued by a financial institution allowing the holder (owner) of the American Option to exercise the option to buy (in the case of Call) or sell (in the case of Put) at a given strike price on or before the expiration date. These contract types are found in all major financial markets, including equity, commodity, insurance, energy, and real estate. The American Option differs from the European Option, where the European Option can only be exercised at the expiration date.
Having the “flexibility” to exercise the option at any time before the expiration date makes it more complex to value the American Option than the European Option. At each time step before its expiry when the American Option is in-the-money (the positive payoff upon exercise), the option owner faces the decision to either exercise the option and collect the payoff or consider “continuing” thinking that if she waits, the payoff might become more in-the-money (the bigger payoff). This dilemma (”When to Exercise?”) is the core challenge of the American Option problem.
At each time step, the option owner must compare the “immediate exercise value” against the “expected payoff from continuation”. Consequently, the valuation of the American Option hinges upon “conditional expectation”, dictating the decision between “exercising” or “continuing” the option. As will be demonstrated in the subsequent section, the fair value of the American Option can be conceptualized within the framework as an “Optimal Stopping Time” problem and Sequential Decision-Making.
In [1], a finite difference method was proposed for pricing Black–Scholes [2] partial differential Equation (PDE) in American Put Option. The Binomial Option Pricing Model (BOPM) is another method for pricing the American Option proposed by [3]. The key idea is to represent the underlying uncertainty in discrete time using a binomial lattice (Tree), then move backward from to to find the option value. In [4], derived an analytical approximation for the American Option pricing by fitting the empirical function to the results of the American put option prices generated by BOPM.
Following a different approach, if the underlying asset is modeled as the Markov Process, the Bellman principle of Dynamic Programming (DP) [5] can be used to compute the option value. However, this approach becomes impractical when the state space has many dimensions since the DP algorithm requires an exponential memory space in the number of dimensions. This problem is known as the “curse of dimensionality” for DP. To overcome this problem, ref. [6] proposed that the state space domain was partitioned into a finite number of regions, and the continuation value was approximated. Similarly, Least Squares Monte Carlo (LSM) method was proposed to approximate the continuation value with regression on a set of basis functions to develop a low-dimensional approximation to the expected continuation value [7].
Since the optimal exercise time of the American Option is a Sequential Decision-Making problem, with two possible actions at each time step, the Reinforcement Learning (RL) [8] approach has been used to find the fair value of option price by learning the optimal policy.
Reinforcement learning (RL) is a model-free approach for Sequential Decision-Making, where explicit modeling of the environment (quantifying uncertainty) is no longer needed. The initial significant contribution to the field of RL was the advent of Q-Learning by [9]. Further, the integration of the RL algorithm and neural networks led to the development of Deep Reinforcement Learning (DRL), particularly with the introduction of the Deep Q-Network (DQN) [10], enabling RL to be applied to more complex tasks like Atari games using deep convolutional neural networks for approximating Q value function in them. Later developments, such as Proximal Policy Optimization (PPO) by [11] and Soft Actor–Critic (SAC) by [12], further advanced the field of RL by introducing policy gradient methods as a new form of finding an optimal policy in Sequential Decision-Making. More recently, model-based RL like DreamerV2 [13] that learn policy through building the latent space of the world model has improved the sample efficiency of the method. Multi-task Reinforcement Learning [14] has also enhanced the scalability and data efficiency of RL methods.
Least Squares Policy Iteration (LSPI) was proposed to learn the optimal policy for pricing the American Option [15]. They considered both Geometric Brownian Motion (GBM) and a stochastic volatility model as the underlying asset price models and showed good quality of the policy learned by RL. The work of [16] employed a deep learning method to learn the optimal stopping times from Monte Carlo samples. They showed that the resulting stopping policy (neural network policy) can approximate the optimal stopping times. Iteration algorithm for pricing American Options based on Reinforcement Learning was proposed in [17]. At each iteration, the method approximates the expected discounted payoff of stopping times and produces those closer to optimal. A thorough literature review of the modeling and pricing of the American Option (both classical and RL-based methods) can be found in [18]. Reinforcement learning has also been applied to dynamic stock option hedging in markets calibrated with stochastic volatility models [19]. A comprehensive review of the application of Reinforcement Learning for option hedging, as well as recent advances in utilizing Reinforcement Learning for various decision-making processes in finance, can be found in [20,21].
We would like to point out that this research addresses a specific gap in the existing literature on American Options, and while there has been extensive research on American Options, there is a lack of focus on the analysis, emphasis, visualization, and comparison of the “policies (set of exercise decisions)” generated by different algorithms: (1) BOMP: Binomial Option Pricing Model, (2) LSM: Least Squares Monte Carlo, and (3) Reinforcement Learning. Our work is the first to concentrate on the final policy decisions generated by these algorithms, exploring how these policies differ and evolve during the training process (in the RL-based method). To implement this idea and conduct an analysis, we considered three distinct pricing models, encompassing both constant and calibrated stochastic volatility models. The results provide new insights into final policy behaviors under various market conditions (represented by different uncertainty models) and a comparison of their resulting values obtained from these algorithms. These findings offer a better understanding of decision-making processes in the context of American options.
The main contributions of this work are threefold: (a) We explored and implemented three methods for pricing the American Option for both constant and stochastic volatility models of underlying uncertainty. (b) The RL method was studied to understand and shed light on how learning in RL contributes to updating “Decisions” at each batch. (c) We replicated Table 1 from the study by [7] utilizing the Least Squares Monte Carlo Method for pricing American Options, facilitating a comparative analysis with the outcomes derived from the RL approach.
Table 1.
Option price for GBM price models, calculated using BOPM, LSM, and RL methods. Columns indicate spot price (S), volatility (), time to maturity (T).
This paper is organized as follows: Section 2 presents the “Problem Statement”. Section 3 discusses the modeling of the uncertainty of underlying price dynamics. Section 4 presents and briefly reviews the three methods for pricing the American Option. Section 5 presents the work results. This paper concludes with Section 6, “Discussion”, and Section 7, “Conclusions”.
2. Problem Statement
2.1. American Options as Sequential Decision-Making: A Framing Perspective
The optimal exercise time of an American Option can be framed as a Sequential Decision-Making (SDM) problem. SDM is a type of decision problem ubiquitous in many fields, including finance, engineering, and healthcare [22]. SDM involves a sequence of decisions over time that can be represented as a sequence of the following:
After making one decision, additional information is received, which is used to update our uncertainty and support the next decision. We incur costs or receive a contribution (reward) each time we decide. The goal is to make a sequence of decisions (known as policy, guiding a decision maker (DM) on how to act according to information) that maximizes the cumulative reward over time. Reward is a quantity defined by the decision maker to measure what he/she desires. The SDM problem in the context of optimal exercise time for an American Option can be represented in a compact form, as follows:
where
- : is the State variable that captures what we know at time t. In the context of the American Option (AO), is a two-dimensional vector that includes the current time step and the underlying asset’s price , i.e., .
- : is the Decision variable that captures the decision made at time t. The AO problem has two alternatives Exercise the option or Continue holding the option, at each time step before the option is exercised.
- : is the Information varibale that emerges after decision . In an AO context, it is the underlying asset’s price at time .
2.2. Optimal Stopping Time for a Stochastic Process
The optimal exercise time of an AO can be framed as an Optimal Stopping Time concept. Stopping Time is a policy to decide when to stop the stochastic process. In the case of the American Option, the underlying asset’s price follows a stochastic process.
Let be the set of time periods when the option can be exercised. The Optimal Stopping Time is defined as the exercise times at which the expected payoff is maximized upon exercise, as follows:
where in the case of the American Put Option, , where K is the strike price of the option.
The Optimal Stopping Time concept can be intuitively understood as the process of searching through multiple stopping times (i.e., Stopping Policies) and selecting the best one—the stopping policy that maximizes the expected value of a reward function H(·) applied to the stochastic process at the chosen stopping time.
2.3. Stopping Policy Determines the Option Value
The option value (the fair price of an American Option) depends on the stopping time. In other words, the AO’s value is the expected value of discounted payoffs at stopping times given all possible sets of prices over time. The discount factor P formally depends on r, , N and T, as expressed in Equation (2) (). However, since T and N are fixed for each specific option valuation, P is simply is written as ). Let be the discount factor, where r is the risk-free interest rate, defined as
T is expiry time (in year) and N is the number of exercise periods. The value of the American Option is
To illustrate this concept, consider a specific example with an initial asset price of . Suppose we have an American Put option with a strike price of and a maturity time of and . Assume that the underlying asset’s price over time follows the sequence in one possible realization of the stochastic process. If the stopping time is chosen as , the payoff for this specific path is given by . The present value for this specific path (single path), denoted as , is calculated as
Here, represents the value of the option for this particular realization of the asset’s price path, whereas in the general case represents the expected value across all possible paths.
3. Modeling Uncertainty
The value of the American Option is inherently tied to the underlying asset’s price dynamics (X) that is uncertain and needs to be modeled as a stochastic process. In this work, three time series models are considered, as follows:
- Geometric Brownian Motion model (GBM);
- Generalized AutoRegressive Conditional Heteroskedasticity (GARCH);
- Exponential Generalized Autoregressive Conditional Heteroskedasticity (EGARCH).
GBM is chosen as a benchmark due to its simplicity and widespread use in price modeling. GARCH and EGARCH price models are a richer model for underlying asset price dynamics and a more complex model for solving AO valuation problems, respectively.
3.1. Geometric Brownian Motion (GBM) Price Model
An Itô process is a type of stochastic process that is used to model the evolution of variables (here asset price) over time with both deterministic and random components. The deterministic part (known as the Drift Term) represents the average rate of change, while the random component representing fluctuation is known as the Diffusion Term. Consider a stochastic process x described in the form of the following Itô process:
In this equation, represents the drift term, represents the diffusion term, and captures the random shocks to the system. GBM is a special case of the Itô process, with being a constant and being a constant [23]. In GBM, the logarithm of the process, denoted as , follows a normal distribution with mean and variance given by the following:
This equation states that the logarithmic value of the price at the time step (t + 1) is normally distributed, with the mean adjusted for the drift () and volatility (), and variance proportional to the time increment .
3.2. GARCH Price Model
The Generalized Autoregressive Conditional Heteroskedasticity (GARCH) model is widely used to model time-varying volatility in econometrics [24]. The process is defined as
where is the asset return , is a term that captures the mean of the return, and is drawn from a normal distribution with zero mean and variance of , i.e., . The conditional variance of the return is modeled as
A GARCH(1,1) model has and and is thus defined as
For using the GARCH(1,1) model, the four parameters , , , and need to be estimated. In this work, the Maximum Likelihood (ML) method is employed to estimate these parameters based on historical data. The process of how to use ML for parameter estimation in this work has been discussed in Appendix A.
3.3. EGARCH Price Model
The GARCH price model has limitations in capturing asymmetry in volatility. To overcome this shortcoming, the EGARCH model is introduced. In the EGARCH model, the asset return has same equation as Equation (6), though the conditional variance is modeled as
Rather than working with the complete specification, a simpler version, an EGARCH(1,1,1) with a constant mean is defined as
For using the EGARCH(1,1,1) model, five parameters , , , , and need to be estimated. Similar to the previous subsection, the Maximum Likelihood (ML) method is employed to estimate these parameters based on historical data.
4. Optimal Exercise Time of American Option
In this section, three methods for valuing an American Option are examined:
- BOPM: Binomial Options Pricing Model [3];
- LSM: Least Squares Monte Carlo [7];
- RL: Reinforcement Learning (Least Square Policy Iteration (LSPI) with Experience-Replay) [25].
A brief overview of each method is provided below.
4.1. Binomial Options Pricing Model (BOPM)
The original BOPM [3] was developed to price options on underlying whose price evolves according to a lognormal stochastic process. The BOPM assumes the underlying price follows a GBM (Equation (5)). The BOPM is a discrete-time, finite-horizon, finite-state approximation of the continuous process. The BOPM is tree based, where a “recombining tree” is constructed, meaning we have price outcomes after i time steps. For each state (price) random the resulting (two nodes) from are (upper node) and (lower node). In BOPM, u is the “up” factor (by which the asset price increases and ), and q is the probability of an “up” move. The graphical structure of the Binomial Option Pricing Model (BOPM) is illustrated in Figure 1. For a detailed explanation of the calibration of q and u in BOPM, we refer to Appendix B and [23].
Figure 1.
Illustration of a binomial tree of price states for time steps in BOPM.
After defining the parameters u and q, the procedure for pricing an American Option using BOPM is as follows. The process starts at the terminal time and moves backward in time to . At each time step i, two values are calculated for each corresponding price state: the “continuation value” and “immediate exercise value”. The maximum among those two values determines the option value at time step i and price state j. This process is repeated until the time is reached. The value calculated at is the value of the option at the current time and underlying price.
4.2. Least Square Monte Carlo
The widely used approach in the finance industry to price the American Option is the Least Square Monte Carlo (LSM) method (also known as the Longstaff–Schwartz pricing algorithm [7]). The LSM method for pricing an American Option is based on the following features:
- Monte Carlo Simulation (MCS). LSM uses MCS, based on a stochastic process model, to generate a sufficient number (more than 10,000) of price paths from the current time to the terminal time (expiry date). The value of exercising an option at each time step of each price path is calculated.
- Least Square Regression. The continuation value—the conditional expected value of continuing holding the option—given a possible underlying asset price at a time step is approximated using a least squares regression function for in-the-money states. Various regression functions can be used. The simplest one is a linear function. With this specification, a functional form of conditional expected value can be represented withIn Equation (15), represents a sample path, the coefficients to be estimated through the least square regression, and is the basis function. In their work, [7] explored various types of basis functions, including Hermite, Legendre, Chebyshev, Gegenbauer, and Jacobi polynomials. The numerical results demonstrate that the Least Squares Monte Carlo (LSM) algorithm exhibits robustness to the selection of basis functions. Among the possible choices, a set of Laguerre polynomials was used in their paper for numerical convenience [7]. In Equation (15), has been represented as a linear combination of basis functions, where coefficients are constant.
- Option Valuation. After finding the coefficients of the regression function at each time step, a new set of paths for the underlying asset’s price, ‘Test Paths’, is generated. The regression function for the continuation value at each time step is then applied to these test paths to identify the paths where ‘early exercise’ is optimal. The value of the option is determined by computing the average of the discounted payoffs derived from these test paths.
4.3. Reinforcement Learning
As AO valuation can be framed as a Sequential Decision-Making problem, it can be solved using a Reinforcement Learning (RL) framework. Li et al. [15] have shown that Least Square Policy Iteration (LSPI) can be a promising Reinforcement Learning method for pricing the American Option.
The AO elements in an RL framework are as follows:
- State (s): [Time, Underlying Asset Price];
- Action (a): Exercise the option or Continue holding the option;
- Reward (r): For continuation 0, except upon Exercise (when the Reward is equal to the immediate exercise value);
- State transitions: Based on stochastic process model for the underlying asset price.
Since the immediate exercise value is known in any state, we only need to create a linear function approximation for the continuation value (action = continue). The function approximation is linear with respect to weights, although the feature functions themselves are non-linear. We can write the Q-value function as
where is the weight vector, is the feature function vector, is the exercise value (immediate payoff). The state s is a tuple of underlying asset price x and time t, . The feature functions are a set of Weighted Laguerre Polynomials, similar to the work of [7]. However, these feature functions are applied both to underlying price and time. In this work, the following feature functions are used:
where and K is the option’s strike price. The same feature functions are applied to time, with in Equation (17) replaced by , where T is the expiry date of the option (usually in years). In total, there are 12 feature functions, and is a vector of 12 elements.
In LSPI [25], the optimal weight vector is determined by minimizing the following objective function:
where is an immediate reward, is the discount factor, is the subsequent (next) state (), and is the target decision policy (i.e., the policy that RL is trying to learn). The term is the Q-value at time step i for the state-action pair . The term is known as the target value.
Taking the semi-gradient of the objective function (Equation (18)) gives
The term is a policy operator, where we need to apply greedy policy. In the case of AO, it is the max operator between the continuation and exercise values. This max operator can also be found in the classical Bellman optimality equation [5]. The term can have two cases, as follows:
- Case 1 (): If is a non-terminal state (not at the terminal time step N in the American Option) and the continuation value at state is higher than the stopping value (i.e., and ), substitute with .
- Case 2 (): If is a terminal state or (i.e., , substitute with .
Using indicator function I for cases C1 and C2, the semi-gradient equation (Equation (19)) can be rewritten as
where for and otherwise, and the same applies to . Factoring out gives
This can be written in
with
The term atomic experience is each data point in the training data of RL, which consists of a stream of . The matrix and (m is the size of feature function vector, here ) matrix is accumulated at each atomic experience
Thus, is solved as and updates the Q-value function approximation [23]. The update of the Q-value function reflects the update in the “decision policy” during the training phase of the Reinforcement Learning method. Unlike classical RL methods [8], which update weights after each individual experience , a batch (set of experiences) training approach is implemented in this work.
5. Results
5.1. Pricing Methods and Experimental Setup
In this section, the results of applying the three pricing methods—BOPM, LSM, and RL—to the following cases are presented as follows:
- The underlying price is modeled using GBM (Section 5.1).
- The underlying price is modeled using GARCH (Section 5.2).
- The underlying price is modeled using EGARCH (Section 5.3).
The parameters for each pricing model, including their calibration (for GARCH and EGARCH models), are provided within each respective section.
It is worth mentioning that the Binomial Option Pricing Model (BOPM) is appropriate for the Geometric Brownian Motion (GBM) price model because GBM assumes constant volatility and a log-normal price distribution, fitting well with BOPM’s discrete time steps and constant volatility assumptions. However, stochastic volatility models like GARCH and EGARCH feature time-varying volatility, which complicates the BOPM framework to be applicable. Therefore, BOPM is not suitable for these models, and only the RL and LSM methods have been applied to GARCH and EGARCH price models.
5.2. Option Price for GBM Price Model
The use of the GBM price model facilitates comparability with previous studies [7] in the field. In addition, the exact solution of the AO valuation with the GBM price model can be solved using BOPM as the number of steps in the binomial tree approaches infinity.
5.2.1. Comparison of AO Valuation by BOPM, LSM, and RL
We first replicate Table 1 in Longstaff and Schwartz’s work [7] for BOPM and LSM. Then, RL is used to solve the same cases. In all cases (rows of Table 1), the option’s strike price is , the number of available opportunities to exercise is (i.e., T = 1, ) and the risk-free discount rate is set to .
In BOPM, each node branches out into two distinct paths: one upward movement and the other a downward movement. Furthermore, , denotes the total number of time steps in the model. In Table 1, the BOPM results are the exact solutions as a benchmark for comparing the results of the other two methods. The comparison shows that LSM and RL yield consistent results to BOPM, often matching to at least two decimal places.
In this study, the number of paths for training RL is 5000. In our experiments, we increased the number of paths to over 10,000 and up to 50,000; however, these modifications did not yield significant improvements in the results. Due to the stochastic nature of GBM during RL training, the RL results in Table 1 are the mean option prices of ten repetitions of the valuation process (for each case, the option is priced ten times, and the results are averaged). For LSM, the number of paths is 100,000, similar to Longstaff and Schwartz’s work [7].
5.2.2. Decision Boundary
In Table 1, the fair price of each option has been computed using the three methods. These three methods lead to slightly different values because their decision boundaries (the decisions to continue and exercise) are slightly different.
To delve into and understand this difference more, Figure 2 illustrates the decision boundary for the GBM price model with , , and . At any point in time (during the exercise window), if the price of the underlying asset is lower than the decision boundary curve, the option should be exercised. In Figure 2, the x-axis has been normalized (Time = t/Expiration Time), and the y-axis is the price of the underlying asset. The “locally jagged” nature of the decision boundary in the BOMP method is noteworthy. This boundary results from the discrete, stepwise structure of underlying asset price within the binomial lattice. The “diamond-like” local structure of these prices forms clusters, leading to jumps along the decision boundary curve, a phenomenon similarly discussed in [23].
Figure 2.
Plot of decision boundaries for BOPM, LSM, and RL. The underlying asset is modeled using GBM with parameters (S0 = 36, strike price = 40, volatility = 0.2).
5.2.3. Decision Frequency
In this section, we look into the frequency of exercise time for each method. The “frequency of exercise time” in this context is the distribution of exercise time (time of stopping) for 10,000 paths (generated from the GBM model), based on a method (LSM or RL).
Figure 3 shows the frequency of exercise time for the case where spot price , strike price , and volatility are used in GBM. For example, at time step 5, the number of paths, for which the option is exercised, is around 500 for both LSM and RL. That is, for around 500 out of 10,000 paths (5%), LSM and RL suggest exercising the option at step 5.
Figure 3.
Frequency of exercise time for LSM and RL. The underlying price model is GBM with , , and .
Note that in Figure 3, the bars after time step 50 represent the paths where the option is not exercised at the expiry time step or before (i.e., the option is never exercised). Both LSM and RL suggest not exercising the option at the expiry time step or before for around 25% of paths, leading to zero payoff for those paths.
LSM and RL lead to similar frequencies of exercise time in general, implying that RL can solve for a decision policy similar to LSM does.
We consider another case with larger uncertainty in the underlying asset prices over time, by using spot price , strike price , and volatility in GBM. Figure 4 shows the resulting frequency of exercise time for each method. Compared to the case with volatility (Figure 3), the frequencies of exercise time (Figure 4) are different.
Figure 4.
Frequencies of exercise time for LSM and RL. , , and in GBM.
For the case with (Figure 4), RL leads to a different frequency of exercise time compared to LSM; especially for early time steps, RL does not suggest exercising the option before time step 24 for any path, whilst LSM suggests exercising the option before time step 24 for some paths. This behavior contrasts with Figure 3, where the RL and LSM methods exhibited similar frequencies of early exercises across all time steps.
5.2.4. How Well Does RL Decision Policy Improve Throughout Training?
Figure 5 shows the change in the frequency of exercise time for RL, indicating the evolution of RL decision policy throughout the training. At the beginning of the training, the weights are equal. As the training progresses (more batches), the weights converge to optimal values, and the decision policy also converges to the optimal one. The training is conducted with 5000 paths in five batches. As shown in Figure 5, the frequency of exercise time has almost no change after “Batch 2”, which indicates the decision policy converges to the optimal one. As explained in greater detail in the next subsection, the Option Value was computed after each batch and compared with the results from the BOMP method. If the final policy from the RL training resembles this value, it is considered an optimal policy.
Figure 5.
Change in frequency of exercise time throughout RL training. , , and in GBM.
5.2.5. How Option Values Change during Training?
Figure 6 shows the change in option value with each batch during the RL training process. In this example, the initial weights were set to . Subsequently, new weights are found in each batch. The new weights are then applied to a set of test paths to find the resulting option value option.
Figure 6.
Change in Option Value following each batch during Reinforcement Learning training.
5.3. Option Price for GARCH Price Model
This section discusses the outcomes of implementing the two option valuation methods when the underlying price is modeled using the Generalized Autoregressive Conditional Heteroskedasticity (GARCH) model. The GARCH price model was selected to test the robustness of the RL method because GARCH is more complex and volatile than GBM. We calibrate the GARCH to historical Brent Crude Oil price data. In this section, Brent Crude Oil price acts as the Underlying Asset for the American Option. Subsequently, this calibrated GARCH model is utilized to simulate underlying asset price paths.
Note that BOMP is not applicable to GARCH, so only LSM and RL are implemented for the GARCH case. In addition, in this example, the strike price is .
5.3.1. Hitorical Brent Crude Oil Price Data and GARCH Calibration
The historical Brent Crude Oil price data from 1987 to 2024, as shown in Figure 7, are used to calibrate the GARCH price model. Figure 8 contrasts the realized volatilities of the Brent Crude Oil price data and the forecasted volatilities of the calibrated GARCH model.
Figure 7.
Historical Brent Crude Oil price data from 1987 to 2024.
Figure 8.
Contrast of the realized volatilities of the Brent Crude Oil price data and the forecasted volatilities of the calibrated GARCH model.
5.3.2. Decision Boundary
The decision boundary curves solved using LSM and RL for the case with the calibrated GARCH price model are plotted in Figure 9. The decision boundary curves for the LSM and RL methods follow a similar trend, where the boundary curves at the expiry time approach the strike price ().
Figure 9.
Decision boundaries solved using LSM and RL for the case with calibrated GARCH price model. At any given time (x-axis), if the underlying asset price falls below the boundary, the option is to be exercised.
5.3.3. Decision Frequency
The resulting LSM and RL decision policies, as illustrated in Figure 9, are applied to 20,000 price paths sampled from the calibrated GARCH price model with the spot price . The frequency of exercise time for each method is shown in Figure 10. Again, RL leads to a similar frequency of exercise time as LSM does. For approximately 7000 out of the 20,000 paths, LSM and RL suggest not exercising the option within the expiry time; in other words, the option is suggested to be exercised before/at the expiry time for around 14,000 paths.
Figure 10.
Frequencies of exercise time for LSM and RL with the calibrated GARCH model.
5.3.4. How Well Does RL Decision Policy Improve Throughout Training?
Figure 11 shows the development of the frequency of exercise time, implying the development of the RL decision policy over the course of training. Initially (batch 0), the RL decision policy predominantly recommends exercising the option during the middle stages of the timeline (time steps between 15 and 30). However, with more batches of training, adjustments to the RL decision policy result in progressively fewer middle-stage exercises and more early-stage exercises. By the end of the fourth batch (Number 3 on the right side of Figure 11), the frequency of exercise time stabilizes, which indicates the RL decision policy converges.
Figure 11.
Evolution of the frequency of exercise time throughout RL training batches.
5.3.5. Final Option Values
The two methods, RL and LSM, were used to find the decision boundary (policy) shown in Figure 9. The resulting decision boundaries from both methods were then tested on 10,000 paths to estimate the option value. Each method was repeated ten times to account for the stochastic nature of underlying price uncertainty (modeled by GARCH(1,1)). The average option value (, ) was USD 10.75 for the RL method and USD 10.72 for the LSM method.
5.4. Option Price for EGARCH Price Model
This section discusses the implementation of RL and LSM methods for pricing AO, where the underlying price is modeled using the Exponential Generalized Autoregressive Conditional Heteroskedasticity (EGARCH) model. The EGARCH price model was selected since it can handle situations where negative returns increase future volatility more than positive returns. The GARCH model does not easily capture this feature. It makes EGARCH better suited for modeling real-world financial data where such asymmetry is common. We calibrate the EGARCH to historical Brent Crude Oil price data, the same as in the previous section. Subsequently, this calibrated EGARCH model is utilized to simulate underlying asset price paths.
5.4.1. EGARCH Price Model Calibration
Similar to the GARCH method, the Maximum Likelihood Estimation (MLE) method was used to estimate the parameters of the EGARCH model by calibrating it to oil price data (Figure 7). The estimated parameters of the EGARCH model are presented in Table 2. Figure 12 contrasts the realized volatilities of Brent Crude Oil price data with the forecasted volatilities from the calibrated EGARCH model. In Figure 12, the EGARCH(1,1,1) price model, due to its short memory (1,1,1), does not perfectly reconstruct the data. However, it is important to clarify that the primary focus of this paper is not on forecasting or developing models that perfectly fit the data. Rather, our work is centered on “Sequential Decision-Making” in American Options, where these models are used to simulate future paths as inputs to different pricing methods. Thus, even though the EGARCH(1,1,1) model may not reconstruct the data with high accuracy, it still serves as a consistent input for all methods under comparison.
Table 2.
Parameter Values for the EGARCH model.
Figure 12.
Contrast of the realized volatilities of the Brent Crude Oil price data and the forecasted volatilities of the calibrated EGARCH model.
5.4.2. Decision Boundary
The decision boundary curves solved using LSM and RL for the case with the calibrated EGARCH price model are plotted in Figure 13. A notable difference between these boundaries is that the RL method results in “more” early exercises before half the expiry date. In contrast, the LSM method has a lower value boundary (), resulting in fewer paths below this value and, consequently, “fewer” early exercises.
Figure 13.
Decision boundaries solved using LSM and RL for the case with calibrated EGARCH price model.
5.4.3. Decision Frequency
The resulting decision policies using LSM and RL, illustrated in Figure 13, were applied to 10,000 price paths sampled from the calibrated EGARCH price model with a spot price of and strike price . The frequency of exercise times for each method is shown in Figure 14. In both methods, approximately 35% of the paths are not exercised early. However, the RL method exhibits more early exercises before time step 25, as indicated by its decision boundary (discussed in the previous section).
Figure 14.
Frequencies of exercise time for LSM and RL with the calibrated EGARCH model.
5.4.4. Final Option Values
The two methods, RL and LSM, were used to find the decision boundary (policy) shown in Figure 13. The decision boundaries obtained from both methods were then tested on 10,000 paths to estimate the option value. Each method was repeated ten times to account for the stochastic nature of underlying price uncertainty, modeled by EGARCH(1,1,1). The average option value, with , strike price = USD 80 was USD 15.01 for the RL method and USD 14.91 for the LSM method.
6. Discussion
The classical LSM method is based on a predefined probabilistic model of an underlying asset price. The LSM samples future paths of the underlying asset price from the predefined probabilistic model, estimates conditional expected values using regression, and derives the option value through backward induction.
Conversely, the Reinforcement Learning (RL) approach is model free, eliminating the need for a predefined probabilistic price model to value an option. Instead, RL can utilize historical underlying asset price data to train for the optimal decision policy, exemplifying a data-driven approach to pricing American Options. Our study demonstrates that the RL method is effective for valuing American Options and yields results comparable to those obtained using the BOPM and LSM methods.
A limitation of the RL method needs to be mentioned. The stability of the training process can be an issue in finding the optimal policy. Careful attention is required when designing the RL algorithm with the appropriate number of batches, iterations, and learning rates to avoid overfitting and underfitting.
Future work could explore the potential of different feature functions (in this study only Laguerre was considered) in improving policy and valuation of the American Option. Additionally, investigating other types of underlying price models, such as jump models or two-factor models, could provide further insight.
7. Conclusions
Identifying the optimal exercise time of an American Option is a Sequential Decision-Making problem. At each time step, the option holder must decide whether to exercise the option based on the immediate reward or wait for potential future rewards, considering downstream uncertainties and decisions.
Our analysis shows that decision boundaries derived from the Least Squares Method Monte Carlo (LSM) method tend to be non-smooth. Notably, for the EGARCH(1,1,1) price model, the decision boundary changes abruptly at certain time steps when the number of sampled prices in the training set is around 100,000 paths. This occurs because LSM performs regression at each time step, whereas RL embeds time (t) directly within the Q-value function. A further conclusion is that, in the EGARCH (1,1,1) model, the LSM and RL-based methods result in two different exercise policies. However, as demonstrated in Section 5.4.4, applying these different final policies to test paths yields the same option valuation. Despite the different exercise policies, the final option values are identical. This contrasts with the cases of the GBM and GARCH (1,1) models, where both the final policies and the option values are similar.
Finally, the Reinforcement Learning (RL) method learns policies through an iterative process. During this learning process, an iteration step where policy “change” stops is observed, indicating convergence. In this work, by employing the experience-replay method in RL, convergence is achieved within a few batches. The key contributions of this work can be summarized as follows: (a) The implementation, illustration, comparison, and discussion of the three methods—BOPM, LSM, and RL—for valuing American Options in cases with constant and stochastic volatility models—GBM and GARCH—of underlying asset price uncertainty. (b) The study and evaluation of an RL implementation for Sequential Decision-Making sheds light on how learning in RL contributes to improving “decisions”. (c) The authors of [7] presented twenty different GBM-based models and used the LSM technique to determine option values. In this work, we apply the RL method to the same models, as shown in Table 1, and compare it with the LSM approach.
Author Contributions
Conceptualization, P.K., R.B.B. and A.H.; methodology, P.K. and R.B.B.; software, P.K.; validation, P.K.; formal analysis, P.K.; investigation, P.K. and R.B.B.; resources, P.K.; data curation, P.K.; writing—original draft preparation, P.K.; writing—review and editing, R.B.B. and A.H.; visualization, P.K.; supervision, R.B.B.; project administration, R.B.B.; funding acquisition, R.B.B. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
Our data and code are publicly available from 4 September 2024 in https://github.com/Peymankor/rlvother_for_AMoption.
Conflicts of Interest
The submitted work was conducted independently of and outside Aojie Hong’s work in Equinor. Aojie Hong contributed to this paper as a private person without representing any organization. The authors declare no conflicts of interest.
Abbreviations
The following abbreviations are used in this manuscript:
| BOPM | Binomial Option Pricing Model |
| DP | Dynamic Programming |
| LSM | Least Squares Monte Carlo |
| GARCH | Generalized Autoregressive Conditional Heteroskedasticity |
| EGARCH | Exponential Generalized Autoregressive Conditional Heteroskedasticity |
| GBM | Geometric Brownian Motion |
| RL | Reinforcement Learning |
| LSPI | Least Square Policy Iteration |
| AO | American Option |
Appendix A
Maximum Likelihood
Given that the values of are assumed to be conditionally i.i.d (independently and identically distributed), Maximum Likelihood (ML) is a natural choice to estimate the unknown parameters, . The likelihood function for parameter r is defined as
The log-likelihood for T independent variables can be written as
Note that the log-likelihood function is a function of the unknown parameters where they are embeddded into the conditional variance , Equation (8).
A simple Nelder–Mead method was used to maximize the log-likelihood function to estimate the parameters of the GARCH(1,1) model. Having found the optimal parameter values , we can simulate the GARCH(1,1) model to generate the new paths of the underlying asset price. A schematical workflow of this process is shown in Figure A1.
Figure A1.
Workflow for generating GARCH(1,1) simulation paths.
Appendix B
The important point is that u remains a constant across time steps (i) and states (j). Let q be the probability of the “up move” so that is the probability of the “down move”. One needs to calibrate q and u so that the probability distribution of log price-ratios after n time steps serve as a good approximation of the probability distribution of the stochastic process.
We can calibrate q and u using the two steps process, as follows:
Step 1
Step 2
References
- Brennan, M.J.; Schwartz, E.S. The Valuation of American Put Options. J. Financ. 1977, 32, 449–462. [Google Scholar] [CrossRef]
- Black, F.; Scholes, M. The valuation of option contracts and a test of market efficiency. J. Financ. 1972, 27, 399–417. [Google Scholar] [CrossRef]
- Cox, J.C.; Ross, S.A.; Rubinstein, M. Option pricing: A simplified approach. J. Financ. Econ. 1979, 7, 229–263. [Google Scholar] [CrossRef]
- Johnson, H.E. An analytic approximation for the American put price. J. Financ. Quant. Anal. 1983, 18, 141–148. [Google Scholar] [CrossRef]
- Bellman, R. A Markovian decision process. J. Math. Mech. 1957, 679–684. [Google Scholar] [CrossRef]
- Barraquand, J.; Martineau, D. Numerical valuation of high dimensional multivariate American securities. J. Financ. Quant. Anal. 1995, 30, 383–405. [Google Scholar] [CrossRef]
- Longstaff, F.A.; Schwartz, E.S. Valuing American options by simulation: A simple least-squares approach. Rev. Financ. Stud. 2001, 14, 113–147. [Google Scholar] [CrossRef]
- Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction, 2nd ed.; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
- Watkins, C.J.C.H.; Dayan, P. Q-Learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
- Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-Level Control through Deep Reinforcement Learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
- Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
- Haarnoja, T.; Zhou, A.; Abbeel, P.; Levine, S. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proceedings of the International Conference on Machine Learning; PMLR: Cambridge MA, USA, 2018; pp. 1861–1870. [Google Scholar]
- Hafner, D.; Lillicrap, T.; Norouzi, M.; Ba, J. Mastering atari with discrete world models. arXiv 2020, arXiv:2010.02193. [Google Scholar]
- Vithayathil Varghese, N.; Mahmoud, Q.H. A Survey of Multi-Task Deep Reinforcement Learning. Electronics 2020, 9, 1363. [Google Scholar] [CrossRef]
- Li, Y.; Szepesvari, C.; Schuurmans, D. Learning exercise policies for American options. In Proceedings of the 12th International Conference on Artificial Intelligence and Statistics (AISTATS) 2009, Clearwater Beach, FL, USA, 16–18 April 2009; Volume 5. [Google Scholar]
- Becker, S.; Cheridito, P.; Jentzen, A. Deep optimal stopping. J. Mach. Learn. Res. 2018, 20, 1–25. [Google Scholar]
- Li, N. An Iteration Algorithm for American Options Pricing Based on Reinforcement Learning. Symmetry 2022, 14, 1324. [Google Scholar] [CrossRef]
- Bloch, D.A. American Options: Models and Algorithms. SSRN Electron. J. 2023. [Google Scholar] [CrossRef]
- Pickard, R.; Wredenhagen, F.; Lawryshyn, Y. Optimizing Deep Reinforcement Learning for American Put Option Hedging. arXiv 2024, arXiv:2405.08602. [Google Scholar]
- Pickard, R.; Lawryshyn, Y. Deep Reinforcement Learning for Dynamic Stock Option Hedging: A Review. Mathematics 2023, 11, 4943. [Google Scholar] [CrossRef]
- Hambly, B.; Xu, R.; Yang, H. Recent Advances in Reinforcement Learning in Finance. arXiv 2023, arXiv:2112.04553. [Google Scholar] [CrossRef]
- Bratvold, R.; Begg, S. Making Good Decisions; Society of Petroleum Engineers: Richardson, TX, USA, 2010. [Google Scholar]
- Rao, A.; Jelvis, T. Foundations of Reinforcement Learning with Applications in Finance; Chapman and Hall/CRC: Boca Raton, FL, USA, 2022. [Google Scholar]
- Sheppard, K. Financial Econometrics Notes; University of Oxford: Oxford, UK, 2010; pp. 333–426. [Google Scholar]
- Lagoudakis, M.G.; Parr, R. Least-squares policy iteration. J. Mach. Learn. Res. 2003, 4, 1107–1149. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).