1. Introduction
In the domain of Deep Reinforcement Learning (RL), the concept known as Experience Replay (ER) has long since developed to become a well-known standard for many algorithms [
1,
2,
3]. Initially designed as an extension for Q- and AHC-Learning [
4], it has become an integral part of the Deep Q-Network (DQN) family. Here, it is mandatory to overcome instabilities in the learning phase [
5]. Another positive effect that comes along with the use of ERs is an increased sample efficiency, which is achieved by reusing remembered transitions several times.
Being a key component for many Deep RL algorithms makes the concept of ER attractive for improvements and extensions. Most of them store real, actually experienced, transitions. Mnih et al. [
2], for example, used the basic ER version to assist their DQN, and Schaul et al. [
1] extended it to a version called Prioritized Experience Replay, which, instead of uniformly drawing, prefers experiences that promise greater learning successes. However, there are other approaches as well, and these extensions focus on the usage and creation of experiences that are synthetic in some way. An example of this is the so-called Hindsight Experience Replay [
3] that saves trajectories of states and actions together with a corresponding goal. By replacing the goal with the last encountered state, a synthetic trajectory is created and saved together with the real one. The authors could show that this method promises great success in multi-objective problem spaces.
The contribution of our work is an ER-extension that uses synthetic experiences. A (not complete) list of (Deep) RL algorithms that use an ER is the following: DQN, DDPG or classic Q-Learning [
6]. Our ER version is targeted to improve the performance of these algorithms in nondeterministic and discrete environments. To achieve this, we consider all the stored real state-transitions as the gathered knowledge of the underlying problem. Utilizing this knowledge, we are able to create synthetic experiences that contain an average reward of all related surroundings. This approach makes it possible to increase the sample efficiency even more because experiences are now also used for the generation of new and possibly better synthetic experiences. We use observed follow-up states to complete our so-called
interpolated experiences and can support our learner in the exploration phase this way.
The evaluation is performed on the
FrozenLake environment from the OpenAI Gym [
7]. Offering a discrete state space in the form of a grid world, in combination with a non-deterministic state-transition-function, makes it a good choice to evaluate our algorithm on. To increase scientific relevance and validity, we evaluate three different state encodings and corresponding various deep network architectures.
The investigated problem is discrete and non-deterministic, and the averaging is a rather simple method as well, but the intention is to gain the first insights in this highly interesting field. We can reveal promising potential utilizing this very simple technique, and this work serves as a basis to build up further research on.
The present work is an extended version of [
8]. In addition to the original publication, we used actual deep neural networks instead of a linear regression. We also investigated different state encodings, making the problem more difficult and interesting and also increasing the scientific validity. We tied up on the results of this paper and were able to define new questions to investigate. The key idea remains the same but is extended with deeper evaluation.
The paper is structured as follows: We start with a brief introduction of the ER and Deep Q-Learning in
Section 2 and proceed with relevant related work in
Section 3. In
Section 4, we introduce our algorithm alongside with a problem description and the Interpolation Component that was used as an underlying architecture. The evaluation and corresponding discussion, as well as interpretation of the results, are presented in
Section 5. The article closes with a conclusion and gives an outlook on future work in
Section 6.
2. Background
In this section, we start with introducing the idea of the Experience Replay and continue with the presentation of Deep Reinforcement Learning basics as well as an explanation of why the former concept is mandatory here.
2.1. Experience Replay
The ER is a biologically inspired mechanism [
4,
9,
10,
11] to store experiences and reuse them for training later on.
An experience is defined as: , where denotes the start state, the performed action, the corresponding received reward and the follow-up state. To perform Experience Replay, at each time step t, the agent stores its recent experience in a data set . In an non-episodic/infinite environment (and also in an episodic one after enough time has gone by), we would run into the problem of limited storage. To counteract this issue, the vanilla ER is realized via a FiFo buffer, and old experiences are thrown away after reaching the maximum length.
This procedure is repeated over many episodes, where the end of an episode is defined by a terminal state. The stored transitions can then be utilized for training either online or in a specific training phase. It is very easy to implement ER in its basic form, and the cost of using it is mainly determined by the storage space needed.
2.2. Deep Q-Learning
The DQN algorithm is the combination of the classic Q-Learning [
12,
13] with neural networks and was introduced in [
2,
14]. The authors showed that their algorithm is able to play Atari 2600 games on a professional human level utilizing the same architecture, algorithm and hyperparameters for every single game. As DQN is a derivative of classical Q-Learning, it approximates the optimal action-value function:
However, DQN employs a neural network instead of a table to parameterize the Q-function. Equation (
1) displays the maximum sum of rewards
discounted by
at each time-step
t, which is achievable by a behaviour policy
, after making an observation
s and taking an action
a. DQN performs a Q-Learning update at every time step that uses the temporal-difference error defined as follows:
Tsitsiklis et al. [
5] showed that a nonlinear function approximator used in combination with temporal-difference learning, such as Q-Learning, can lead to unstable learning or even divergence of the Q-Function.
As a neural network is a nonlinear function approximator, several problems arise:
the correlations present in the sequence of observations;
the fact that small updates to Q may significantly change the policy and, therefore, impact the data distribution; and
the correlations between the action-values and the target values
present in the td-error shown in Equation (
2).
The last point is crucial because an update to Q will change the values of both the action-values and the target values. This change could lead to oscillations or even divergence of the policy. To counteract these issues, two concrete actions have been proposed:
The use of an ER solves, as stated above, the two first points. Training is performed each step on minibatches of experiences , which are drawn uniformly at random from the ER.
To remove the correlations between the action-values and the target values, a second neural network is introduced that is basically a copy of the network used to predict the action-values. The
target-network is either frozen for a certain interval
C before it is updated again or “soft” updated by slowly tracking the learned networks weights utilizing a factor
. This network is responsible for the computation of the target action-values. [
2,
15]
We use the target network performing “soft” updates as presented above and extend the classic ER with a component to create synthetic experiences.
3. Related Work
The classical ER, introduced in
Section 2.1, has been improved in many further publications. One prominent improvement is the so called Prioritized Experience Replay [
1], which replaces the uniform sampling with a weighted sampling in favour of experience samples that might influence the learning process most. This modification of the distribution in the replay induces bias, and to account for this, importance-sampling has to be used. The authors show that a prioritized sampling leads to great success. This extension of the ER also changes the default distribution but uses real transitions and therefore has a different focus.
The authors of [
16,
17] investigated the composition of experience samples in the ER. They discovered that for some tasks, transitions made in an early phase when exploration is high are important to prevent overfitting. Therefore, they split the ER into two parts: one with samples from the beginning and one with actual samples. They also show that the composition of the data in an ER is vital for the stability of the learning process, and at all times, diverse samples should be included. Following these results, we try to achieve a broad distribution over the state space utilizing synthetic experiences (for most of our configurations).
Jiang et al. [
18] investigated ERs combined with model-based RL and implemented a tree structure to represent a model of the environment. In their research, they learned a model of the problem and invented a tree structure to represent it. Using this model, they could simulate virtual experiences that they used in the planning phase to support learning. To increase sample efficiency, experience samples are stored in an ER. This approach has some similarities to the interpolation-based approach presented in this work but addresses other aspects, such as learning a model of the problem first.
Gu et al. [
19] presented an interpolation of on-policy and off-policy model-free Deep Reinforcement Learning techniques. In this publication, an approach of interpolation between on- and off-policy gradient mixes likelihood ratio gradient with Q-Learning, which provides unbiased but high-variance gradient estimations. This approach does not use an ER and therefore differs from our work.
This work draws on the methods proposed in [
20,
21,
22,
23,
24]. The authors used interpolation in combination with an XCS classifier System to speed up learning in single-step problems by using previous experiences as sampling points for interpolation. Our approach focuses on a DQN as a learning component and, more importantly, multi-step problems and therefore differs from this work. Anyhow, we adopted the so-called Interpolation Component that is introduced in more detail in
Section 4.3.
4. Interpolated Experience Replay
In this Section, we present the FrozenLake problem and introduce our algorithm to solve it. We also introduce the Interpolation Component that serves as architectural concept.
4.1. Problem Description
“FrozenLake” is one example of a non-deterministic world in which an action
realised in a state
may not consistently lead to the same follow-up state
. FrozenLake is basically a grid world consisting of an initial state
I, a final state
G and frozen, as well as unfrozen tiles. The unfrozen tiles equal holes
H in the lake and if the agent falls into one of such, he has to start from the initial state again. If the agent reaches
G, he receives a reward of 1. The set of possible actions
A consists of the four cardinal directions
. Executing a concrete action (e.g.,
N) only results with a probability of
in the corresponding field, but it is also possible that the agent instead performs one of the orthogonal actions (in our example:
W or
E) with the same probability of
for each case. This behaviour makes the environment non-deterministic. Because there is a discrete number of states the agent can reach, we can denote the problem as discrete as well. The environment used for evaluation is the “FrozenLake8x8-v0” environment from OpenAI Gym [
7], as depicted in
Figure 1.
In addition to the described version, we changed the reward function to return a reward of in the case of falling into a hole and 5 for reaching the goal. As the first adaption is crucial for our approach, the second change helps the learner to solve the environment. Both changes intensify the received rewards and therefore the experienced transitions. Assigning a negative reward to the end of an episode (hole) makes it possible to calculate an average reward containing an additional value (see below). By testing different final rewards (goal), we could observe that the agent performed best with a reward of 5.
The decision to focus on the presented environment was taken because: (1) it is a relatively well-known problem in the RL community (OpenAI Gym); (2) our presented approach is designed for discrete and non-deterministic environments. To add more variability, we used three different state encodings (see
Section 5.1).
The non-deterministic character of the problem comes with difficulties that are described in the following paragraph: If an action is chosen that leads the agent in the direction of the goal, but because of the slippery factor, it is falling into a hole, it additionally receives a negative reward and creates the following experience:
. If this experience is used for a
Q update, it misleadingly shifts the state-action value away from a positive value. We denote the slippery factor for executing a neighbouring action as
, the resulting rewards for executing the two neighbouring actions as
and
and the reward for executing the intended action as
and can then define the true expected reward for executing
in
as follows:
Following Equation (
3), we define the experience that takes the state-transition function into account and that does not confuse the learner as the expected experience
:
The learner will converge its state-action value
after seeing enough experiences to:
A Q update with received (misleading) experiences comes with the effect of oscillation as the non-deterministic property of the environment creates rewards and follow-up states that might lead in completely opposite directions (e.g., brings the agent closer to the goal vs. ends the episode in a hole). In the original environment (reward for falling into a hole equals 0), this effect would also appear because the different follow-up states hold the same information. If the learner only receives experiences in the form of , the amount of time required to converge to could be decreased.
4.2. Averaging Rewards
The intention of our solution is to reduce the amount of training by the creation of synthetic experiences that are as similar as possible to . As a current limitation, we focus on estimating and use real observed follow-up states . That approach is possible because the environment is discrete (this represents a mandatory precondition of our algorithm). Discrete environments provide a limited amount of states and more important corresponding follow-up states, and it is possible to observe and remember them. In continuous environments, we would also need to predict the follow-up state next to the reward, and for this first investigation of the concept of interpolated experiences, we decided to keep it simple. To compute an accurate estimation of , we need to estimate first.
The set of all rewards that belong to the experiences that start in the same state
and execute the same action
can be defined as:
We use the rewards in
to calculate the average and denote it as
. This value holds as a good estimation of
.
following this, we can then define
as our estimation of
as:
with
The accuracy of this interpolation correlates with the amount of transitions stored in the ER, which start in
and execute
. This comes from the fact that the effect of outliers can be mitigated from enough normal distributed samples. To achieve this, we defined an algorithm that triggers an interpolation after every step the agent takes. A query point
is drawn via a sampling method from the state space, and all matching experiences:
where their starting point
is equal to the query point
, are collected from the ER. Then for every action
, all experiences that satisfy
are selected from
in:
The resulting transitions are used to compute an average reward value
. Utilizing this estimation, a synthetic experience
for every distinct next state:
is created. This results in a minimum of 0 and a maximum of 3 synthetic experiences per action and sums up to a maximum of 12 synthetic transitions per interpolation depending on the amount of stored transitions in the ER. As with the amount of stored real transitions, which can be seen as the combined knowledge of the model, the quality of the interpolated experiences may get better. A parameter
is introduced that determines the minimum amount of stored experiences before the first interpolation is executed. The associated pseudocode is depicted in Algorithm 1.
4.3. Interpolation Component
Stein et al. introduce their Interpolation Component (IC) in [
22]. As already mentioned in
Section 3, we adopted it for our approach. We use it as the underlying basic structure for our interpolation tasks and present it in more detail in the following chapter.
This IC, depicted in
Figure 2, serves as an abstract pattern and consists of a Machine Learning Interface (MLI), an Interpolant, an Adjustment Component, an Evaluation Component and the Sampling Points (SP). The MLI acts as an interface to attached ML components and as a controller for the IC. If it receives a sample, it is handed to the Adjustment Component; there, following a decision function, it is added to or removed from SP. If an interpolation is required, the Interpolation Component fetches the required sampling points from SP and computes, depending on an interpolation technique, an output. The Evaluation Component provides a so-called trust-level as a metric of interpolation accuracy.
Algorithm 1: Reward averaging in IER. |
|
We replaced the SP with the ER. It is realized by a FiFo queue with a maximum length. This queue represents the classic ER and is filled only with real experiences. To store the synthetic transitions another queue, a so-called
ShrinkingMemory is introduced. This second storage is characterized by its decreasing size. Starting at a predefined maximum, it gets smaller depending on the length of the real experience queue. The
Interpolated Experience Replay (IER) has a total size, comprising the sum of the lengths of both queues, as can be seen in
Figure 3. If this size is reached, the length of the ShrinkingMemory is decreased, and the oldest items are removed. This goes on as long as either the real valued queue reaches its maximum length and there is some space left for interpolated experiences or the IER fills up with real experiences. As interpolation is a lot of extra work, it might seem counterproductive to throw such examples away, but this decision was made because of two reasons:
As the learner comes near convergence, randomly distributed experiences might harm the real distribution that is derived by following the actual policy. Following this point, the learner benefits more from real experiences as time goes by.
The quality of the interpolated experiences is unclear and a bad interpolation could harm the learner even more than a misleading real experience. By throwing them away and regularly replacing them with new ones, we try to mitigate this effect.
We also introduced a minimum length for the interpolated storage that is never fallen below. This results in a differing behaviour from the above-explained procedure. If the ShrinkingMemory is instructed to reduce its length, it does this only until it reaches this threshold. Therefore, the maximum length of the IER consists of the real experience buffers maximum length and the minimum length of the synthetic part .
The IER algorithm, as described in
Section 4.2, is located in the Interpolant, and, as stated above, executed in every step. An exhaustive search would need a computation time of
and therefore is not practical for large sized IERs because this operation is executed in every single step. A possible solution for this problem is to employ a so called kd-tree, which represents a multidimensional data structure. Using such a tree, the computation time could be decreased to
[
25]. As the examined problem is very small and consists out of
discrete states, we use another approach to reduce the computation time further to
. To achieve this, we use a dictionary
of size
with keys:
and corresponding values:
with:
This equals an entry for every state-action pair with associated average rewards and distinct next states of all seen transitions. The dictionary is updated after every transition the agent makes.
To evaluate the quality of computed interpolations, an appropriate metric cloud be used in the Evaluation part. This is not implemented yet and left for future work.
5. Evaluation
This chapter first introduces the experimental setup and is followed with a detailed evaluation of the results.
5.1. State Encodings
We investigated three different state encodings:
State Vector Encoding (VE): The state vector encoding is realized with an array of the length of the state space (). The whole vector is filled with zeros, and the entry that corresponds to the actual state is set to 1. This results in an input layer of size 64.
Coordinates Encoding (CE): The coordinates state encoding is realized via a vector with two entries that hold the value for the normalized x- and y-coordinate. An input layer of size two is used here.
Local Knowledge Encoding (LKE): In the local knowledge state encoding, the agent receives a vector with eight entries that corresponds to the surrounding fields of the actual state. The different state types are shown in
Table 1. We utilized an input layer of size 8. In this encoding, we face the problem of perceptual aliasing [
26], as some states have the exact same encoding but in fact are different. These states share their collected follow-up states and rewards, and therefore, it is expected that a small real experience buffer should perform better than a big one because stored experiences are replaced with a higher frequency. Furthermore, the complexity of the problem increases because of these states.
A graphical illustration of the different encodings can be observed in
Figure 4.
5.2. Query Methods
We investigated four different query methods of how to receive :
Random (R): This method draws randomly from the state space. We expect this mode to assist the learner with exploration. Because the new inserted experiences induce a completely differing distribution (), this might also harm the learner. The concept is not feasible for LKE because the sampling of eight random numbers out of five possibilities results in illegal or not existing states most of the time.
Policy Distribution (PD): This method draws a random state from the real experience buffer. The received state distribution resembles the one that is created by the policy. We expect the (possible) harmful effect of inducing a different sample distribution to be mitigated.
Last State (LS): To stay even closer to the distribution created by the policy, this method takes the last state that was saved to the real experience buffer.
Last State—On Policy Action Selection (LS-OnPol): In an attempt to stay even closer to the distribution that the policy creates, we use the LS query method in combination with an altered interpolation step. We only create synthetic experiences for one action that is given by the actual policy. Using this technique, we create synthetic samples for the same experiences the agent observed with a small deviation in the form of the actual of the exploration.
5.3. IER Modes
We investigated three different methods of how the IER was used:
Synthetic Min Size Zero (MSZ): In this mode, we used a minimum size for the synthetic buffer of 0. In this configuration, the learner starts with a lot of synthetic samples in its buffer, but if he comes near convergence, they are replaced with real experiences, and in the end, the agent learns from the real data.
Synthetic Min Size Equals Real Size (MSERS): In this mode, we set the to . The synthetic buffer fills up completely and stays like this. Overt time, the real experience buffer fills to the same size, and in the end, both buffers have the same length. This brings a ratio of synthetic to real experience samples in favour of the interpolated ones in the beginning and ends in an equal distribution of both.
Only Use Interpolated Buffer (OIER): Because (as described in
Section 4.2) we assume that our synthetic samples are even better than the real ones, we also investigated how our approach performs if we only train on them. The maximum length of the IER, in this case, is only related to
because the real examples are never used for learning.
5.4. Hyperparameter
Preliminary experiments revealed the hyperparameters given in
Table 2, which are shared by all experiments.
The used network architectures for the different state encodings are presented in
Table 3. All networks use the same amount of output nodes.
Table 4 shows the different hyperparameter for the IER modes. The value of
for OIER and MSERS varies for LKE from the other encodings and is much smaller; this comes from the assumption (as described in
Section 5.1) that a smaller buffer would help to handle aliasing states because the rotation of the sampling points is increased that way. In contrast, we expect that a bigger buffer helps in the other cases because it corresponds to a bigger knowledge base and consequentially better synthetic experiences.
We used a linearly decaying -greedy as the exploration technique and investigated three different durations (500, 750 and 1000 episodes).
5.5. Experiments
As baseline, we used a DQN with vanilla ER and 100,000 for all three state encodings to compare the different configurations with.
The different constellations of the individual experiments are shown in
Table 5. We measure the average return over the last 100 episodes to obtain a moving average that indicates how often the agent is able to reach the goal in this time. Each experiment was repeated 20 times, and the results are reported as the overall mean values and the observed standard deviations (
SD) over the repetitions.
Each configuration was tested against the baseline, and the differences have been assessed for statistical significance. Therefore, we first conducted
Shapiro–Wilk tests in conjunction with visual inspection of
QQ-plots to determine whether a normal distribution can be assumed. Since this criterion could not be confirmed for any of the experiments, the
Mann–Whitney-U test has been chosen. All measured statistics, comprising the corresponding p-values for the hypothesis tests, can be found in the
Appendix A.
5.6. Experimental Results
Figure 5,
Figure 6 and
Figure 7 depict the results of the best IER configurations, as given in
Table 6. Each Figure holds the results for all investigated exploration phases in this order:
.
Figure 5 shows the result for the VE state encoding,
Figure 6 the result for the CE state encoding and
Figure 7 the result for the LKE state encoding. The graphs for all conducted experiments can be found in the
Appendix B.
Figure 8,
Figure 9 and
Figure 10 picture the size of the IER at a given episode. As the graphs for different
only differ marginally, we chose
to present here. The graphs for experiment 1 and 2 look quite similar, but experiment 3 differs from them. This comes from the fact that we chose smaller values for
for MSERS and OIER because of the aforementioned reasons. The orange curve, indicating OIER-OnPol, shows a less steep increase than the OIER configuration resulting from the creation of fewer synthetic samples.
5.7. Interpretation
All presented best configurations (
Figure 5,
Figure 6 and
Figure 7) outperform their corresponding baseline (red line) and converge on a higher value alongside a steeper increase. This effect even increases with a smaller value of
. The IER approach and the baseline for
remain close to each other after 1400 episodes, and we can expect a similar behaviour from the other experiments if we give them more time. This shows that our approach helps the agent to understand the underlying model of the environment in the early to mid stages of learning (exploration phase). This fits our expectations from
Section 4.2, where we noted that
will converge to
after seeing enough samples, but our synthetic experiences can accelerate this process. The enhanced information encapsulated in a synthetic experience as well as the distribution created by the interpolation process help to reduce oscillations that come from the effect of shifting the Q-function away from
.
The IER graph in
Figure 6a ends slightly below the baseline. As it reaches a higher value, both overall and faster, we still consider it to outperform the baseline.
A closer look at
Table 6 reveals a clear best configuration for experiment 3 (OIER-LS-OnPol). Experiment 2 favours OIER-PD with an outlier for
, but the difference between OIER-PD and MSERS-LS is marginal here (cf.
Appendix A), and therefore, we can declare OIER-PD as the best-performing configuration for this experiment. For the CE encoding, on the other hand, it is not as obvious, and we obtain three different configurations here. Overall, we can observe that the random query method performs poorly in comparison with the others, which fits our expectations from
Section 5.2. The PD query method seems to be the most efficient method for encodings with global knowledge (VE and CE).
The agent limited to local knowledge benefited most from staying as close as possible to the distribution created by the policy. All interpolation techniques, except OnPol, create synthetic experiences for every action that has been performed from the actual state. A synthetic experience is created for every follow-up state that has been reached from this state-action-pair. This should result in a maximum of 12 created experiences for every state. In the case of conceptual aliasing, the agent observes more than three follow-up states, which leads to the creation of many synthetic experiences. This might harm the learner more than it helps. Reducing this amount (by only interpolating experiences for one action) helps the learner. Additionally, here comes the effect of a smaller value of into play because, as stored experiences are replaced more often, the used follow-up states for interpolation are located closer to the trajectory created by the policy, which results in lesser follow-up states for aliasing states.
The fact that the agents from experiment 1 and 2 do not benefit from the OnPol configuration could be explained by the bigger effect of exploration that is induced into the ER by interpolating all actions. This increased spread over the state space seems to assist the learner.
The best results can be observed in the LKE experiments. The problem of the conceptual aliasing that arises here brings an additional difficulty into play and complicates the whole learning process. Our synthetic experiences use the gathered problem knowledge stored in the real transitions and help the learner to understand the problem. As this fact holds for all state encodings, the LKE encoding benefits from the additional focus on promising follow-up states that are replicated in the ER (OnPol).
In conclusion, our approach outperforms the baseline in most of the configurations in every state encoding. Using only synthetic experiences and querying in a way that follows the policy distribution to some extend promises better results.
6. Conclusions and Future Work
We presented an extension for the classic ER used in Deep RL that includes synthetic experiences to speed up and improve learning in non-deterministic and discrete environments. The proposed algorithm interprets stored, actually experienced transitions as an (inaccurate) model of the environment and calculates synthetic -tuples by means of interpolation. The synthetic experiences comprise a more accurate estimate of the expected long-term return a state-action pair promises than a real transition does. We investigated three different state encodings of the FrozenLake8x8-v0 environment from the OpenAI Gym to evaluate our approach for varying network structures and different challenges. To date, the employed interpolation technique is a simple, equally weighted averaging that serves as an initial approach. More complex methods in even more complex problem spaces have to be investigated in the future. The IER approach was compared to the default ER in the FrozenLake8x8-v0 environment from the OpenAI Gym and showed an increased performance in terms of a 3–19% increased overall mean reward of the best-performing configurations. By the investigation of different state encodings, query methods and IER modes, we were able to show that using only synthetic experiences and querying from the distribution created by the policy can assist the learner in terms of performance and speed.
As of yet, the proposed approach is limited to discrete and nondeterministic environments. We plan to develop the IER further to solve more complex problems (increased/continuous state and action space) as well. To achieve this, a solution for the unknown follow-up state is needed, which could also be interpolated or even predicted by a state-transition function that is learned in parallel. Here, the work from [
18] could serve as a possible approach to begin with. A simple, yet nevertheless, more complex problem because of its continuity that is beyond the domain of grid worlds is the MountainCar problem. Other, more complex interpolation techniques have to be examined to adapt our IER approach in this environment. Furthermore, at last, the impact of interpolated experiences on more sophisticated Experience Replay mechanisms, such as Hindsight ER and Prioritized ER, have to be investigated as well.