Exploration in Meta-Reinforcement Learning

$24.99 $18.99

By turning in this assignment, I agree by the Stanford honor code and declare that all of this is my own work. Overview In this assignment, we will be exploring meta-reinforcement learning algorithms. In par-ticular, you will Experiment with black-box meta-RL methods [1,2] trained end-to-end. Answer conceptual questions about a popular decoupled meta-RL algorithm, Pearl…

5/5 – (2 votes)

You’ll get a: zip file solution

 

Categorys:

Description

5/5 – (2 votes)

By turning in this assignment, I agree by the Stanford honor code and declare that all of this is my own work.

Overview

In this assignment, we will be exploring meta-reinforcement learning algorithms. In par-ticular, you will

  1. Experiment with black-box meta-RL methods [1,2] trained end-to-end.

  1. Answer conceptual questions about a popular decoupled meta-RL algorithm, Pearl [4].

  1. Conceptually compare end-to-end with decoupled optimization.

  1. Implement components of Dream [3] to achieve the best of both groups of approaches.

To answer the questions in this homework, it may be helpful to refer to the Dream paper, which we’ve included in the zip of the starter code under dream paper.pdf.

Submission: To submit your homework, submit one PDF report to Gradescope containing written answers/plots to the questions below and a zip file of your code. The PDF should also include your name and any students you talked to or collaborated with.

Setup: Please download and unzip the starter code, and then follow the instructions in the README. Download the dependencies, which we recommend you do via a virtualenv as follows:

  • python3 -m virtualenv venv

  • source venv/bin/activate

  • pip install -r requirements.txt

Please ensure that you’re using Python3.7. Otherwise, installing the dependencies will not work.

Code Overview: The main entry point for the code is via dream.py and rl2.py, which are the training scripts for Dream and RL2 respectively. Both of these can be invoked as follows:

$ python3 fscriptg.py fexperiment nameg -b environment=n”mapn”

In this invocation, fscriptg can either be dream.py or rl2.py and fexperiment nameg can be any string with no white spaces. Results from this invocation are saved under experiments/fexperiment nameg. For example, to launch the Dream training script and save the results to experiments/dream, you would run:

$ python3 dream.py dream -b environment=n”mapn”

You can pass the –force overwrite flag to run another experiment with the same ex-periment name, which will overwrite any previously saved files at the corresponding ex-periment directory. If you do not pass this flag, the scripts will not allow you to run two experiments with the same experiment name.

There are two important subdirectories in each experiment directory:

Tensorboard: Each experiment includes a experiments/experiment name/tensorboard subdirectory, which will log important statistics about the training run, including the meta-testing returns under the rewards/test tag and the meta-training returns un-der the rewards/training tag. To view these, point Tensorboard at the appropriate directories. All curves are plotted with two versions, one where the x-axis is number of meta-training trials under tensorboard/episode and one where the x-axis is the number of environment steps tensorboard/step.

Visualizations: Each experiment also includes a experiments/experiment name/visualize subdirectory. This directory includes .gif videos of the agent during meta-testing

and is structured as follows. The top level of subdirectories identify how many meta-training trials have elapsed before the video.

In experiments run with dream.py, the exploration episode is saved under fvideo numg-exploration.gif and the exploration episode is saved under

fvideo numg-exploitation.gif. For example, the video under experiments/dream/visualize/10000/0-exploration.gif is the first exploration meta-testing episode after 10000 meta-training trials.

In experiments run with rl2.py, fvideo numg.gif contains both the exploration and exploitation episodes, with the exploration episode first. For example, the video under experiments/rl2/visualize/10000/0.gif contains the first exploration and exploitation episode after 10000 meta-training trials.

You will implement two short methods inside the embed.py file.

Problem 0: Grid World Navigation with Buses (4 Points)

quickly as possible. The episode ends if either the agent goes to any goal location (correct or incorrect) or if 20 timesteps have passed.

Each task is associated with a problem ID . The only thing that changes between tasks is the bus destinations: i.e., which colored bus appears in which outer corner. Therefore, there are 4! = 24 different tasks. These tasks are uniformly sampled during meta-training and meta-testing.

Throughout the assignment, we consider the meta-RL setting with one exploration episode and one exploitation episode. The objective is to maximize the returns achieved in the ex-ploitation episode, which we refer to as the exploitation returns. Note that the returns achieved in the exploration episode do not matter. During the exploitation episode, the agent is allowed to condition on the exploration episode exp = (s0; a0; r0; : : :).

  1. What returns are achieved by only taking the move action to get to the goal, without riding any buses: i.e., directly walking to the goal? (1 point)

  2. If the bus destinations (i.e., the problem ID) were known, what is the optimal returns that could be achieved in a single exploitation episode? Describe an exploitation policy that achieves such returns given knowledge of the bus destinations. (1 point)

  1. Describe the exploration policy that discovers all of the bus destinations within the fewest number of timesteps. (1 point)

  2. Given your answers in b) and c), what is the optimal exploitation returns achievable by a meta-RL agent? (1 point)

For Problems 1 and 3, note that in the visualizations saved under experiments/experiment name/visualize:

The agent is rendered as a red square.

The grid cells that the agent has visited in the episode are rendered as small origin squares.

There are four pairs of buses, rendered as blue, pink, cyan, and yellow squares. The map is rendered as a black square.

The goal state is rendered as a green square, which obscures one of the buses.

Problem 1: End-to-End Meta-Reinforcement Learning (5 Points)

In this problem, we’ll analyze the performance of end-to-end meta-RL algorithms on the grid world. To do this, start by running the RL2 agent on the grid world navigation task for 50; 000 trials by running the below command. This should take approximately 1 hour.

python3 rl2.py rl2 -b environment=n”mapn”

  1. Examine the Tensorboard results under the tag reward/test in the experiments/rl2 directory. To 1 decimal place, what is the average meta-testing exploitation returns RL2 achieves after training? (1 point)

  1. Examine the videos saved under experiments/rl2/visualize/36000/. Describe the exploration and exploitation behaviors that RL2 learns. (2 points).

  1. Does RL2 achieve the optimal returns? Based on what you know about end-to-end meta-RL, do these results align with your expectations? Why or why not? (2 points).

Problem 2: Decoupled Meta-Reinforcement Learning (8 Points)

In this problem, we’ll examine Pearl [4], a common decoupled meta-RL algorithm. Recall that Pearl consists of 3 key components: a prior over latent z’s p(z), a policy (a j s; z) that conditions on the latents, and a posterior q(z j f igNi=1) over the latents, given the past N episodes 1; : : : ; N . Intuitively, the latent z can be thought of as an approximation of the task, so the prior and posterior model uncertainty over what the current task is, and the policy attempts to achieve high returns, assuming that z identifies the current task.

On a new task, Pearl explores via Thompson sampling. First, Pearl samples a z p(z) from the prior, and rolls out an episode from the policy (a j; s; z), which yields an episode 1. Then, a new latent is sampled from the posterior z q(z j 1) and another episode is rolled out from the policy given the new z. This process is repeated for as many episodes as is allowed in the setting. Since we’re considering only a single exploration episode and single exploitation episode, a latent is only sampled once from the posterior in our setting.

In this problem, we consider an idealized version of Pearl, where z is actually the problem ID , and the policy (a j s; ) is the optimal policy for each task . That is, (a j s; ) acts optimally assuming that is the true problem ID. Below, we’ll directly refer to instead of z.

  1. In the grid world, the prior p( ) is uniform over the 24 tasks. After observing a single exploration episode , what is the new posterior over tasks p( j )? Hint: think about what the policy does. (2 points)

  2. What is the expected returns achieved by Pearl given a single exploration episode? Hint: there are two cases to consider. Show your work. (4 points)

  1. Does this idealized version of Pearl achieve the optimal returns? Based on what you know about decoupled meta-RL algorithms, do these results align with your expectations? Why or why not? (2 points)

Problem 3: Dream (9 Points)

In Problem 1d) and Problem 2c) we observed some shortcomings of end-to-end and exist-ing decoupled meta-RL approaches. In this problem, we’ll implement some components

of Dream, which attempts to address these shortcomings, given the assumption that each meta-training task is assigned a unique problem ID . During meta-testing, Dream does not assume access to the problem ID.

From a high level, Dream works by separately learning exploration and exploitation. To learn exploitation, Dream learns an exploitation policy task(a j s; z) that tries to maximize returns during exploitation episodes, conditioned on a task encoding z . Dream learns a encoder F (z j ) to produce the task encoding z from the problem ID . Critically, this encoder is trained in such a way that z contains only the information necessary for the exploitation policy task to solve the task and achieve high returns.

By training the encoder in this way, Dream can then learn to explore by trying to recover the information contained in z . To achieve this, Dream learns an exploration policy exp, which produces an exploration trajectory exp = (s0; a0; r0; : : 🙂 when rolled out during the exploration episode. To recover the information contained in z, Dream tries to maximize the mutual information between the encoding z and the exploration trajectory exp:

max I(F (z j ); exp):

This objective can be optimized by maximizing the following variational lower bound:

J (!; ) = E ;z F ; exp exp [log q!(z j exp)]

where q!(z j exp) is a learned decoder. Note that this decoder enables us to convert an exploration trajectory exp into a task encoding z that the exploitation policy uses. This is critical for meta-test time, where the problem ID is unavailable, and z cannot be computed via the encoder F (z j ).

The objective J (!; ) is optimized with respect to both the decoder q! and the exploration policy exp:

a) For simplicity, we parametrize the decoder q!(z j exp) as a Gaussian N (g!( exp); 2I) centered at a learned g!( exp) with unit variance. Then, log q!(z j exp) equals neg-ative mean-squared error g!( exp) zk22 plus some constants independent of !. Overall, maximizing J (!; ) with respect to the decoder parameters ! is equal to minimizing the below mean-squared error with respect to !:

Ez F ( ) kg!( exp) zk22 :

Code: Fill in the compute losses method of the EncoderDecoder in embed.py to implement the above equation for optimize J (!; ) with respect to the decoder pa-rameters !. (2 points)

  1. To optimize J (!; ) with respect to the exploration policy parameters , we expand out J (!; ) as a telescoping series:

J (!; ) = E ;z F ( )[log q!(z j s0)]+E ;z F ( ); exp exp

T 1

log q!(z j :expt+1) log q!(z j :expt)i

;

h t=0

X

where :expt denotes the exploration trajectory up to timestep t: (s0; a0; r0; : : : ; st). Only the second term depends on the exploration policy, and since it occurs per timestep, it can be maximized by treating it as the following intrinsic reward function rtexp, which we can maximize with standard reinforcement learning:

rtexp(at; rt; st+1; texp1; ) = Ez F ( ) log q!(z j :expt+1]) log q!(z j :expt) :

(1)

Note that :expt+1 is equal to :expt with the additional observations of (at; rt; st+1). The exploration policy is optimized to maximize this intrinsic reward rtexp instead of the extrinsic rewards rt, which will maximize the objective J (!; ). Intuitively, rtexp is the “information gain” representing how much additional information about z – which encodes all the information to solve the task – the tuple (at; rt; st+1) contains over what was already observed in :expt.

Code: Implement the reward function rtexp(at; rt; st+1; texp1; ) by filling in the label rewards function of EncoderDecoder in embed.py. Note that you’ll need to make the same substitution for log q!(z j exp) in Equation (1) that we used in part a). (2 points)

c) Check your implementation by running Dream:

python3 dream.py dream -b environment=n”mapn”

Submit the plot for test returns under the tag rewards/test from the experiments/dream directory. Submit the plot under tensorboard/step, not the plot under tensorboard/episode. If your implementation in part a) and b) is correct, you should see the test returns training curve improve within 10 minutes of training. By around 25 minutes, the test returns curve should begin to look different from RL2. (2 points)

  1. Does Dream achieve optimal returns in your results from c)? Based on what you know about Dream, do these results align with your expectations? Why or why not?

(2 points)

  1. Inspect the videos saved under experiments/dream/visualize/28000 or a later step after Dream converges. Describe the exploration and exploitation behaviors that Dream has learned. (1 point)

References

  1. Yan Duan, John Schulman, Xi Chen, Peter L. Bartlett, Ilya Sutskever, Pieter Abbeel. RL2: Fast Reinforcement Learning via Slow Reinforcement Learning. https://arxiv.org/abs/ 1611.02779

  1. Jane X Wang, Zeb Kurth-Nelson, Dhruva Tirumala, Hubert Soyer, Joel Z Leibo, Remi Munos, Charles Blundell, Dharshan Kumaran, Matt Botvinick. Learning to Reinforcement Learn. The Annual Meeting of the Cognitive Science Society. (CogSci) 2017. https:// arxiv.org/abs/1611.05763

Evan Zheran Liu, Aditi Raghunathan, Percy Liang, Chelsea Finn. Decoupling Ex-ploration and Exploitation for Meta-Reinforcement Learning without Sacrifices. https: //arxiv.org/abs/2008.02790

  1. Kate Rakelly, Aurick Zhou, Deirdre Quillen, Chelsea Finn, Sergey Levine. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables. https://arxiv. org/abs/1903.08254

8

Exploration in Meta-Reinforcement Learning
$24.99 $18.99