Skip to content

alexlioralexli/rllab-finetuning

Repository files navigation

Docs Circle CI License Join the chat at https://gitter.im/rllab/rllab

rllab

Sub-policy Adaptation for Hierarchical Reinforcement Learning

To run experiments for the paper Sub-policy Adaptation for Hierarchical Reinforcement Learning, navigate to sandbox/finetuning/README.md to view instructions.

Citing Sub-policy Adaptation for Hierarchical Reinforcement Learning

If you use our code for academic research, you are highly encouraged to cite the following paper:

Credit for project code

We built off of the original rllab code as well as the SNN4HRL code developed by Carlos Florensa (UC Berkeley / Covariant). Alexander Li (UC Berkeley / CMU) was the main developer on this project.

rllab

rllab is a framework for developing and evaluating reinforcement learning algorithms. It includes a wide range of continuous control tasks plus implementations of the following algorithms:

rllab is fully compatible with OpenAI Gym. See here for instructions and examples.

rllab only officially supports Python 3.5+. For an older snapshot of rllab sitting on Python 2, please use the py2 branch.

rllab comes with support for running reinforcement learning experiments on an EC2 cluster, and tools for visualizing the results. See the documentation for details.

The main modules use Theano as the underlying framework, and we have support for TensorFlow under sandbox/rocky/tf.

Documentation

Documentation is available online: https://rllab.readthedocs.org/en/latest/.

Citing rllab

If you use rllab for academic research, you are highly encouraged to cite the following paper:

Credits

rllab was originally developed by Rocky Duan (UC Berkeley / OpenAI), Peter Chen (UC Berkeley), Rein Houthooft (UC Berkeley / OpenAI), John Schulman (UC Berkeley / OpenAI), and Pieter Abbeel (UC Berkeley / OpenAI). The library is continued to be jointly developed by people at OpenAI and UC Berkeley.

Slides

Slides presented at ICML 2016: https://www.dropbox.com/s/rqtpp1jv2jtzxeg/ICML2016_benchmarking_slides.pdf?dl=0

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published