Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Official implementation of the AAAI 2021 paper Deep Bayesian Quadrature Policy Optimization.

License

NotificationsYou must be signed in to change notification settings

Akella17/Deep-Bayesian-Quadrature-Policy-Optimization

Repository files navigation

Akella Ravi Tej1, Kamyar Azizzadenesheli1, Mohammad Ghavamzadeh2, Anima Anandkumar3, Yisong Yue3
1Purdue University,2Google Research,3Caltech


Preprint:arxiv.org/abs/2006.15637
Publication:AAAI-21 (also presented at NeurIPSDeep RL andReal-World RL Workshops 2020)
Project Website:akella17.github.io/publications/Deep-Bayesian-Quadrature-Policy-Optimization/

Bayesian Quadrature for Policy Gradient

MIT licensecontributions welcome

Bayesian quadrature is an approach in probabilistic numerics for approximating a numerical integration. When estimating the policy gradient integral, replacing standard Monte-Carlo estimation with Bayesian quadrature provides

  1. more accurate gradient estimates with a significantly lower variance
  2. a consistent improvement in the sample complexity and average return for several policy gradient algorithms
  3. a methodological way to quantify the uncertainty in gradient estimation.

This repository contains a computationally efficient implementation of BQ for estimating thepolicy gradient integral (gradient vector) and theestimation uncertainty (gradient covariance matrix). The source code is written in amodular fashion, currently supporting three policy gradient estimators and three policy gradient algorithms (9 combinations overall):

Policy Gradient Estimators :-

  1. Monte-Carlo Estimation
  2. Deep Bayesian Quadrature Policy Gradient (DBQPG)
  3. Uncertainty Aware Policy Gradient (UAPG)

Policy Gradient Algorithms :-

  1. Vanilla Policy Gradient
  2. Natural Policy Gradient (NPG)
  3. Trust-Region Policy Optimization (TRPO)

Project Setup

This codebase requires Python 3.6 (or higher). We recommend using Anaconda or Miniconda for setting up the virtual environment. Here's a walk through for the installation and project setup.

git clone https://github.com/Akella17/Deep-Bayesian-Quadrature-Policy-Optimization.gitcd Deep-Bayesian-Quadrature-Policy-Optimizationconda create -n DBQPG python=3.6conda activate DBQPGpip install -r requirements.txt

Supported Environments

  1. Classic Control
  2. MuJoCo
  3. PyBullet
  4. Roboschool
  5. DeepMind Control Suite (viadm_control2gym)

Training

Modular implementation:

python agent.py --env-name <gym_environment_name> --pg_algorithm <VanillaPG/NPG/TRPO> --pg_estimator <MC/BQ> --UAPG_flag

All the experiments will run for 1000 policy updates and thelogs get stored insession_logs/ folder. To reproduce the results in the paper, refer the following command:

# Running Monte-Carlo baselinespython agent.py --env-name <gym_environment_name> --pg_algorithm <VanillaPG/NPG/TRPO> --pg_estimator MC# DBQPG as the policy gradient estimatorpython agent.py --env-name <gym_environment_name> --pg_algorithm <VanillaPG/NPG/TRPO> --pg_estimator BQ# UAPG as the policy gradient estimatorpython agent.py --env-name <gym_environment_name> --pg_algorithm <VanillaPG/NPG/TRPO> --pg_estimator BQ --UAPG_flag

For more customization options, kindly take a look at thearguments.py.

Visualization

visualize.ipynb can be used to visualize the Tensorboard files stored insession_logs/ (requiresjupyter andtensorboard installed).

Results

Vanilla Policy Gradient

Average of 10 runs.

Natural Policy Gradient

Average of 10 runs.

Trust Region Policy Optimization

Average of 10 runs.

Implementation References

Contributing

Contributions are very welcome. If you know how to make this code better, please open an issue. If you want to submit a pull request, please open an issue first. Also see the todo list below.

TODO

  • Implement policy network for discrete action space and test on Arcade Learning Environment (ALE).
  • Add other policy gradient algorithms.

Citation

If you find this work useful, please consider citing:

@article{ravi2020DBQPG,    title={Deep Bayesian Quadrature Policy Optimization},    author={Akella Ravi Tej and Kamyar Azizzadenesheli and Mohammad Ghavamzadeh and Anima Anandkumar and Yisong Yue},    journal={arXiv preprint arXiv:2006.15637},    year={2020}}

[8]ページ先頭

©2009-2026 Movatter.jp