Skip to content

Latest commit

 

History

History
34 lines (24 loc) · 2.07 KB

README.md

File metadata and controls

34 lines (24 loc) · 2.07 KB

Important: Deprecated version. Update to latest arxiv version coming soon. Please use the resources in our main repository in the meanwhile.

RUDDER efficiently learns optimal policies in finite Markov decision processes with delayed rewards. With the following links you can find:

Baselines incl. RUDDER

Uses RUDDER on PPO to solve environments with delayed rewards. Our RUDDER paper is available at https://arxiv.org/abs/1806.07857.

See https://github.com/widmi/rudder-a-practical-tutorial for a quick tutorial on reward redistribution using a pytorch implementation and code-snippets that you can run on your CPU.

Our code is based on the OpenAI Baselines package, in which we included our implementation of RUDDER for PPO for ATARI games with delayed rewards. Additionally, the Tensorflow Layer Library (TeLL) package (v1.0.0) is required for RUDDER.

RUDDER for PPO2 and documentation for the configuration is located in the folder baselines/ppo2_rudder.

Modified/new files:

  • baselines/common/atari_wrappers.py
  • baselines/common/distributions.py
  • baselines/common/vec_env/vec_frame_stack.py
  • baselines/ppo2_rudder/
  • logger.py
  • README.md

Videos are available at:

RUDDER - Venture

RUDDER - Bowling