Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The effect of NormalizedEnv #12

Open
pengzhi1998 opened this issue Jan 21, 2021 · 2 comments
Open

The effect of NormalizedEnv #12

pengzhi1998 opened this issue Jan 21, 2021 · 2 comments

Comments

@pengzhi1998
Copy link

Hi, thank you for this great implementation!!
However, I'm not very sure about the effect of normalized_env.py. Actually, if I remove it, the results seem to be worse than not removing it. What does it do?
Look forward to your reply!

@zhihanyang2022
Copy link

zhihanyang2022 commented Apr 15, 2021

There are several levels for my answer:

  • Strangely, I had to change the method names from _action and _reverse_action to action and reverse_action for the code to work - maybe this has to do with the gym version (mine has version 0.18.0).
  • If you try adding print statements in the action and reverse_action methods, you would see that action is being called repeatedly but reverse_action is never called.
  • The reason why Gym has the ActionWrapper class and the action method is that: "Used to modify the actions passed to the environment." See reference at https://alexandervandekleut.github.io/gym-wrappers/. I'm actually not sure about the reverse_action method so I would welcome any other thoughts on this.
  • So basically the actor network would output something in the range of Tanh, but you can imagine an environment in which the range of possible actions is, for example, [2, 7].

@pengzhi1998
Copy link
Author

Thank you for your response!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants