-
Notifications
You must be signed in to change notification settings - Fork 323
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RL models clean up #112
RL models clean up #112
Conversation
Hello @djbyrne! Thanks for updating this PR. There are currently no PEP 8 issues detected in this Pull Request. Cheers! 🍻 Comment last updated at 2020-07-14 15:48:24 UTC |
Codecov Report
@@ Coverage Diff @@
## master #112 +/- ##
=======================================
Coverage 91.91% 91.92%
=======================================
Files 77 78 +1
Lines 3944 4010 +66
=======================================
+ Hits 3625 3686 +61
- Misses 319 324 +5
Continue to review full report at Codecov.
|
self.episode_reward = 0 | ||
self.episode_steps = 0 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
shall this be rater at the beginning rather than the end?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Im not sure I understand
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
that you reset the episode_steps
and the other when it is done... so shall it be more logical to reset it before you start rather?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah I see. The done
is a local variable that is retrieved after taking a step on line 72, so the done check must come after that, so I think it makes more sense to do it at the end
…jbyrne/pytorch-lightning-bolts into enhancement/rl_models_clean_up
This pull request is now in conflict... :( |
Before submitting
What does this PR do?
Refactors RL models to use train_batch structure as seen in #107
general clean up of docstrings and methods
PR review
Did you have fun?
👍