Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fixes for issues #3, #5 and #7. Agent learns better #8

Open
wants to merge 2 commits into
base: master
Choose a base branch
from

Conversation

praveen-palanisamy
Copy link

Below is the summary of the contributions made by this PR:

This was referenced Nov 1, 2017
@SSARCandy
Copy link

thanks for your contribution! It works on pytorch 0.2 👍

BTW, can I ask why doing gradient clipping, is that matters to performance?
thanks for your code again :)

@praveen-palanisamy
Copy link
Author

Glad to hear that my contributions helped you.

Clipping the gradient will make sure that the gradients don't "explode" which is a common problem encountered when using gradient descent algorithms with neural networks.
In this case with DQN, gradient clipping will ensure that the optimization algorithm only takes small (in magnitude) steps in the direction pointed to by the gradient. Making a larger descent step and hence a big update to the Q-value function approximation might throw the approximation off from (converging to) the optimal values.

Hope the explanation helps.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants