-
Notifications
You must be signed in to change notification settings - Fork 123
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DDPG implementation #65
Comments
Hi Roman, In deer, the actor and the critic are two different objects (two different neural networks): The place where the actions are given as input in the NN depends on the architecture you decide for your neural network. So indeed you can decide to provide them deeper in the NN. It's a design architecture but it does not really change anything in the logic of the algorithm. Best, |
That's true. In the ddpg paper the author used furthermore different learning rates for the actor and critic. The final layer for the actor is a 'tanh'. Do you think it is worth to add this to deer? |
Yes, that may possibly be easier for the end user. So we would have two classes: a NN_actor and a NN_critic. For the Q-network, we would then use the NN_critic class. If you want to suggest a PR, we can look into it. |
I will work out a proposal. Best, |
Hey VinF,
thanks for your work!
I have questions about the DDPG implementation in deer.
Patrick Emami recommends in http://pemami4911.github.io/blog/2016/08/21/ddpg-rl.html to use for the actor and critic two functions in separate classes.
Additionally, he adds the action tensor in the 2nd hidden layer of the Critic Network.
Is my assumption correct that the ddpg implementation in deer is different?
King regards,
Roman
The text was updated successfully, but these errors were encountered: