Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Chaper 14 Deterministic policy gradients results are quite noisy. #86

Open
isu10503054a opened this issue Oct 27, 2020 · 3 comments
Open

Comments

@isu10503054a
Copy link

isu10503054a commented Oct 27, 2020

In the results of Chapter 14 Deterministic policy gradients in the book,
why the training is not very stable and noisy?


擷取
擷取2

I read the content repeatedly, but I still don’t understand why.

@Shmuma
Copy link
Collaborator

Shmuma commented Oct 27, 2020 via email

@isu10503054a
Copy link
Author

isu10503054a commented Oct 28, 2020

Random weights initialization adds randomness to initial starting point. Usage if different parallel environments also might add stochastisity вт, 27 окт. 2020 г., 12:01 isu10503054a notifications@github.com:

In the results of Chapter 14 Deterministic policy gradients in the book, why the training is not very stable and noisy? I read the content repeatedly, but I still don’t understand why. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub <#86>, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAAQE2WTJOWPGQGYY3MOTRLSM2D5XANCNFSM4TAQL7BQ .

Is there any hyperparameter in the source code that can modification to improve this situation?
thx

@Shmuma
Copy link
Collaborator

Shmuma commented Oct 28, 2020 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

No branches or pull requests

2 participants