Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Entropy calculation not useful #8

Open
Myrkiriad-coder opened this issue Apr 18, 2021 · 1 comment
Open

Entropy calculation not useful #8

Myrkiriad-coder opened this issue Apr 18, 2021 · 1 comment

Comments

@Myrkiriad-coder
Copy link

Describe the bug
In ppo_continous_tensorflow.py, when you calculate entropy with:
dist_entropy = tf.math.reduce_mean(self.distributions.entropy(action_mean, self.std))
since entropy only depends on std and std is a static parameter, dist_entropy has always the same value all the time.
Thus, entropy loss has no effect on learning.

To Reproduce
Launch any env and stop your debugger on dist_entropy. Check that it has the same value for every batch at any given point during learning.

Expected behavior
Std shall not be static but somehow represent real prediction confidence of the network.

@wisnunugroho21
Copy link
Owner

Sorry if I'm so late to reply. Thank you for the advice.
Actually, you can set the entropy coefficient to 0 if you use static parameters.

I really think using a neural network to calculate the std is much better than using static parameters.
I forgot to do it in this repository

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants