Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

training for classification does not converge #19

Open
zwangab91 opened this issue Jul 27, 2018 · 2 comments
Open

training for classification does not converge #19

zwangab91 opened this issue Jul 27, 2018 · 2 comments

Comments

@zwangab91
Copy link

zwangab91 commented Jul 27, 2018

I tried to train the classification models for alexnet and inception, with the hyperparameters in train.py ( 'learning_rate_decay_type': 'exponential', 'learning_rate': '0.01', 'learning_rate_decay_factor': '0.1'), but the loss fluctuates around 6 and 11 respectively for the two models. I tried to tune the learning rate in the range from 1e-5 to 0.1, but the training still shows no sign of convergence (even after 10,000 steps). Could you inform me of the hyperparameters chosen for the training of the classification models in order to reproduce the results, and the final values of the cross-entropy loss?
screen shot 2018-07-27 at 3 08 18 pm

@yuantailing
Copy link
Owner

yuantailing commented Jul 27, 2018

  1. We didn't tune hyper-parameters. The hyper-parameters we used is what you find in git.
  2. I forget the cross-entropy loss. But the loss is only cross-entropy loss as I know.
  3. 10,000 steps is far from convergence, we trained 100,000 steps. 1 epoch is 800,000 / 64 = 12,500. Don't pray net learning well before 1 epoch.

Please be patient, I believe you can reproduce the exact result (the only problem is random seed) without any modification.

@zwangab91
Copy link
Author

Thanks! The loss did drop down to around 2 after 5 epochs of training.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants