Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why can't my training loss decrease with ArcFace? It drops to about 21 and stops falling. #3

Open
LHH20000923 opened this issue Jan 8, 2022 · 3 comments

Comments

@LHH20000923
Copy link

No description provided.

@fdbtrs
Copy link
Owner

fdbtrs commented Jan 14, 2022

which database do you use for training, loss function, batch size etc...?

@LHH20000923
Copy link
Author

I use CASIA-WebFace, optim is SGD ,lr=0.1, batchsize=256/512, and didn't use knowledge distillation。
The loss curve is shown below。Is it a problem with the data set?
1

@fdbtrs
Copy link
Owner

fdbtrs commented Jan 26, 2022

For casia dataset, you may need to train the model for at least 50 epochs with lr step of 20 30 and 40. The reported FR result in the paper is based on MS1Mv2

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants