Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

gradient_accumulation_batch_size missing in trainer #65

Open
carbonz0 opened this issue Oct 24, 2019 · 3 comments
Open

gradient_accumulation_batch_size missing in trainer #65

carbonz0 opened this issue Oct 24, 2019 · 3 comments

Comments

@carbonz0
Copy link

I found
gradient_accumulation_batch_size exists in several scibert conf, such as
https://github.com/allenai/scibert/blob/8562a120e6788dcbadbe05ef7fd4463dee17ee59/allennlp_config/ner.json
but allennlp trainer doesn't have this param,
https://github.com/allenai/allennlp/blob/master/allennlp/training/trainer.py

@ibeltagy
Copy link
Collaborator

Yes, AllenNLP doesn't support gradient accumulation. We have it implemented in our fork of allennlp (check requirements: https://github.com/allenai/scibert/blob/master/requirements.txt)

@carbonz0
Copy link
Author

carbonz0 commented Nov 1, 2019

got it, thank you!

@xiaoruijiang
Copy link

Hi, The installation command in the readme returned the following "allennlp version not found error". Can I ask what is the right way to install the tool and reproduce the results?

Thank you very much!

Obtaining allennlp from git+git://github.com/ibeltagy/allennlp@fp16_and_others#egg=allennlp (from -r requirements.txt (line 1))
WARNING: Discarding git+git://github.com/ibeltagy/allennlp@fp16_and_others#egg=allennlp. Command errored out with exit status 128: git rev-parse HEAD Check the logs for full command output.
ERROR: Could not find a version that satisfies the requirement allennlp (unavailable)
ERROR: No matching distribution found for allennlp (unavailable)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants