Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

No optimizer.bin in Step 2 #18

Open
mihara-bot opened this issue May 20, 2024 · 3 comments
Open

No optimizer.bin in Step 2 #18

mihara-bot opened this issue May 20, 2024 · 3 comments

Comments

@mihara-bot
Copy link

Hi,
When I run "Step 2: Building the gradient datastore"
FileNotFoundError: [Errno 2] No such file or directory: '../out/llama2-7b-p0.05-lora-seed3/checkpoint-1688/optimizer.bin'
I check the folder "llama2-7b-p0.05-lora-seed3" generate from Step 1, only files optimizer.pt in checkpoint-1688.

I noticed that in other issues, some scholars had problem on generating optimizer.pt. I think my problem is different from these.
May I kindly ask for you help?

@xiamengzhou
Copy link
Collaborator

Could you confirm that you are using transformers==4.36.2? Transformers package gets frequently updated especially the part that supports distributed training :(

@pengming617
Copy link

i have the same error while using transformers==4.36.2

No such file or directory: 'out/llama2-7b-p0.05-lora-seed3/checkpoint-1688/optimizer.bin'

@xiamengzhou
Copy link
Collaborator

@pengming617 Please refer to this thread! I think you need to use fsdp to get optimizer.pt.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants