-
Notifications
You must be signed in to change notification settings - Fork 974
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
chat/base.py: extend checkpoint_dir before accessing it #1575
Conversation
The reason why auto-download is so far down there (compared to the other places) is that I had to move it below the LoRA merging, because otherwise it will download a model if you want to use LoRA weights. |
I also didn't notice the missing CPU warning because I was running it on GPU 😅. The following reorg might work ... |
I think this should be good now. Feel free to merge if you agree |
There is a problem. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Oh I see now what you meant previously.
Are we good to merge? What do you think? |
Yes, I think so. But one last question, after the update, have you double-checked / tested it on custom paths that don't start with `"checkpoints", like litgpt chat my_custom_dir/google/gemma-2-9b-it |
No 😊.
|
Nice, thanks for checking! Looks all good to me now :) |
Hi there 👋
Inside
auto_download_checkpoint
the code also extends the path by prependingcheckpoints/
if it's not provided.Since currently it's done after
check_file_size_on_cpu_and_warn
, the commandwill fail.