-
Notifications
You must be signed in to change notification settings - Fork 301
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adding the following models: Vicuna, Koala, Pythia-Chat-Base-7B, GPT-NeoXT-Chat-Base-20B #1268
Adding the following models: Vicuna, Koala, Pythia-Chat-Base-7B, GPT-NeoXT-Chat-Base-20B #1268
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Seems good. Not sure about the license implications.
Co-authored-by: Cameron Fairchild <cameron@fairchild.dev>
Co-authored-by: Cameron Fairchild <cameron@fairchild.dev>
Co-authored-by: Cameron Fairchild <cameron@fairchild.dev>
Co-authored-by: Cameron Fairchild <cameron@fairchild.dev>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good. Thanks for adding tweaks/suggestions @camfairchild
Co-authored-by: Cameron Fairchild <cameron@fairchild.dev>
@ViktorThink Nice work! A few Qs:
Command ran:
Error message (indicating there is no tokenizer found, which makes sense because there is not one mentioned in the readme at all and is not discussed how to move from the final Koala weights to prepare the miner.)
The error is on line 45: |
@ifrit98 Ah, nice spotting. Yes, it is the same tokenizer as llama. I'll update the documentation later today to show exactly how to do it. |
@ifrit98 I had missed to add the last step of the conversion process before starting the miner.
EasyLM has not released any converted version of the models, only delta, but I saw on Huggingface there are already converted checkpoints that could be loaded directly. I don't actually have the Llama weights in torch format, so I will look into adding a second option for conversion for those who have Llama in HF format but not in torch. |
Converted and miner is up and running. Nice work! Closing. |
Adding support for using Vicuna, Koala, Pythia-Chat-Base-7B, GPT-NeoXT-Chat-Base-20B for mining on netuid 1.
Closing my pervious pull requests 1266 and 1261 since this new pull request includes those models, and it was cleaner to have them all in the same branch.
Vicuna, Pythia and NeoXT has been tested that they work properly on netuid 1. I plan to test Koala today on netuid 1 if there are active validators.