Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support Qwen type llm's #638

Closed
5 tasks done
CyberTimon opened this issue Sep 26, 2023 · 7 comments
Closed
5 tasks done

Support Qwen type llm's #638

CyberTimon opened this issue Sep 26, 2023 · 7 comments
Labels
enhancement New feature or request

Comments

@CyberTimon
Copy link

⚠️ Please check that this feature request hasn't been suggested before.

  • I searched previous Ideas in Discussions didn't find any similar feature requests.
  • I searched previous Issues didn't find any similar feature requests.

🔖 Feature description

With the new qwen 14b model it would be very nice to have axolotl support.

✔️ Solution

-Support finetuning of Qwen type llm's

❓ Alternatives

Normal finetuning but it's unoptimized

📝 Additional Context

No response

Acknowledgements

  • My issue title is concise, descriptive, and in title casing.
  • I have searched the existing issues to make sure this feature has not been requested yet.
  • I have provided enough information for the maintainers to understand and evaluate this request.
@CyberTimon CyberTimon added the enhancement New feature or request label Sep 26, 2023
@NanoCode012
Copy link
Collaborator

Could you please try with trust_remote_code: true and using AutoModelCausalLM and AutoTokenizer? Is there an error?

@Peter-Devine
Copy link

I'd like to second this. trust_remote_code: true seems to work for full finetuning, but I get errors for qlora training. Is fully tested support for Qwen on the roadmap for this project? It would be really useful for CJK languages!

@CheshireAI
Copy link

I get errors: raise ValueError("Adding unknown special tokens is not supported")
ValueError: Adding unknown special tokens is not supported
trying to use qlora with qwen

@kostum123
Copy link

Has anyone been able to fine-tune it with qlora?

@NanoCode012
Copy link
Collaborator

I get errors: raise ValueError("Adding unknown special tokens is not supported") ValueError: Adding unknown special tokens is not supported trying to use qlora with qwen

@CheshireAI , Hm, this is an upstream problem. If you want this feature, you should ask at transformers.

@Peter-Devine @kostum123 , what's the error with running qlora?

@NanoCode012
Copy link
Collaborator

Hello, I apologize for the long delay. This should be fixed here: #894

Please let me know how it goes.

@CyberTimon
Copy link
Author

Thank you very much! Will try it out soon.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

5 participants