Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

📌 AutoAWQ Roadmap #32

Closed
21 of 30 tasks
casper-hansen opened this issue Sep 6, 2023 · 11 comments
Closed
21 of 30 tasks

📌 AutoAWQ Roadmap #32

casper-hansen opened this issue Sep 6, 2023 · 11 comments

Comments

@casper-hansen
Copy link
Owner

casper-hansen commented Sep 6, 2023

Optimization

More models

Ease of access

Software integration and quality

@casper-hansen casper-hansen pinned this issue Sep 6, 2023
@flexchar
Copy link

Hey Casper, first of all, amazing work!

I'm actually really curious - what's the reasoning behind supporting legacy models such as GPT-2 or GPT-J/OPT that are already in?

In my perception, the latest developments mostly on MPT/Llama 2 are by orders of magnitude better than the legacy models.

@casper-hansen
Copy link
Owner Author

Hey Casper, first of all, amazing work!

I'm actually really curious - what's the reasoning behind supporting legacy models such as GPT-2 or GPT-J/OPT that are already in?

In my perception, the latest developments mostly on MPT/Llama 2 are by orders of magnitude better than the legacy models.

Supporting older models is on the roadmap because people still use those models and ask for them. However, I do try to focus my efforts on optimizing the newer models.

@heiqilin1985
Copy link

yi-34b 能支持吗?看数据这个模型很牛叉啊。

@casper-hansen
Copy link
Owner Author

yi-34b 能支持吗?看数据这个模型很牛叉啊。

Yi is now supported on the main branch

@SinanAkkoyun
Copy link

Can you please implement Phi 1.5 support? Thank you for all the amazing work!

@xTayEx
Copy link

xTayEx commented Dec 3, 2023

Hi Casper, thank you for your wonderful work! I wonder if there is some tutorial for adding support for new model? I have noticed that Baichuan is on the roadmap. I would like try to add support for this model, could you please give me some pointer on how to support new model?

@casper-hansen
Copy link
Owner Author

@xTayEx I do not have a written guide, but here are the steps:

  1. Create a model class BaichuanAWQForCausalLM
  2. Add the model to the model map https://github.com/casper-hansen/AutoAWQ/blob/main/awq/models/auto.py#L6
  3. Import the model here https://github.com/casper-hansen/AutoAWQ/blob/main/awq/models/__init__.py

For creating the model class, look into the llama class or other classes to see how they are defined.

@casper-hansen
Copy link
Owner Author

Can you please implement Phi 1.5 support? Thank you for all the amazing work!

Phi 1.5 support has been attempted, but they have a very unusual model definition. Until it's been standardized, I am not sure I will support it.

@SinanAkkoyun
Copy link

Phi 1.5 support has been attempted, but they have a very unusual model definition. Until it's been standardized, I am not sure I will support it.

Oh :( Do you mean until a new phi model comes out?
Phi 1.5 is such an amazing model for so many applications

What would roughly be the steps to implement it on our own?

@christian-ci
Copy link

Hi @casper-hansen First of all thank you for the Amazing work. From my understanding there is an AWQ TheBloke Mixtral 8x7b Base Instruct version. I tried to run inference on it and ran into issues. Would this model be supported? Also is there a way to contribute with a donation?

@casper-hansen casper-hansen unpinned this issue Mar 1, 2024
@casper-hansen
Copy link
Owner Author

We achieved most items on the roadmap, so closing this for now to focus on other things.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants