Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

📌 AutoAWQ Roadmap #32

Closed
21 of 30 tasks
casper-hansen opened this issue Sep 6, 2023 · 11 comments
Closed
21 of 30 tasks

📌 AutoAWQ Roadmap #32

casper-hansen opened this issue Sep 6, 2023 · 11 comments

Comments

@casper-hansen
Copy link
Owner

casper-hansen commented Sep 6, 2023

Optimization

More models

Ease of access

Software integration and quality

@casper-hansen casper-hansen pinned this issue Sep 6, 2023
@flexchar
Copy link

Hey Casper, first of all, amazing work!

I'm actually really curious - what's the reasoning behind supporting legacy models such as GPT-2 or GPT-J/OPT that are already in?

In my perception, the latest developments mostly on MPT/Llama 2 are by orders of magnitude better than the legacy models.

@casper-hansen
Copy link
Owner Author

Hey Casper, first of all, amazing work!

I'm actually really curious - what's the reasoning behind supporting legacy models such as GPT-2 or GPT-J/OPT that are already in?

In my perception, the latest developments mostly on MPT/Llama 2 are by orders of magnitude better than the legacy models.

Supporting older models is on the roadmap because people still use those models and ask for them. However, I do try to focus my efforts on optimizing the newer models.

@heiqilin1985
Copy link

yi-34b 能支持吗?看数据这个模型很牛叉啊。

@casper-hansen
Copy link
Owner Author

yi-34b 能支持吗?看数据这个模型很牛叉啊。

Yi is now supported on the main branch

@SinanAkkoyun
Copy link

Can you please implement Phi 1.5 support? Thank you for all the amazing work!

@xTayEx
Copy link

xTayEx commented Dec 3, 2023

Hi Casper, thank you for your wonderful work! I wonder if there is some tutorial for adding support for new model? I have noticed that Baichuan is on the roadmap. I would like try to add support for this model, could you please give me some pointer on how to support new model?

@casper-hansen
Copy link
Owner Author

@xTayEx I do not have a written guide, but here are the steps:

  1. Create a model class BaichuanAWQForCausalLM
  2. Add the model to the model map https://github.com/casper-hansen/AutoAWQ/blob/main/awq/models/auto.py#L6
  3. Import the model here https://github.com/casper-hansen/AutoAWQ/blob/main/awq/models/__init__.py

For creating the model class, look into the llama class or other classes to see how they are defined.

@casper-hansen
Copy link
Owner Author

Can you please implement Phi 1.5 support? Thank you for all the amazing work!

Phi 1.5 support has been attempted, but they have a very unusual model definition. Until it's been standardized, I am not sure I will support it.

@SinanAkkoyun
Copy link

Phi 1.5 support has been attempted, but they have a very unusual model definition. Until it's been standardized, I am not sure I will support it.

Oh :( Do you mean until a new phi model comes out?
Phi 1.5 is such an amazing model for so many applications

What would roughly be the steps to implement it on our own?

@christian-ci
Copy link

Hi @casper-hansen First of all thank you for the Amazing work. From my understanding there is an AWQ TheBloke Mixtral 8x7b Base Instruct version. I tried to run inference on it and ran into issues. Would this model be supported? Also is there a way to contribute with a donation?

@casper-hansen casper-hansen unpinned this issue Mar 1, 2024
@casper-hansen
Copy link
Owner Author

We achieved most items on the roadmap, so closing this for now to focus on other things.

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants