-
-
Notifications
You must be signed in to change notification settings - Fork 5.8k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
[Bug]: AttributeError: 'Qwen2Model' object has no attribute 'rotary_emb' #10773
Labels
bug
Something isn't working
Comments
im getting same thing but i think this is a bug AutoAwq |
bro,How did you solve this problem?
…------------------ 原始邮件 ------------------
发件人: ***@***.***>;
发送时间: 2024年11月30日(星期六) 中午11:18
收件人: ***@***.***>;
抄送: ***@***.***>; ***@***.***>;
主题: Re: [vllm-project/vllm] [Bug]: AttributeError: 'Qwen2Model' object has no attribute 'rotary_emb' (Issue #10773)
im getting same thing but i think this is a bug AutoAwq
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
+1 |
This is a bug in the weight saving method provided by Huggingface. This is being tracked in Transformers below and will be fixed in the future - until then, recommend to use v0.2.6 of AutoAWQ or patch your model like this: casper-hansen/AutoAWQ#665 (comment) |
This issue can now be closed as it has been resolved upstream in |
# for free
to join this conversation on GitHub.
Already have an account?
# to comment
Your current environment
The output of `python collect_env.py`
Model Input Dumps
from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer
model_path = '/home/root123/workspace/model/qwen2-0-5/'
quant_path = '/home/root123/workspace/model/qwen2-0-5-awq-4/'
quant_config = { "zero_point": True, "q_group_size": 128, "w_bit": 4, "version": "GEMM" }
Load model
model = AutoAWQForCausalLM.from_pretrained(
model_path, **{"low_cpu_mem_usage": True, "use_cache": False}
)
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
Quantize
model.quantize(tokenizer, quant_config=quant_config)
Save quantized model
model.save_quantized(quant_path)
tokenizer.save_pretrained(quant_path)
print(f'Model is quantized at "{quant_path}"')
🐛 Describe the bug
在进行awq模型量化时报错
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: