-
Notifications
You must be signed in to change notification settings - Fork 254
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
How can I merge the LoRA weights into the base model? #74
Comments
You can do something like this
|
How to perform inference with a LoRA model using Python code, if save_adapters = True? |
When I run this, I got 'ImportError: cannot import name 'Transformer' from 'mistral_inference.model'', the version of mistral_inference=1.2.0, how can I fix this problem? Thx! |
Try with |
A single GPU doesn't seem to be able to load the entire Mixtral-8x7B-v0.1-Instruct model, how should I merge the model using multiple cards? Thanks! |
Apparently, the only merging method available today relies on loading everything on the same device, which forces us to rent out a 40GB GPU instance like the p4d.24xlarge for the 7B model. Someone (please) correct me if I'm wrong. |
mistral-finetune has a requirement of torch==2.2, whereas mistral-inference has a requirement of torch==2.3.0 for all but the first release. |
Is there a script for this?
The text was updated successfully, but these errors were encountered: