-
Notifications
You must be signed in to change notification settings - Fork 500
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
KeyError model[0] did not exist in tensor? #446
Comments
Confirming exact same error ; mergekit can not find the "base_model" ; including if the path is local (absolute) on windows. Funny thing is some mergekits work fine - no issue, where as others fail for the reasons below. Example: L3 models -> merge fine, no issue This works fine: models:
BUT THIS DIES: models:
But exact SAME as above (3 models, base, dare_ties) , for Llama 3/3.1 merge - works fine (??) Other GEMMA merges of the same type (3 models, base, dare_ties) that DID work (sept 2024) now crash and burn. Even if I change this: Still dies, no matter what. Likewise any "Gemma" merges like the one above that DID WORK fine, now crash and burn. Please advise. Traceback (most recent call last): |
@FrozzDay @David-AU-github If you're able, could you try this merge on a commit from before #429 (if it's Llama) or #406 (if it's Gemma)? I'm working on more robust handling for cases like this but it'd be great to get confirmation that the issue you're experiencing is what I have in mind. Thanks! |
@cg123 Thank you so much.; |
I am performing a Mega Merge using LLaMA 3.2 3B, both the base model and fine-tuning/instruction tuning, with the DARE linear method. Following the successful completion of the initial merge, I encountered an error when attempting to merge the second one. The error message:
The config is something like this
The text was updated successfully, but these errors were encountered: