-
Notifications
You must be signed in to change notification settings - Fork 458
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
KeyError model[0] did not exist in tensor? #446
Comments
Confirming exact same error ; mergekit can not find the "base_model" ; including if the path is local (absolute) on windows. Funny thing is some mergekits work fine - no issue, where as others fail for the reasons below. Example: L3 models -> merge fine, no issue This works fine: models:
BUT THIS DIES: models:
But exact SAME as above (3 models, base, dare_ties) , for Llama 3/3.1 merge - works fine (??) Other GEMMA merges of the same type (3 models, base, dare_ties) that DID work (sept 2024) now crash and burn. Even if I change this: Still dies, no matter what. Likewise any "Gemma" merges like the one above that DID WORK fine, now crash and burn. Please advise. Traceback (most recent call last): |
@FrozzDay @David-AU-github If you're able, could you try this merge on a commit from before #429 (if it's Llama) or #406 (if it's Gemma)? I'm working on more robust handling for cases like this but it'd be great to get confirmation that the issue you're experiencing is what I have in mind. Thanks! |
@cg123 Thank you so much.; |
I am performing a Mega Merge using LLaMA 3.2 3B, both the base model and fine-tuning/instruction tuning, with the DARE linear method. Following the successful completion of the initial merge, I encountered an error when attempting to merge the second one. The error message:
The config is something like this
The text was updated successfully, but these errors were encountered: