Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

After the mm_vision_tower and mm_mlp_adapter of llava-onevision-qwen2-7b-ov were fine-tuned, the model parameter shape did not match after the fine-tuning #342

Open
zyf8818 opened this issue Nov 18, 2024 · 2 comments

Comments

@zyf8818
Copy link

zyf8818 commented Nov 18, 2024

After the mm_vision_tower and mm_mlp_adapter of llava-onevision-qwen2-7b-ov were fine-tuned, the model parameter shape did not match after the fine-tuning.
微信图片_20241118104341

This is my training setting:after fine-tuning, when I load the fine-tuned model there is an error:

微信图片_20241118104707

@zyf8818
Copy link
Author

zyf8818 commented Nov 18, 2024

I used zero2 and zero3 respectively, but the fine-tuned model would report errors during the loading process. Is this caused by improper parameter setting?

@zyf8818
Copy link
Author

zyf8818 commented Nov 18, 2024

another information for this error:
Uploading 微信图片_20241118110501.png…

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant