Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RuntimeError: Error(s) in loading state_dict for MPLUGOwl2LlamaForCausalLM: size mismatch for norm.weight: copying a param with shape torch.Size([768]) from checkpoint, the shape in current model is torch.Size([4096]). You may consider adding ignore_mismatched_sizes=True in the model from_pretrained method. #33

Open
style09131111 opened this issue Sep 6, 2024 · 1 comment

Comments

@style09131111
Copy link

RuntimeError: Error(s) in loading state_dict for MPLUGOwl2LlamaForCausalLM: size mismatch for norm.weight: copying a param with shape torch.Size([768]) from checkpoint, the shape in current model is torch.Size([4096]).
How can i solve this question,thank you!

@CarlCloudWang
Copy link

你试试直接把模型拆开来自己写个__init__加载模型,从scorer.py看,实际上这模型就三个部分:
tokenizer = AutoTokenizer.from_pretrained(model_path, model_path=model_path, use_fast=False,
cache_dir=cache_dir)
model = MPLUGOwl2LlamaForCausalLM.from_pretrained(model_path, model_path=model_path, local_files_only=True,
cache_dir=cache_dir, low_cpu_mem_usage=True,
device_map="auto")
image_processor = CLIPImageProcessor.from_pretrained(model_path)
在Modeling_llama.py的结尾,作者很巧妙地直接把llama2的函数动态加载了,所以随便加载权重。

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants