Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

rwkv-7 代码和模型不一致 #285

Open
qxde01 opened this issue Jan 26, 2025 · 1 comment
Open

rwkv-7 代码和模型不一致 #285

qxde01 opened this issue Jan 26, 2025 · 1 comment

Comments

@qxde01
Copy link

qxde01 commented Jan 26, 2025

rwkv_v7_demo.py : args.vocab_size = 50304
01.b 实际:65536

raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format( RuntimeError: Error(s) in loading state_dict for RWKV: Missing key(s) in state_dict: "blocks.0.att.v0", "blocks.0.att.v1", "blocks.0.att.v2". size mismatch for emb.weight: copying a param with shape torch.Size([65536, 768]) from checkpoint, the shape in current model is torch.Size([50304, 768]). size mismatch for head.weight: copying a param with shape torch.Size([65536, 768]) from checkpoint, the shape in current model is torch.Size([50304, 768]).

@BlinkDL
Copy link
Owner

BlinkDL commented Jan 26, 2025

pile系列模型,和world系列模型,使用的tokenizer不同,所以vocab_size不同

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants