Issue merging peft model...

#1
by typeof - opened

Seems I'm getting the same error with this model too... adapted gist from here from @TheBloke

@teknium seems to have figured it out? perhaps this issue is some bug in config of qlora?

RuntimeError: Error(s) in loading state_dict for MixFormerSequentialForCausalLM: size mismatch for layers.25.linear.lora_B.default.weight: copying a param with shape torch.Size([50304, 64]) from checkpoint, the shape in current model is torch.Size([51200, 64]).
note: only occurs on the final layer!?....

typeof changed discussion status to closed

Sign up or log in to comment