兄弟能运行起来吗?我做了个8卡int4的,结果报错
File "/home/ai/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1671, in load_state_dict
raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(
RuntimeError: Error(s) in loading state_dict for GLM130B:
size mismatch for transformer.layers.0.attention.query_key_value.weight: copying a param with shape torch.Size([4608 , 6144]) from checkpoint, the shape in current model is torch.Size([4608, 12288]).
size mismatch for transformer.layers.0.attention.dense.weight: copying a param with shape torch.Size([12288, 768]) f rom checkpoint, the shape in current model is torch.Size([12288, 1536]).
size mismatch for transformer.layers.0.mlp.dense_4h_to_h.weight: copying a param with shape torch.Size([12288, 2048] ) from checkpoint, the shape in current model is torch.Size([12288, 4096]).
size mismatch for transformer.layers.0.mlp.dense_h_to_4h.weight: copying a param with shape torch.Size([8192, 6144]) from checkpoint, the shape in current model is torch.Size([8192, 12288]).
size mismatch for transformer.layers.1.attention.query_key_value.weight: copying a param with shape torch.Size([4608 , 6144]) from checkpoint, the shape in current model is torch.Size([4608, 12288]).