Is vision_model config now correct?

#11
by AisingioroHao0 - opened
image_encoder=transformers.CLIPVisionModelWithProjection.from_pretrained("laion/CLIP-ViT-bigG-14-laion2B-39B-b160k")
RuntimeError: Error(s) in loading state_dict for CLIPVisionModelWithProjection:
    size mismatch for visual_projection.weight: copying a param with shape torch.Size([1280, 1664]) from checkpoint, the shape in current model is torch.Size([512, 1664]).
    You may consider adding `ignore_mismatched_sizes=True` in the model `from_pretrained` method.
LAION eV org

fixed

rwightman changed discussion status to closed

Sign up or log in to comment