Error converting Video-LLaVA-7B

#71
by nahidalam - opened

I tried to use https://huggingface.co/spaces/ggml-org/gguf-my-repo for creating GGUF for this repo https://huggingface.co/LanguageBind/Video-LLaVA-7B-hf/tree/main but got this error

Error: Error converting to fp16: b'INFO:hf-to-gguf:Loading model: Video-LLaVA-7B-hf\nTraceback (most recent call last):\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2774, in \n main()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2752, in main\n model_class = Model.from_model_architecture(hparams["architectures"][0])\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 370, in from_model_architecture\n raise NotImplementedError(f'Architecture {arch!r} not supported!') from None\nNotImplementedError: Architecture 'VideoLlavaForConditionalGeneration' not supported!\n'

Is this likely due to https://huggingface.co/spaces/ggml-org/gguf-my-repo having older version of transformer?

ggml.ai org

Hi @nahidalam - I don't think this is a transformers issue - for context - we rebuild from source every couple of hours.

The model is not supported by the quantisation script (as you can see in the error stack). I'm not sure if it is even supported by llama.cpp :/

reach-vb changed discussion status to closed

Sign up or log in to comment