Not able to load the model using transformers

#1
by Prajwal231 - opened

Could not load model ChocoWu/nextgpt_7b_tiva_v0 with any of the following classes: (<class 'transformers.models.llama.modeling_llama.LlamaForCausalLM'>,). See the original errors: while loading with LlamaForCausalLM, an error is thrown: Traceback (most recent call last): File "/src/transformers/src/transformers/pipelines/base.py", line 269, in infer_framework_load_model model = model_class.from_pretrained(model, **kwargs) File "/src/transformers/src/transformers/modeling_utils.py", line 3063, in from_pretrained raise EnvironmentError( OSError: ChocoWu/nextgpt_7b_tiva_v0 does not appear to have a file named pytorch_model.bin, tf_model.h5, model.ckpt or flax_model.msgpack.

same here.

Hi, @minar09 , @Prajwal231 , thx for interests.
We haven't integrated the model into the transformers framework yet, so you can't load the model directly by its name;
One way is to download it offline. For specific instructions, please refer to: https://github.com/NExT-GPT/NExT-GPT
In the future, we will consider enabling automatic loading of the model using transformers.

@ChocoWu Hi, I have tried navigating your model via the instructions you have provided (https://github.com/NExT-GPT/NExT-GPT) but with no success. After the setup is completed, I tried running a simple text-to-text inference but with no success. The model outputs repetitive word tokens that do not make sense (image below).
The only way to make this error go away is if the LoRA weights are ignored all together. However, then this of course cannot produce any other modality.

Screenshot 2024-04-29 at 15.03.07.png

Sign up or log in to comment