OSError: no file named pytorch_model.bin

#3
by ymoslem - opened

Hello!

I tried this code:

from transformers import AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("amazon/FalconLite", trust_remote_code=True)

It results in this error:

OSError: amazon/FalconLite does not appear to have a file named pytorch_model.bin, tf_model.h5, model.ckpt or flax_model.msgpack.

I am not using SageMaker. Just a regular GPU.

Thanks!

Amazon Web Services org

You'll need to add the model_basename as well -> gptq_model-4bit-128g.safetensors
However, it looks like generation_config.json is missing, so it won't work locally
Error:
OSError: amazon/FalconLite does not appear to have a file named generation_config.json. Checkout 'https://huggingface.co/amazon/FalconLite/main' for available files.

I guess they only made change in TGI codebase, so it won't run locally with scaledRoPE because there is no code change from the original Falcon repo. Please correct me if I am wrong and eager to learn.

Does it means there is no way to finetune it (again)?

Sign up or log in to comment