Error message while downloading the model

#1
by atakanince - opened

I am trying to download the model on Google Colab with the following command:

"

Load model directly

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("osunlp/TableLlama")
model = AutoModelForCausalLM.from_pretrained("osunlp/TableLlama")
"

However, I get the following error message:

"ValueError: Couldn't instantiate the backend tokenizer from one of:
(1) a tokenizers library serialization file,
(2) a slow tokenizer instance to convert or
(3) an equivalent slow tokenizer class to instantiate and convert.
You need to have sentencepiece installed to convert a slow tokenizer to a fast one."

transformers version: 4.35.2

OSU NLP Group org

I tried "transformers version: 4.35.2" this version. You can try to install sentencepiece package using "!pip install sentencepiece" on Colab.

I can download it by the following codes:

from transformers import AutoModel, AutoTokenizer

def download_model(model_name):
    # Replace "model_name" with the specific model you want to download
    model = AutoModel.from_pretrained(model_name)
    tokenizer = AutoTokenizer.from_pretrained(model_name)
    # Save the model and tokenizer to a directory
    model.save_pretrained('./model_directory/')
    tokenizer.save_pretrained('./model_directory/')

if __name__ == "__main__":
    model_name = "osunlp/TableLlama" 
    download_model(model_name)

Sign up or log in to comment