Error while trying to use the inference API with this model.

#2
by SnoozingSimian - opened

I tried using the inference API using this model and got the following error:

Can't load tokenizer using from_pretrained, please update its configuration: Can't load tokenizer for 'hivemind/gpt-j-6B-8bit'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'hivemind/gpt-j-6B-8bit' is the correct path to a directory containing all relevant files for a GPT2TokenizerFast tokenizer.

This error occurs even when using the inference box in the model card page.

Something is wrong here, could you please have a look?

I get the same error trying to run locally on my machine.

I had the same problem. There's a example notebook in model card (https://colab.research.google.com/drive/1ft6wQU0BhqG5PRlwgaZJv2VukKKjU4Es) that shows usage of this model. You have to use tokenizer from EleutherAI/gpt-j-6B and custom GPTJForCausalLM class that pass params.

For the record:

  • this model did not support hosted API, it could only run via the colab notebook (see model card)
  • currently, there is a better way to run gpt-J in 8 bit (also see model card)
justheuristic changed discussion status to closed

Sign up or log in to comment