Tokenizer files missing
#2
by
munish0838
- opened
The tokenizer files are missing in model repo
same issue here:
Traceback (most recent call last):
File "", line 1, in
File "/home/Ubuntu/miniconda3/envs/ipt/lib/python3.11/site-packages/transformers/tokenization_utils_base.py", line 2094, in from_pretrained
raise EnvironmentError(
OSError: Can't load tokenizer for 'instruction-pretrain/medicine-Llama3-8B'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'instruction-pretrain/medicine-Llama3-8B' is the correct path to a directory containing all relevant files for a LlamaTokenizerFast tokenizer.
Thanks, we've uploaded tokenizer files.
Thanks so much for this! π