Not able to Convert llama3 based model

#67
by indirajith-jithu - opened

Error: Error converting to fp16: b'INFO:hf-to-gguf:Loading model: llama-3-8b-tenjin\nINFO:gguf.gguf_writer:gguf: This GGUF file is for Little Endian only\nINFO:hf-to-gguf:Set model parameters\nINFO:hf-to-gguf:gguf: context length = 8192\nINFO:hf-to-gguf:gguf: embedding length = 4096\nINFO:hf-to-gguf:gguf: feed forward length = 14336\nINFO:hf-to-gguf:gguf: head count = 32\nINFO:hf-to-gguf:gguf: key-value head count = 8\nINFO:hf-to-gguf:gguf: rope theta = 500000.0\nINFO:hf-to-gguf:gguf: rms norm epsilon = 1e-05\nINFO:hf-to-gguf:gguf: file type = 1\nINFO:hf-to-gguf:Set model tokenizer\nSpecial tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\nWARNING:hf-to-gguf:\n\nWARNING:hf-to-gguf:**********************************************************************\nWARNING:hf-to-gguf: WARNING: The BPE pre-tokenizer was not recognized!\nWARNING:hf-to-gguf: There are 2 possible reasons for this:\nWARNING:hf-to-gguf: - the model has not been added to convert-hf-to-gguf-update.py yet\nWARNING:hf-to-gguf: - the pre-tokenization config has changed upstream\nWARNING:hf-to-gguf: Check your model files and convert-hf-to-gguf-update.py and update them accordingly.\nWARNING:hf-to-gguf: ref: https://github.com/ggerganov/llama.cpp/pull/6920\nWARNING:hf-to-gguf:**\nWARNING:hf-to-gguf:** chkhsh: c136ed14d01c2745d4f60a9596ae66800e2b61fa45643e72436041855ad4089d\nWARNING:hf-to-gguf:**********************************************************************************\nWARNING:hf-to-gguf:\n\nTraceback (most recent call last):\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 1248, in set_vocab\n self. _set_vocab_sentencepiece()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 572, in _set_vocab_sentencepiece\n raise FileNotFoundError(f"File not found: {tokenizer_path}")\nFileNotFoundError: File not found: llama-3-8b-tenjin/tokenizer.model\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 1251, in set_vocab\n self._set_vocab_llama_hf()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 634, in _set_vocab_llama_hf\n vocab = LlamaHfVocab(self.dir_model)\n File "/home/user/app/llama.cpp/convert.py", line 577, in init\n raise TypeError('Llama 3 must be converted with BpeVocab')\nTypeError: Llama 3 must be converted with BpeVocab\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2548, in \n main()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2533, in main\n model_instance.set_vocab()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 1254, in set_vocab\n self._set_vocab_gpt2()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 499, in _set_vocab_gpt2\n tokens, toktypes, tokpre = self.get_vocab_base()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 381, in get_vocab_base\n tokpre = self.get_vocab_base_pre(tokenizer)\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 490, in get_vocab_base_pre\n raise NotImplementedError("BPE pre-tokenizer was not recognized - update get_vocab_base_pre()")\nNotImplementedError: BPE pre-tokenizer was not recognized - update get_vocab_base_pre()\n'

ggml.ai org

Hi @indirajith-jithu - This is fixed in the latest version of the space - please check and feel free to open a new issue if it doesn't work. ๐Ÿค—

reach-vb changed discussion status to closed

It's working

Sign up or log in to comment