Can't load in TextGen w/llamacpp_HF

#3
by biship - opened

Error: Could not load the model because a tokenizer in Transformers format was not found

However, i can load it it with llama.cpp.
Do you have the a tokenizer_config.json ?

Only a LoRA adapter for Mistral-Instruct-7B-v0.2 and a GGUF quantization for llama.cpp are provided here, not the full FP16 weights; so tokenizer_config.json is not needed. The tokenizer is the same as Mistral-7B.

Ah ok. Is there a difference between running it the two different ways?

There shouldn't be; the provided GGUF quantization should be almost lossless.

lemonilia changed discussion status to closed

Sign up or log in to comment