Text Generation
Transformers
Safetensors
English
stablelm_epoch
causal-lm
code
custom_code
Eval Results
4-bit precision
7 papers

issue loading

#2
by baelof - opened

10:54:10-928608 INFO Loading "TheBloke_stable-code-3b-GPTQ"
!! Warning, unknown architecture: StableLMEpochForCausalLM
!! Loading as LlamaForCausalLM
10:54:10-931120 ERROR Failed to load the model.
Traceback (most recent call last):
File "D:\ai\installs\text-generation-webui\modules\ui_model_menu.py", line 245, in load_model_wrapper
shared.model, shared.tokenizer = load_model(selected_model, loader)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ai\installs\text-generation-webui\modules\models.py", line 87, in load_model
output = load_func_maploader
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ai\installs\text-generation-webui\modules\models.py", line 357, in ExLlamav2_loader
model, tokenizer = Exllamav2Model.from_pretrained(model_name)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ai\installs\text-generation-webui\modules\exllamav2.py", line 45, in from_pretrained
config.prepare()
File "D:\ai\installs\text-generation-webui\installer_files\env\Lib\site-packages\exllamav2\config.py", line 164, in prepare
self.norm_eps = read(read_config, float, self.arch.norm_eps_key)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ai\installs\text-generation-webui\installer_files\env\Lib\site-packages\exllamav2\config.py", line 41, in read
raise ValueError(f"Missing any of the following keys: {keys}")
ValueError: Missing any of the following keys: ['rms_norm_eps']

@baelof i believe only transformers(maybe auto-gptq?) supports this model currently. Use llama.cpp(still slower then exllamav2) and the gguf variant if you want faster inference then transformers

Sign up or log in to comment