Text Generation
Transformers
PyTorch
mosaic_gpt
custom_code

SyntaxError: (self.logit_scale=) When trying to load this model

#2
by CreativeUsername - opened

When loading this model using: AutoModelForCausalLM.from_pretrained("anas-awadalla/mpt-1b-redpajama-200b-dolly") I keep getting the following Error:
Traceback (most recent call last):
File "of_test.py", line 4, in
AutoModelForCausalLM.from_pretrained("mosaicml/mpt-1b-redpajama-200b-dolly")
File "/home/myName/miniconda3/envs/omnigibson/lib/python3.7/site-packages/transformers/models/auto/auto_factory.py", line 461, in from_pretrained
**kwargs,
File "/home/myName/miniconda3/envs/omnigibson/lib/python3.7/site-packages/transformers/models/auto/configuration_auto.py", line 953, in from_pretrained
config_class = get_class_from_dynamic_module(class_ref, pretrained_model_name_or_path, **kwargs)
File "/home/myName/miniconda3/envs/omnigibson/lib/python3.7/site-packages/transformers/dynamic_module_utils.py", line 443, in get_class_from_dynamic_module
return get_class_in_module(class_name, final_module.replace(".py", ""))
File "/home/myName/miniconda3/envs/omnigibson/lib/python3.7/site-packages/transformers/dynamic_module_utils.py", line 164, in get_class_in_module
module = importlib.import_module(module_path)
File "/home/myName/miniconda3/envs/omnigibson/lib/python3.7/importlib/init.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "", line 1006, in _gcd_import
File "", line 983, in _find_and_load
File "", line 967, in _find_and_load_unlocked
File "", line 677, in _load_unlocked
File "", line 724, in exec_module
File "", line 860, in get_code
File "", line 791, in source_to_code
File "", line 219, in _call_with_frames_removed
File "", line 1
(self.logit_scale=)
^
SyntaxError: invalid syntax

Im running transformers-3.40.2 on python 3.7 and an Ubuntu 22.04.2 machine with a NVIDIA GeForce RTX 2080 SUPER
Has anyone else experienced this? How do I fix this?

Can you try adding the trust code parameter? Like this:

model = AutoModelForCausalLM.from_pretrained("anas-awadalla/mpt-1b-redpajama-200b-dolly", trust_remote_code=True)

That didn't work. I still got the same error

Sign up or log in to comment