configuration_RW.py missing in latest commit

#60
by ravikiran3690 - opened

Could not locate the configuration_RW.py inside tiiuae/falcon-7b.

HTTPError Traceback (most recent call last)
/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py in hf_raise_for_status(response, endpoint_name)
260 try:
--> 261 response.raise_for_status()
262 except HTTPError as e:

12 frames
HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/tiiuae/falcon-7b/resolve/main/configuration_RW.py

The above exception was the direct cause of the following exception:

EntryNotFoundError Traceback (most recent call last)
EntryNotFoundError: 404 Client Error. (Request ID: Root=1-64af215f-17743c4c17dd4b075a71a76c;cc383af2-79fb-445b-b791-6819c7f1439f)

Entry Not Found for url: https://huggingface.co/tiiuae/falcon-7b/resolve/main/configuration_RW.py.

During handling of the above exception, another exception occurred:

OSError Traceback (most recent call last)
/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py in cached_file(path_or_repo_id, filename, cache_dir, force_download, resume_download, proxies, use_auth_token, revision, local_files_only, subfolder, repo_type, user_agent, _raise_exceptions_for_missing_entries, _raise_exceptions_for_connection_errors, _commit_hash)
461 if revision is None:
462 revision = "main"
--> 463 raise EnvironmentError(
464 f"{path_or_repo_id} does not appear to have a file named {full_filename}. Checkout "
465 f"'https://huggingface.co/{path_or_repo_id}/{revision}' for available files."

OSError: tiiuae/falcon-7b does not appear to have a file named configuration_RW.py. Checkout 'https://huggingface.co/tiiuae/falcon-7b/main' for available files.

ravikiran3690 changed discussion title from configuration_RW.py in latest commit to configuration_RW.py missing in latest commit

I'm getting the same error .

this is causing all the models which are based off this (ex: ybelkada/falcon-7b-sharded-bf16) to stop loading.

Same issue faced by me. This commit is very disappointing.I think we can think of 2 below workarounds

A. Can we use the missing file from local storage where the script is executing from?

B. Also if can load a previous revision of a model commit in the above code. Then I think we can bypass this.

I'm getting the same error .

this is causing all the models which are based off this (ex: ybelkada/falcon-7b-sharded-bf16) to stop loading.

My code also referring this repo. Thanks for pointing it out

Using FalconForCausalLM seems to work as the other configuration file was removed.

@ditchtech Thank you, using FalconForCausalLM worked for me as well

Yes @ditchtech worked for me too. Thank you :)

Yes @ditchtech worked for me too. Thank you :)

may I know how did you use FalconForCausalLM? I can't seem to find how to import it

Yes @ditchtech worked for me too. Thank you :)

may I know how did you use FalconForCausalLM? I can't seem to find how to import it

import FalconForCausalLM

@ditchtech thanks a million :) !

Using FalconForCausalLM seems to work as the other configuration file was removed.

from transformers import FalconForCausalLM
model_name_or_path = "huggingface/falcon-40b-gptq"

tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)

model = FalconForCausalLM.from_pretrained(model_name_or_path,
device_map="auto",
trust_remote_code=True,
revision="main")

It works. Thanks @ditchtech

Sign up or log in to comment