Datasets:
Tasks:
Token Classification
Sub-tasks:
named-entity-recognition
Languages:
English
Size:
1K<n<10K
License:
Dataset Viewer issue
#4
by
Kriyans
- opened
The dataset viewer is not working.
Error details:
Error code: ConfigNamesError
Exception: DatasetWithScriptNotSupportedError
Message: The dataset viewer doesn't support this dataset because it runs arbitrary python code. Please open a discussion in the discussion tab if you think this is an error and tag
@lhoestq
and
@severo
.
Traceback: Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 65, in compute_config_names_response
for config in sorted(get_dataset_config_names(path=dataset, token=hf_token))
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 351, in get_dataset_config_names
dataset_module = dataset_module_factory(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1512, in dataset_module_factory
raise e1 from None
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1481, in dataset_module_factory
return HubDatasetModuleFactoryWithScript(
File "/src/services/worker/src/worker/utils.py", line 400, in raise_unsupported_dataset_with_script_or_init
raise DatasetWithScriptNotSupportedError(
libcommon.exceptions.DatasetWithScriptNotSupportedError: The dataset viewer doesn't support this dataset because it runs arbitrary python code. Please open a discussion in the discussion tab if you think this is an error and tag
@lhoestq
and
@severo
.
Hi. The reason is that we don't support the dataset scripts in the dataset viewer anymore. See https://discuss.huggingface.co/t/dataset-repo-requires-arbitrary-python-code-execution/59346/5
We had to disable the viewer for datasets with a script for now, because some people were abusing it. Sorry for the inconvenience.
In the meantime, if you want the dataset viewer to work you need to remove the dataset script and use a supported data format (csv, parquet, etc.) Personnally I’d recommend uploading the dataset using the datasets library and push_to_hub().