id
int64
599M
3.29B
url
stringlengths
58
61
html_url
stringlengths
46
51
number
int64
1
7.72k
title
stringlengths
1
290
state
stringclasses
2 values
comments
int64
0
70
created_at
timestamp[s]date
2020-04-14 10:18:02
2025-08-05 09:28:51
updated_at
timestamp[s]date
2020-04-27 16:04:17
2025-08-05 11:39:56
closed_at
timestamp[s]date
2020-04-14 12:01:40
2025-08-01 05:15:45
user_login
stringlengths
3
26
labels
listlengths
0
4
body
stringlengths
0
228k
is_pull_request
bool
2 classes
851,229,399
https://api.github.com/repos/huggingface/datasets/issues/2172
https://github.com/huggingface/datasets/pull/2172
2,172
Pin fsspec lower than 0.9.0
closed
0
2021-04-06T09:19:09
2021-04-06T09:49:27
2021-04-06T09:49:26
lhoestq
[]
Today's release of `fsspec` 0.9.0 implied a new release of `s3fs` 0.6.0 but this version breaks the CI (see [here](https://app.circleci.com/pipelines/github/huggingface/datasets/5312/workflows/490f3240-cd1c-4dd1-bb60-b416771c5584/jobs/32734) for example) I'm pinning `fsspec` until this has been resolved
true
851,090,662
https://api.github.com/repos/huggingface/datasets/issues/2171
https://github.com/huggingface/datasets/pull/2171
2,171
Fixed the link to wikiauto training data.
closed
3
2021-04-06T07:13:11
2021-04-06T16:05:42
2021-04-06T16:05:09
mounicam
[]
true
850,913,228
https://api.github.com/repos/huggingface/datasets/issues/2170
https://github.com/huggingface/datasets/issues/2170
2,170
Wikipedia historic dumps are deleted but hf/datasets hardcodes dump date
open
1
2021-04-06T03:13:18
2021-06-16T01:10:50
null
leezu
[]
Wikimedia does not keep all historical dumps. For example, as of today https://dumps.wikimedia.org/kowiki/ only provides ``` 20201220/ 02-Feb-2021 01:36 - 20210101/ 21-Feb-2021 01:26 - 20210120/ 02-Mar-2021 01:25 - 20210201/ 21-Mar-2021 01:26 - 20210220/ 02-Apr-2021 01:26 - 20210301/ 03-Mar-2021 08:10 - 20210320/ 21-Mar-2021 18:13 - 20210401/ 03-Apr-2021 10:08 - latest/ 03-Apr-2021 10:08 - ``` However, the wikipedia dataset provided in the library, only supports the following configs, none of which are applicable anymore when disregarding the cached datasets: ``` ValueError: BuilderConfig 20210401.ko not found. Available: ['20200501.aa', '20200501.ab', '20200501.ace', '20200501.ady', '20200501.af', '20200501.ak', '20200501.als', '20200501.am', '20200501.an', '20200501.ang', '20200501.ar', '20200501.arc', '20200501.arz', '20200501.as', '20200501.ast', '20200501.atj', '20200501.av', '20200501.ay', '20200501.az', '20200501.azb', '20200501.ba', '20200501.bar', '20200501.bat-smg', '20200501.bcl', '20200501.be', '20200501.be-x-old', '20200501.bg', '20200501.bh', '20200501.bi', '20200501.bjn', '20200501.bm', '20200501.bn', '20200501.bo', '20200501.bpy', '20200501.br', '20200501.bs', '20200501.bug', '20200501.bxr', '20200501.ca', '20200501.cbk-zam', '20200501.cdo', '20200501.ce', '20200501.ceb', '20200501.ch', '20200501.cho', '20200501.chr', '20200501.chy', '20200501.ckb', '20200501.co', '20200501.cr', '20200501.crh', '20200501.cs', '20200501.csb', '20200501.cu', '20200501.cv', '20200501.cy', '20200501.da', '20200501.de', '20200501.din', '20200501.diq', '20200501.dsb', '20200501.dty', '20200501.dv', '20200501.dz', '20200501.ee', '20200501.el', '20200501.eml', '20200501.en', '20200501.eo', '20200501.es', '20200501.et', '20200501.eu', '20200501.ext', '20200501.fa', '20200501.ff', '20200501.fi', '20200501.fiu-vro', '20200501.fj', '20200501.fo', '20200501.fr', '20200501.frp', '20200501.frr', '20200501.fur', '20200501.fy', '20200501.ga', '20200501.gag', '20200501.gan', '20200501.gd', '20200501.gl', '20200501.glk', '20200501.gn', '20200501.gom', '20200501.gor', '20200501.got', '20200501.gu', '20200501.gv', '20200501.ha', '20200501.hak', '20200501.haw', '20200501.he', '20200501.hi', '20200501.hif', '20200501.ho', '20200501.hr', '20200501.hsb', '20200501.ht', '20200501.hu', '20200501.hy', '20200501.ia', '20200501.id', '20200501.ie', '20200501.ig', '20200501.ii', '20200501.ik', '20200501.ilo', '20200501.inh', '20200501.io', '20200501.is', '20200501.it', '20200501.iu', '20200501.ja', '20200501.jam', '20200501.jbo', '20200501.jv', '20200501.ka', '20200501.kaa', '20200501.kab', '20200501.kbd', '20200501.kbp', '20200501.kg', '20200501.ki', '20200501.kj', '20200501.kk', '20200501.kl', '20200501.km', '20200501.kn', '20200501.ko', '20200501.koi', '20200501.krc', '20200501.ks', '20200501.ksh', '20200501.ku', '20200501.kv', '20200501.kw', '20200501.ky', '20200501.la', '20200501.lad', '20200501.lb', '20200501.lbe', '20200501.lez', '20200501.lfn', '20200501.lg', '20200501.li', '20200501.lij', '20200501.lmo', '20200501.ln', '20200501.lo', '20200501.lrc', '20200501.lt', '20200501.ltg', '20200501.lv', '20200501.mai', '20200501.map-bms', '20200501.mdf', '20200501.mg', '20200501.mh', '20200501.mhr', '20200501.mi', '20200501.min', '20200501.mk', '20200501.ml', '20200501.mn', '20200501.mr', '20200501.mrj', '20200501.ms', '20200501.mt', '20200501.mus', '20200501.mwl', '20200501.my', '20200501.myv', '20200501.mzn', '20200501.na', '20200501.nah', '20200501.nap', '20200501.nds', '20200501.nds-nl', '20200501.ne', '20200501.new', '20200501.ng', '20200501.nl', '20200501.nn', '20200501.no', '20200501.nov', '20200501.nrm', '20200501.nso', '20200501.nv', '20200501.ny', '20200501.oc', '20200501.olo', '20200501.om', '20200501.or', '20200501.os', '20200501.pa', '20200501.pag', '20200501.pam', '20200501.pap', '20200501.pcd', '20200501.pdc', '20200501.pfl', '20200501.pi', '20200501.pih', '20200501.pl', '20200501.pms', '20200501.pnb', '20200501.pnt', '20200501.ps', '20200501.pt', '20200501.qu', '20200501.rm', '20200501.rmy', '20200501.rn', '20200501.ro', '20200501.roa-rup', '20200501.roa-tara', '20200501.ru', '20200501.rue', '20200501.rw', '20200501.sa', '20200501.sah', '20200501.sat', '20200501.sc', '20200501.scn', '20200501.sco', '20200501.sd', '20200501.se', '20200501.sg', '20200501.sh', '20200501.si', '20200501.simple', '20200501.sk', '20200501.sl', '20200501.sm', '20200501.sn', '20200501.so', '20200501.sq', '20200501.sr', '20200501.srn', '20200501.ss', '20200501.st', '20200501.stq', '20200501.su', '20200501.sv', '20200501.sw', '20200501.szl', '20200501.ta', '20200501.tcy', '20200501.te', '20200501.tet', '20200501.tg', '20200501.th', '20200501.ti', '20200501.tk', '20200501.tl', '20200501.tn', '20200501.to', '20200501.tpi', '20200501.tr', '20200501.ts', '20200501.tt', '20200501.tum', '20200501.tw', '20200501.ty', '20200501.tyv', '20200501.udm', '20200501.ug', '20200501.uk', '20200501.ur', '20200501.uz', '20200501.ve', '20200501.vec', '20200501.vep', '20200501.vi', '20200501.vls', '20200501.vo', '20200501.wa', '20200501.war', '20200501.wo', '20200501.wuu', '20200501.xal', '20200501.xh', '20200501.xmf', '20200501.yi', '20200501.yo', '20200501.za', '20200501.zea', '20200501.zh', '20200501.zh-classical', '20200501.zh-min-nan', '20200501.zh-yue', '20200501.zu'] ``` The cached datasets: ``` % aws s3 --no-sign-request --endpoint-url https://storage.googleapis.com ls s3://huggingface-nlp/cache/datasets/wikipedia/ PRE 20200501.de/ PRE 20200501.en/ PRE 20200501.fr/ PRE 20200501.frr/ PRE 20200501.it/ PRE 20200501.simple/ ```
false
850,456,180
https://api.github.com/repos/huggingface/datasets/issues/2169
https://github.com/huggingface/datasets/pull/2169
2,169
Updated WER metric implementation to avoid memory issues
closed
1
2021-04-05T15:43:20
2021-04-06T15:02:58
2021-04-06T15:02:58
diego-fustes
[]
This is in order to fix this issue: https://github.com/huggingface/datasets/issues/2078
true
849,957,941
https://api.github.com/repos/huggingface/datasets/issues/2168
https://github.com/huggingface/datasets/pull/2168
2,168
Preserve split type when realoding dataset
closed
5
2021-04-04T20:46:21
2021-04-19T10:57:05
2021-04-19T09:08:55
mariosasko
[]
Fixes #2167 Using `eval` is not ideal for security reasons (in web apps I assume), but without it the code would be much more complex IMO. In terms of style, instead of explicitly importing a private member (`_RelativeInstruction`), we can add these imports at the top of the module: ```python from . import arrow_reader # gives us access to ReadInstruction and _RelativeInstruction from . import splits # gives us access to NamedSplit ``` and then define the `eval` globals as follows: ```python {**arrow_reader.__dict__, **splits.__dict__} ```
true
849,944,891
https://api.github.com/repos/huggingface/datasets/issues/2167
https://github.com/huggingface/datasets/issues/2167
2,167
Split type not preserved when reloading the dataset
closed
0
2021-04-04T19:29:54
2021-04-19T09:08:55
2021-04-19T09:08:55
mariosasko
[]
A minimal reproducible example: ```python >>> from datasets import load_dataset, Dataset >>> dset = load_dataset("sst", split="train") >>> dset.save_to_disk("sst") >>> type(dset.split) <class 'datasets.splits.NamedSplit'> >>> dset = Dataset.load_from_disk("sst") >>> type(dset.split) # NamedSplit expected <class 'str'> ``` It seems like this bug was introduced in #2025.
false
849,778,545
https://api.github.com/repos/huggingface/datasets/issues/2166
https://github.com/huggingface/datasets/issues/2166
2,166
Regarding Test Sets for the GEM datasets
closed
2
2021-04-04T02:02:45
2021-04-06T08:13:12
2021-04-06T08:13:12
vyraun
[ "Dataset discussion" ]
@yjernite Hi, are the test sets for the GEM datasets scheduled to be [added soon](https://gem-benchmark.com/shared_task)? e.g. ``` from datasets import load_dataset DATASET_NAME="common_gen" data = load_dataset("gem", DATASET_NAME) ``` The test set doesn't have the target or references. ``` data['test'][0] {'concept_set_id': 0, 'concepts': ['drill', 'field', 'run', 'team'], 'gem_id': 'common_gen-test-0', 'gem_parent_id': 'common_gen-test-0', 'references': [], 'target': ''} ```
false
849,771,665
https://api.github.com/repos/huggingface/datasets/issues/2165
https://github.com/huggingface/datasets/issues/2165
2,165
How to convert datasets.arrow_dataset.Dataset to torch.utils.data.Dataset
closed
7
2021-04-04T01:01:48
2021-08-24T15:55:35
2021-04-07T15:06:04
y-rokutan
[]
Hi, I'm trying to pretraine deep-speed model using HF arxiv dataset like: ``` train_ds = nlp.load_dataset('scientific_papers', 'arxiv') train_ds.set_format( type="torch", columns=["input_ids", "attention_mask", "global_attention_mask", "labels"], ) engine, _, _, _ = deepspeed.initialize( args=args, model=model, model_parameters=[p for p in model.parameters() if p.requires_grad], training_data=train_ds) ``` but deepspeed.initialize accepts torch.utils.data.Dataset only. How can I convert HF-style dataset to torch-style dataset?
false
849,739,759
https://api.github.com/repos/huggingface/datasets/issues/2164
https://github.com/huggingface/datasets/pull/2164
2,164
Replace assertTrue(isinstance with assertIsInstance in tests
closed
0
2021-04-03T21:07:02
2021-04-06T14:41:09
2021-04-06T14:41:08
mariosasko
[]
Replaces all the occurrences of the `assertTrue(isinstance(` pattern with `assertIsInstance`.
true
849,669,366
https://api.github.com/repos/huggingface/datasets/issues/2163
https://github.com/huggingface/datasets/pull/2163
2,163
Concat only unique fields in DatasetInfo.from_merge
closed
3
2021-04-03T14:31:30
2021-04-06T14:40:00
2021-04-06T14:39:59
mariosasko
[]
I thought someone from the community with less experience would be interested in fixing this issue, but that wasn't the case. Fixes #2103
true
849,129,201
https://api.github.com/repos/huggingface/datasets/issues/2162
https://github.com/huggingface/datasets/issues/2162
2,162
visualization for cc100 is broken
closed
3
2021-04-02T10:11:13
2022-10-05T13:20:24
2022-10-05T13:20:24
dorost1234
[ "nlp-viewer" ]
Hi visualization through dataset viewer for cc100 is broken https://huggingface.co/datasets/viewer/ thanks a lot
false
849,127,041
https://api.github.com/repos/huggingface/datasets/issues/2161
https://github.com/huggingface/datasets/issues/2161
2,161
any possibility to download part of large datasets only?
closed
6
2021-04-02T10:06:46
2022-10-05T13:26:51
2022-10-05T13:26:51
dorost1234
[]
Hi Some of the datasets I need like cc100 are very large, and then I wonder if I can download first X samples of the shuffled/unshuffled data without going through first downloading the whole data then sampling? thanks
false
849,052,921
https://api.github.com/repos/huggingface/datasets/issues/2160
https://github.com/huggingface/datasets/issues/2160
2,160
data_args.preprocessing_num_workers almost freezes
closed
2
2021-04-02T07:56:13
2021-04-02T10:14:32
2021-04-02T10:14:31
dorost1234
[]
Hi @lhoestq I am running this code from huggingface transformers https://github.com/huggingface/transformers/blob/master/examples/language-modeling/run_mlm.py to speed up tokenization, since I am running on multiple datasets, I am using data_args.preprocessing_num_workers = 4 with opus100 corpus but this moves on till a point and then this freezes almost for sometime during tokenization steps and then this is back again, overall to me taking more time than normal case, I appreciate your advice on how I can use this option properly to speed up. thanks
false
848,851,962
https://api.github.com/repos/huggingface/datasets/issues/2159
https://github.com/huggingface/datasets/issues/2159
2,159
adding ccnet dataset
closed
1
2021-04-01T23:28:36
2021-04-02T10:05:19
2021-04-02T10:05:19
dorost1234
[ "dataset request" ]
## Adding a Dataset - **Name:** ccnet - **Description:** Common Crawl - **Paper:** https://arxiv.org/abs/1911.00359 - **Data:** https://github.com/facebookresearch/cc_net - **Motivation:** this is one of the most comprehensive clean monolingual datasets across a variety of languages. Quite important for cross-lingual reseach Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md). thanks
false
848,506,746
https://api.github.com/repos/huggingface/datasets/issues/2158
https://github.com/huggingface/datasets/issues/2158
2,158
viewer "fake_news_english" error
closed
2
2021-04-01T14:13:20
2022-10-05T13:22:02
2022-10-05T13:22:02
emanuelevivoli
[ "nlp-viewer" ]
When I visit the [Huggingface - viewer](https://huggingface.co/datasets/viewer/) web site, under the dataset "fake_news_english" I've got this error: > ImportError: To be able to use this dataset, you need to install the following dependencies['openpyxl'] using 'pip install # noqa: requires this pandas optional dependency for reading xlsx files' for instance' as well as the error Traceback.
false
847,205,239
https://api.github.com/repos/huggingface/datasets/issues/2157
https://github.com/huggingface/datasets/pull/2157
2,157
updated user permissions based on umask
closed
0
2021-03-31T19:38:29
2021-04-06T07:19:19
2021-04-06T07:19:19
bhavitvyamalik
[]
Updated user permissions based on running user's umask (#2065). Let me know if `0o666` is looking good or should I change it to `~umask` only (to give execute permissions as well)
true
847,198,295
https://api.github.com/repos/huggingface/datasets/issues/2156
https://github.com/huggingface/datasets/pull/2156
2,156
User permissions
closed
0
2021-03-31T19:33:48
2021-03-31T19:34:24
2021-03-31T19:34:24
bhavitvyamalik
[]
Updated user permissions based on running user's umask. Let me know if `0o666` is looking good or should I change it to `~umask` only (to give execute permissions as well)
true
846,786,897
https://api.github.com/repos/huggingface/datasets/issues/2155
https://github.com/huggingface/datasets/pull/2155
2,155
Add table classes to the documentation
closed
1
2021-03-31T14:36:10
2021-04-01T16:46:30
2021-03-31T15:42:08
lhoestq
[]
Following #2025 , I added the table classes to the documentation cc @albertvillanova
true
846,763,960
https://api.github.com/repos/huggingface/datasets/issues/2154
https://github.com/huggingface/datasets/pull/2154
2,154
Adding the NorNE dataset for Norwegian POS and NER
closed
1
2021-03-31T14:22:50
2021-04-01T09:27:00
2021-04-01T09:16:08
versae
[]
NorNE is a manually annotated corpus of named entities which extends the annotation of the existing Norwegian Dependency Treebank. Comprising both of the official standards of written Norwegian (Bokmål and Nynorsk), the corpus contains around 600,000 tokens and annotates a rich set of entity types including persons, organizations, locations, geo-political entities, products, and events, in addition to a class corresponding to nominals derived from names. See #1720.
true
846,181,502
https://api.github.com/repos/huggingface/datasets/issues/2153
https://github.com/huggingface/datasets/issues/2153
2,153
load_dataset ignoring features
closed
3
2021-03-31T08:30:09
2022-10-05T13:29:12
2022-10-05T13:29:12
GuillemGSubies
[ "bug" ]
First of all, I'm sorry if it is a repeated issue or the changes are already in master, I searched and I didn't find anything. I'm using datasets 1.5.0 ![image](https://user-images.githubusercontent.com/37592763/113114369-8f376580-920b-11eb-900d-94365b59f04b.png) As you can see, when I load the dataset, the ClassLabels are ignored, I have to cast the dataset in order to make it work. Code to reproduce: ```python import datasets data_location = "/data/prueba_multiclase" features = datasets.Features( {"texto": datasets.Value("string"), "label": datasets.features.ClassLabel(names=["false", "true"])} ) dataset = datasets.load_dataset( "csv", data_files=data_location, delimiter="\t", features=features ) ``` Dataset I used: [prueba_multiclase.zip](https://github.com/huggingface/datasets/files/6235022/prueba_multiclase.zip) (it has to be unzipped) Thank you! ❤️
false
845,751,273
https://api.github.com/repos/huggingface/datasets/issues/2152
https://github.com/huggingface/datasets/pull/2152
2,152
Update README.md
closed
0
2021-03-31T03:21:19
2021-04-01T10:20:37
2021-04-01T10:20:36
JieyuZhao
[]
Updated some descriptions of Wino_Bias dataset.
true
844,886,081
https://api.github.com/repos/huggingface/datasets/issues/2151
https://github.com/huggingface/datasets/pull/2151
2,151
Add support for axis in concatenate datasets
closed
5
2021-03-30T16:58:44
2021-06-23T17:41:02
2021-04-19T16:07:18
albertvillanova
[ "enhancement" ]
Add support for `axis` (0 or 1) in `concatenate_datasets`. Close #853.
true
844,776,448
https://api.github.com/repos/huggingface/datasets/issues/2150
https://github.com/huggingface/datasets/pull/2150
2,150
Allow pickling of big in-memory tables
closed
0
2021-03-30T15:51:56
2021-03-31T10:37:15
2021-03-31T10:37:14
lhoestq
[]
This should fix issue #2134 Pickling is limited to <4GiB objects, it's not possible to pickle a big arrow table (for multiprocessing for example). For big tables, we have to write them on disk and only pickle the path to the table.
true
844,734,076
https://api.github.com/repos/huggingface/datasets/issues/2149
https://github.com/huggingface/datasets/issues/2149
2,149
Telugu subset missing for xtreme tatoeba dataset
closed
2
2021-03-30T15:26:34
2022-10-05T13:28:30
2022-10-05T13:28:30
cosmeowpawlitan
[]
from nlp import load_dataset train_dataset = load_dataset('xtreme', 'tatoeba.tel')['validation'] ValueError: BuilderConfig tatoeba.tel not found. but language tel is actually included in xtreme: https://github.com/google-research/xtreme/blob/master/utils_preprocess.py def tatoeba_preprocess(args): lang3_dict = { 'afr':'af', 'ara':'ar', 'bul':'bg', 'ben':'bn', 'deu':'de', 'ell':'el', 'spa':'es', 'est':'et', 'eus':'eu', 'pes':'fa', 'fin':'fi', 'fra':'fr', 'heb':'he', 'hin':'hi', 'hun':'hu', 'ind':'id', 'ita':'it', 'jpn':'ja', 'jav':'jv', 'kat':'ka', 'kaz':'kk', 'kor':'ko', 'mal':'ml', 'mar':'mr', 'nld':'nl', 'por':'pt', 'rus':'ru', 'swh':'sw', 'tam':'ta', **_'tel':'te'_**, 'tha':'th', 'tgl':'tl', <----here 'tur':'tr', 'urd':'ur', 'vie':'vi', 'cmn':'zh', 'eng':'en', }
false
844,700,910
https://api.github.com/repos/huggingface/datasets/issues/2148
https://github.com/huggingface/datasets/issues/2148
2,148
Add configurable options to `seqeval` metric
closed
1
2021-03-30T15:04:06
2021-04-15T13:49:46
2021-04-15T13:49:46
marrodion
[]
Right now `load_metric("seqeval")` only works in the default mode of evaluation (equivalent to conll evaluation). However, seqeval library [supports](https://github.com/chakki-works/seqeval#support-features) different evaluation schemes (IOB1, IOB2, etc.), which can be plugged in just by supporting additional kwargs in `Seqeval._compute` https://github.com/huggingface/datasets/blob/85cf7ff920c90ca2e12bedca12b36d2a043c3da2/metrics/seqeval/seqeval.py#L109 Things that would be relevant are, for example, supporting `mode="strict", scheme=IOB2` to count only full entity match as a true positive and omit partial matches. The only problem I see is that the spirit of `metrics` seems to not require additional imports from user. `seqeval` only supports schemes as objects, without any string aliases. It can be solved naively with mapping like `{"IOB2": seqeval.scheme.IOB2}`. Or just left as is and require user to explicitly import scheme from `seqeval` if he wants to configure it past the default implementation. If that makes sense, I am happy to implement the change.
false
844,687,831
https://api.github.com/repos/huggingface/datasets/issues/2147
https://github.com/huggingface/datasets/pull/2147
2,147
Render docstring return type as inline
closed
0
2021-03-30T14:55:43
2021-03-31T13:11:05
2021-03-31T13:11:05
albertvillanova
[ "documentation" ]
This documentation setting will avoid having the return type in a separate line under `Return type`. See e.g. current docs for `Dataset.to_csv`.
true
844,673,244
https://api.github.com/repos/huggingface/datasets/issues/2146
https://github.com/huggingface/datasets/issues/2146
2,146
Dataset file size on disk is very large with 3D Array
open
6
2021-03-30T14:46:09
2021-04-16T13:07:02
null
jblemoine
[]
Hi, I have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. The actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. `{ "description": "", "citation": "", "homepage": "", "license": "", "features": { "image": { "shape": [224, 224, 3], "dtype": "uint8", "id": null, "_type": "Array3D", } }, "post_processed": null, "supervised_keys": null, "builder_name": "shot_type_image_dataset", "config_name": "default", "version": { "version_str": "0.0.0", "description": null, "major": 0, "minor": 0, "patch": 0, }, "splits": { "train": { "name": "train", "num_bytes": 520803408, "num_examples": 1479, "dataset_name": "shot_type_image_dataset", } }, "download_checksums": { "": { "num_bytes": 16940447118, "checksum": "5854035705efe08b0ed8f3cf3da7b4d29cba9055c2d2d702c79785350d72ee03", } }, "download_size": 16940447118, "post_processing_size": null, "dataset_size": 520803408, "size_in_bytes": 17461250526, }` I have created the same dataset with tensorflow_dataset and it takes only 125MB on disk. I am wondering, is it normal behavior ? I understand `Datasets` uses Arrow for serialization wheres tf uses TF Records. This might be a problem for large dataset. Thanks for your help.
false
844,603,518
https://api.github.com/repos/huggingface/datasets/issues/2145
https://github.com/huggingface/datasets/pull/2145
2,145
Implement Dataset add_column
closed
1
2021-03-30T14:02:14
2021-04-29T14:50:44
2021-04-29T14:50:43
albertvillanova
[ "enhancement" ]
Implement `Dataset.add_column`. Close #1954.
true
844,352,067
https://api.github.com/repos/huggingface/datasets/issues/2144
https://github.com/huggingface/datasets/issues/2144
2,144
Loading wikipedia 20200501.en throws pyarrow related error
open
6
2021-03-30T10:38:31
2021-04-01T09:21:17
null
TomPyonsuke
[]
**Problem description** I am getting the following error when trying to load wikipedia/20200501.en dataset. **Error log** Downloading and preparing dataset wikipedia/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to /usr/local/workspace/NAS_NLP/cache/wikipedia/20200501.en/1.0.0/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931... Downloading: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 14.6k/14.6k [00:00<00:00, 5.41MB/s] Downloading: 59%|███████████████████████████████████████████████████████████████████████████████████████▊ | 10.7G/18.3G [11:30<08:08, 15.5MB/s] Dataset wikipedia downloaded and prepared to /usr/local/workspace/NAS_NLP/cache/wikipedia/20200501.en/1.0.0/50aa706aa417bb77d910ad61211cc672c0ef3e0f224225a5e0a18277ade8b931. Subsequent calls will reuse this data. Traceback (most recent call last): File "load_wiki.py", line 2, in <module> ds = load_dataset('wikipedia', '20200501.en', cache_dir='/usr/local/workspace/NAS_NLP/cache') File "/usr/local/lib/python3.6/dist-packages/datasets/load.py", line 751, in load_dataset ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory) File "/usr/local/lib/python3.6/dist-packages/datasets/builder.py", line 746, in as_dataset map_tuple=True, File "/usr/local/lib/python3.6/dist-packages/datasets/utils/py_utils.py", line 204, in map_nested _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm) File "/usr/local/lib/python3.6/dist-packages/datasets/utils/py_utils.py", line 204, in <listcomp> _single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm) File "/usr/local/lib/python3.6/dist-packages/datasets/utils/py_utils.py", line 142, in _single_map_nested return function(data_struct) File "/usr/local/lib/python3.6/dist-packages/datasets/builder.py", line 763, in _build_single_dataset in_memory=in_memory, File "/usr/local/lib/python3.6/dist-packages/datasets/builder.py", line 835, in _as_dataset in_memory=in_memory, File "/usr/local/lib/python3.6/dist-packages/datasets/arrow_reader.py", line 215, in read return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory) File "/usr/local/lib/python3.6/dist-packages/datasets/arrow_reader.py", line 236, in read_files pa_table = self._read_files(files, in_memory=in_memory) File "/usr/local/lib/python3.6/dist-packages/datasets/arrow_reader.py", line 171, in _read_files pa_table: pa.Table = self._get_dataset_from_filename(f_dict, in_memory=in_memory) File "/usr/local/lib/python3.6/dist-packages/datasets/arrow_reader.py", line 302, in _get_dataset_from_filename pa_table = ArrowReader.read_table(filename, in_memory=in_memory) File "/usr/local/lib/python3.6/dist-packages/datasets/arrow_reader.py", line 324, in read_table pa_table = f.read_all() File "pyarrow/ipc.pxi", line 544, in pyarrow.lib.RecordBatchReader.read_all File "pyarrow/error.pxi", line 99, in pyarrow.lib.check_status OSError: Expected to be able to read 9176784 bytes for message body, got 4918712 **Detailed version info** datasets==1.5.0 - dataclasses [required: Any, installed: 0.8] - dill [required: Any, installed: 0.3.3] - fsspec [required: Any, installed: 0.8.7] - importlib-metadata [required: Any, installed: 1.7.0] - zipp [required: >=0.5, installed: 3.1.0] - huggingface-hub [required: <0.1.0, installed: 0.0.7] - filelock [required: Any, installed: 3.0.12] - importlib-metadata [required: Any, installed: 1.7.0] - zipp [required: >=0.5, installed: 3.1.0] - requests [required: Any, installed: 2.24.0] - certifi [required: >=2017.4.17, installed: 2020.6.20] - chardet [required: >=3.0.2,<4, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.6] - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10] - tqdm [required: Any, installed: 4.49.0] - importlib-metadata [required: Any, installed: 1.7.0] - zipp [required: >=0.5, installed: 3.1.0] - multiprocess [required: Any, installed: 0.70.11.1] - dill [required: >=0.3.3, installed: 0.3.3] - numpy [required: >=1.17, installed: 1.17.0] - pandas [required: Any, installed: 1.1.5] - numpy [required: >=1.15.4, installed: 1.17.0] - python-dateutil [required: >=2.7.3, installed: 2.8.0] - six [required: >=1.5, installed: 1.15.0] - pytz [required: >=2017.2, installed: 2020.1] - pyarrow [required: >=0.17.1, installed: 3.0.0] - numpy [required: >=1.16.6, installed: 1.17.0] - requests [required: >=2.19.0, installed: 2.24.0] - certifi [required: >=2017.4.17, installed: 2020.6.20] - chardet [required: >=3.0.2,<4, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.6] - urllib3 [required: >=1.21.1,<1.26,!=1.25.1,!=1.25.0, installed: 1.25.10] - tqdm [required: >=4.27,<4.50.0, installed: 4.49.0] - xxhash [required: Any, installed: 2.0.0]
false
844,313,228
https://api.github.com/repos/huggingface/datasets/issues/2143
https://github.com/huggingface/datasets/pull/2143
2,143
task casting via load_dataset
closed
0
2021-03-30T10:00:42
2021-06-11T13:20:41
2021-06-11T13:20:36
theo-m
[]
wip not satisfied with the API, it means as a dataset implementer I need to write a function with boilerplate and write classes for each `<dataset><task>` "facet".
true
843,919,420
https://api.github.com/repos/huggingface/datasets/issues/2142
https://github.com/huggingface/datasets/pull/2142
2,142
Gem V1.1
closed
0
2021-03-29T23:47:02
2021-03-30T00:10:02
2021-03-30T00:10:02
yjernite
[]
This branch updates the GEM benchmark to its 1.1 version which includes: - challenge sets for most tasks - detokenized TurkCorpus to match the rest of the text simplification subtasks - fixed inputs for TurkCorpus and ASSET test sets - 18 languages in WikiLingua cc @sebastianGehrmann
true
843,914,790
https://api.github.com/repos/huggingface/datasets/issues/2141
https://github.com/huggingface/datasets/pull/2141
2,141
added spans field for the wikiann datasets
closed
3
2021-03-29T23:38:26
2021-03-31T13:27:50
2021-03-31T13:27:50
rabeehk
[]
Hi @lhoestq I tried to add spans to the wikiann datasets. Thanks a lot for kindly having a look. This addresses https://github.com/huggingface/datasets/issues/2130. Best regards Rabeeh
true
843,830,451
https://api.github.com/repos/huggingface/datasets/issues/2140
https://github.com/huggingface/datasets/pull/2140
2,140
add banking77 dataset
closed
1
2021-03-29T21:32:23
2021-04-09T09:32:18
2021-04-09T09:32:18
dkajtoch
[]
Intent classification/detection dataset from banking category with 77 unique intents.
true
843,662,613
https://api.github.com/repos/huggingface/datasets/issues/2139
https://github.com/huggingface/datasets/issues/2139
2,139
TypeError when using save_to_disk in a dataset loaded with ReadInstruction split
closed
2
2021-03-29T18:23:54
2021-03-30T09:12:53
2021-03-30T09:12:53
PedroMLF
[]
Hi, Loading a dataset with `load_dataset` using a split defined via `ReadInstruction` and then saving it to disk results in the following error: `TypeError: Object of type ReadInstruction is not JSON serializable`. Here is the minimal reproducible example: ```python from datasets import load_dataset from datasets import ReadInstruction data_1 = load_dataset( "wikiann", "en", split="validation", ) data_1.save_to_disk("temporary_path_1") print("Save with regular split works.") data_2 = load_dataset( "wikiann", "en", split=ReadInstruction("validation", to=50, unit="%"), ) data_2.save_to_disk("temporary_path_2") ``` and the corresponding output: ``` Reusing dataset wikiann (/xxxxx/.cache/huggingface/datasets/wikiann/en/1.1.0/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9) Save with regular split works. Reusing dataset wikiann (/xxxxx/.cache/huggingface/datasets/wikiann/en/1.1.0/0b11a6fb31eea02f38ca17610657bfba3206100685283014daceb8da291c3be9) Traceback (most recent call last): File "bug.py", line 20, in <module> data_2.save_to_disk("temporary_path_2") File "/xxxxx/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 645, in save_to_disk json.dump(state, state_file, indent=2, sort_keys=True) File "/usr/lib/python3.7/json/__init__.py", line 179, in dump for chunk in iterable: File "/usr/lib/python3.7/json/encoder.py", line 431, in _iterencode yield from _iterencode_dict(o, _current_indent_level) File "/usr/lib/python3.7/json/encoder.py", line 405, in _iterencode_dict yield from chunks File "/usr/lib/python3.7/json/encoder.py", line 438, in _iterencode o = _default(o) File "/usr/lib/python3.7/json/encoder.py", line 179, in default raise TypeError(f'Object of type {o.__class__.__name__} ' TypeError: Object of type ReadInstruction is not JSON serializable ``` Let me know if there is some misuse from my end. Thanks in advance.
false
843,508,402
https://api.github.com/repos/huggingface/datasets/issues/2138
https://github.com/huggingface/datasets/pull/2138
2,138
Add CER metric
closed
0
2021-03-29T15:52:27
2021-04-06T16:16:11
2021-04-06T07:14:38
chutaklee
[]
Add Character Error Rate (CER) metric that is used in evaluation in ASR. I also have written unittests (hopefully thorough enough) but I'm not sure how to integrate them into the existed codebase. ```python from cer import CER cer = CER() class TestCER(unittest.TestCase): def test_cer_case_senstive(self): refs = ['White House'] preds = ['white house'] # S = 2, D = 0, I = 0, N = 11, CER = 2 / 11 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.1818181818) < 1e-6) def test_cer_whitespace(self): refs = ['were wolf'] preds = ['werewolf'] # S = 0, D = 0, I = 1, N = 9, CER = 1 / 9 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.1111111) < 1e-6) refs = ['werewolf'] preds = ['weae wolf'] # S = 1, D = 1, I = 0, N = 8, CER = 0.25 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.25) < 1e-6) # consecutive whitespaces case 1 refs = ['were wolf'] preds = ['were wolf'] # S = 0, D = 0, I = 0, N = 9, CER = 0 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.0) < 1e-6) # consecutive whitespaces case 2 refs = ['were wolf'] preds = ['were wolf'] # S = 0, D = 0, I = 0, N = 9, CER = 0 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.0) < 1e-6) def test_cer_sub(self): refs = ['werewolf'] preds = ['weaewolf'] # S = 1, D = 0, I = 0, N = 8, CER = 0.125 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.125) < 1e-6) def test_cer_del(self): refs = ['werewolf'] preds = ['wereawolf'] # S = 0, D = 1, I = 0, N = 8, CER = 0.125 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.125) < 1e-6) def test_cer_insert(self): refs = ['werewolf'] preds = ['wereolf'] # S = 0, D = 0, I = 1, N = 8, CER = 0.125 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.125) < 1e-6) def test_cer_equal(self): refs = ['werewolf'] char_error_rate = cer.compute(predictions=refs, references=refs) self.assertEqual(char_error_rate, 0.0) def test_cer_list_of_seqs(self): refs = ['werewolf', 'I am your father'] char_error_rate = cer.compute(predictions=refs, references=refs) self.assertEqual(char_error_rate, 0.0) refs = ['werewolf', 'I am your father', 'doge'] preds = ['werxwolf', 'I am your father', 'doge'] # S = 1, D = 0, I = 0, N = 28, CER = 1 / 28 char_error_rate = cer.compute(predictions=preds, references=refs) self.assertTrue(abs(char_error_rate - 0.03571428) < 1e-6) def test_cer_unicode(self): ref = [u'我能吞下玻璃而不伤身体'] pred = [u' 能吞虾玻璃而 不霜身体啦'] # S = 3, D = 2, I = 0, N = 11 # CER = 5 / 11 char_error_rate = cer.compute(predictions=pred, references=ref) self.assertTrue(abs(char_error_rate - 0.4545454545) < 1e-6) ref = [u'我能吞', u'下玻璃而不伤身体'] pred = [u'我 能 吞 下 玻 璃', u'而不伤身体'] # S = 0, D = 5, I = 0, N = 11 # CER = 5 / 11 char_error_rate = cer.compute(predictions=pred, references=ref) self.assertTrue(abs(char_error_rate - 0.454545454545) < 1e-6) ref = [u'我能吞下玻璃而不伤身体'] char_error_rate = cer.compute(predictions=ref, references=ref) self.assertFalse(char_error_rate, 0.0) def test_cer_empty(self): ref = '' pred = 'Hypothesis' with self.assertRaises(ValueError): char_error_rate = cer.compute(predictions=pred, references=ref) if __name__ == '__main__': unittest.main() ```
true
843,502,835
https://api.github.com/repos/huggingface/datasets/issues/2137
https://github.com/huggingface/datasets/pull/2137
2,137
Fix missing infos from concurrent dataset loading
closed
0
2021-03-29T15:46:12
2021-03-31T10:35:56
2021-03-31T10:35:55
lhoestq
[]
This should fix issue #2131 When calling `load_dataset` at the same time from 2 workers, one of the worker could have missing split infos when reloading the dataset from the cache.
true
843,492,015
https://api.github.com/repos/huggingface/datasets/issues/2136
https://github.com/huggingface/datasets/pull/2136
2,136
fix dialogue action slot name and value
closed
0
2021-03-29T15:34:13
2021-03-31T12:48:02
2021-03-31T12:48:01
adamlin120
[]
fix #2128
true
843,246,344
https://api.github.com/repos/huggingface/datasets/issues/2135
https://github.com/huggingface/datasets/issues/2135
2,135
en language data from MLQA dataset is missing
closed
3
2021-03-29T10:47:50
2021-03-30T10:20:23
2021-03-30T10:20:23
rabeehk
[]
Hi I need mlqa-translate-train.en dataset, but it is missing from the MLQA dataset. could you have a look please? @lhoestq thank you for your help to fix this issue.
false
843,242,849
https://api.github.com/repos/huggingface/datasets/issues/2134
https://github.com/huggingface/datasets/issues/2134
2,134
Saving large in-memory datasets with save_to_disk crashes because of pickling
closed
6
2021-03-29T10:43:15
2021-05-03T17:59:21
2021-05-03T17:59:21
prokopCerny
[ "bug" ]
Using Datasets 1.5.0 on Python 3.7. Recently I've been working on medium to large size datasets (pretokenized raw text sizes from few gigabytes to low tens of gigabytes), and have found out that several preprocessing steps are massively faster when done in memory, and I have the ability to requisition a lot of RAM, so I decided to do these steps completely out of the datasets library. So my workflow is to do several .map() on datasets object, then for the operation which is faster in memory to extract the necessary columns from the dataset and then drop it whole, do the transformation in memory, and then create a fresh Dataset object using .from_dict() or other method. When I then try to call save_to_disk(path) on the dataset, it crashes because of pickling, which appears to be because of using old pickle protocol which doesn't support large files (over 4 GiB). ``` Traceback (most recent call last): File "./tokenize_and_chunkify_in_memory.py", line 80, in <module> main() File "./tokenize_and_chunkify_in_memory.py", line 75, in main tokenize_and_chunkify(config) File "./tokenize_and_chunkify_in_memory.py", line 60, in tokenize_and_chunkify contexts_dataset.save_to_disk(chunked_path) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 457, in save_to_disk self = pickle.loads(pickle.dumps(self)) OverflowError: cannot serialize a bytes object larger than 4 GiB ``` From what I've seen this issue may be possibly fixed, as the line `self = pickle.loads(pickle.dumps(self))` does not appear to be present in the current state of the repository. To save these datasets to disk, I've resorted to calling .map() over them with `function=None` and specifying the .arrow cache file, and then creating a new dataset using the .from_file() method, which I can then safely save to disk. Additional issue when working with these large in-memory datasets is when using multiprocessing, is again to do with pickling. I've tried to speed up the mapping with function=None by specifying num_proc to the available cpu count, and I again get issues with transferring the dataset, with the following traceback. I am not sure if I should open a separate issue for that. ``` Traceback (most recent call last): File "./tokenize_and_chunkify_in_memory.py", line 94, in <module> main() File "./tokenize_and_chunkify_in_memory.py", line 89, in main tokenize_and_chunkify(config) File "./tokenize_and_chunkify_in_memory.py", line 67, in tokenize_and_chunkify contexts_dataset.map(function=None, cache_file_name=str(output_dir_path / "tmp.arrow"), writer_batch_size=50000, num_proc=config.threads) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1485, in map transformed_shards = [r.get() for r in results] File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1485, in <listcomp> transformed_shards = [r.get() for r in results] File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/multiprocess/pool.py", line 657, in get raise self._value File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/multiprocess/pool.py", line 431, in _handle_tasks put(task) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/multiprocess/connection.py", line 209, in send self._send_bytes(_ForkingPickler.dumps(obj)) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/multiprocess/reduction.py", line 54, in dumps cls(buf, protocol, *args, **kwds).dump(obj) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/dill/_dill.py", line 454, in dump StockPickler.dump(self, obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 437, in dump self.save(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 789, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/dill/_dill.py", line 941, in save_module_dict StockPickler.save_dict(pickler, obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 859, in save_dict self._batch_setitems(obj.items()) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 885, in _batch_setitems save(v) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 662, in save_reduce save(state) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/dill/_dill.py", line 941, in save_module_dict StockPickler.save_dict(pickler, obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 859, in save_dict self._batch_setitems(obj.items()) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 885, in _batch_setitems save(v) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 638, in save_reduce save(args) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 774, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 843, in _batch_appends save(x) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 638, in save_reduce save(args) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 774, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 846, in _batch_appends save(tmp[0]) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 638, in save_reduce save(args) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 774, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 789, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 846, in _batch_appends save(tmp[0]) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 789, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 846, in _batch_appends save(tmp[0]) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 789, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 843, in _batch_appends save(x) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 638, in save_reduce save(args) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 774, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 732, in save_bytes self._write_large_bytes(BINBYTES + pack("<I", n), obj) struct.error: 'I' format requires 0 <= number <= 4294967295Traceback (most recent call last): File "./tokenize_and_chunkify_in_memory.py", line 94, in <module> main() File "./tokenize_and_chunkify_in_memory.py", line 89, in main tokenize_and_chunkify(config) File "./tokenize_and_chunkify_in_memory.py", line 67, in tokenize_and_chunkify contexts_dataset.map(function=None, cache_file_name=str(output_dir_path / "tmp.arrow"), writer_batch_size=50000, num_proc=config.threads) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1485, in map transformed_shards = [r.get() for r in results] File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1485, in <listcomp> transformed_shards = [r.get() for r in results] File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/multiprocess/pool.py", line 657, in get raise self._value File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/multiprocess/pool.py", line 431, in _handle_tasks put(task) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/multiprocess/connection.py", line 209, in send self._send_bytes(_ForkingPickler.dumps(obj)) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/multiprocess/reduction.py", line 54, in dumps cls(buf, protocol, *args, **kwds).dump(obj) File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/dill/_dill.py", line 454, in dump StockPickler.dump(self, obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 437, in dump self.save(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 789, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/dill/_dill.py", line 941, in save_module_dict StockPickler.save_dict(pickler, obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 859, in save_dict self._batch_setitems(obj.items()) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 885, in _batch_setitems save(v) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 662, in save_reduce save(state) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/home/cernypro/dev/envs/huggingface_gpu/lib/python3.7/site-packages/dill/_dill.py", line 941, in save_module_dict StockPickler.save_dict(pickler, obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 859, in save_dict self._batch_setitems(obj.items()) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 885, in _batch_setitems save(v) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 638, in save_reduce save(args) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 774, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 843, in _batch_appends save(x) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 638, in save_reduce save(args) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 774, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 846, in _batch_appends save(tmp[0]) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 638, in save_reduce save(args) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 774, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 789, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 846, in _batch_appends save(tmp[0]) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 789, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 846, in _batch_appends save(tmp[0]) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 789, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 819, in save_list self._batch_appends(obj) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 843, in _batch_appends save(x) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 549, in save self.save_reduce(obj=obj, *rv) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 638, in save_reduce save(args) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 774, in save_tuple save(element) File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 504, in save f(self, obj) # Call unbound method with explicit self File "/mnt/appl/software/Python/3.7.4-GCCcore-8.3.0/lib/python3.7/pickle.py", line 732, in save_bytes self._write_large_bytes(BINBYTES + pack("<I", n), obj) struct.error: 'I' format requires 0 <= number <= 4294967295 ```
false
843,149,680
https://api.github.com/repos/huggingface/datasets/issues/2133
https://github.com/huggingface/datasets/issues/2133
2,133
bug in mlqa dataset
closed
3
2021-03-29T09:03:09
2021-03-30T17:40:57
2021-03-30T17:40:57
dorost1234
[]
Hi Looking into MLQA dataset for langauge "ar": ``` "question": [ "\u0645\u062a\u0649 \u0628\u062f\u0627\u062a \u0627\u0644\u0645\u062c\u0644\u0629 \u0627\u0644\u0645\u062f\u0631\u0633\u064a\u0629 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645 \u0628\u0627\u0644\u0646\u0634\u0631?", "\u0643\u0645 \u0645\u0631\u0629 \u064a\u062a\u0645 \u0646\u0634\u0631\u0647\u0627 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?", "\u0645\u0627 \u0647\u064a \u0627\u0644\u0648\u0631\u0642\u0629 \u0627\u0644\u064a\u0648\u0645\u064a\u0629 \u0644\u0644\u0637\u0644\u0627\u0628 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?", "\u0643\u0645 \u0639\u062f\u062f \u0627\u0644\u0627\u0648\u0631\u0627\u0642 \u0627\u0644\u0627\u062e\u0628\u0627\u0631\u064a\u0629 \u0644\u0644\u0637\u0644\u0627\u0628 \u0627\u0644\u062a\u064a \u0648\u062c\u062f\u062a \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?", "\u0641\u064a \u0627\u064a \u0633\u0646\u0629 \u0628\u062f\u0627\u062a \u0648\u0631\u0642\u0629 \u0627\u0644\u0637\u0627\u0644\u0628 \u0627\u0644\u062d\u0633 \u0627\u0644\u0633\u0644\u064a\u0645 \u0628\u0627\u0644\u0646\u0634\u0631 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645?" ] ``` the questions are in the wrong format, and not readable, could you please have a look? thanks @lhoestq
false
843,142,822
https://api.github.com/repos/huggingface/datasets/issues/2132
https://github.com/huggingface/datasets/issues/2132
2,132
TydiQA dataset is mixed and is not split per language
open
3
2021-03-29T08:56:21
2021-04-04T09:57:15
null
dorost1234
[]
Hi @lhoestq Currently TydiQA is mixed and user can only access the whole training set of all languages: https://www.tensorflow.org/datasets/catalog/tydi_qa for using this dataset, one need to train/evaluate in each separate language, and having them mixed, makes it hard to use this dataset. This is much convenient for user to have them split and I appreciate your help on this. Meanwhile, till hopefully this is split per language, I greatly appreciate telling me how I can preprocess and get data per language. thanks a lot
false
843,133,112
https://api.github.com/repos/huggingface/datasets/issues/2131
https://github.com/huggingface/datasets/issues/2131
2,131
When training with Multi-Node Multi-GPU the worker 2 has TypeError: 'NoneType' object
closed
3
2021-03-29T08:45:58
2021-04-10T11:08:55
2021-04-10T11:08:55
andy-yangz
[ "bug" ]
version: 1.5.0 met a very strange error, I am training large scale language model, and need train on 2 machines(workers). And sometimes I will get this error `TypeError: 'NoneType' object is not iterable` This is traceback ``` 71 |   | Traceback (most recent call last): -- | -- | -- 72 |   | File "run_gpt.py", line 316, in <module> 73 |   | main() 74 |   | File "run_gpt.py", line 222, in main 75 |   | delimiter="\t", column_names=["input_ids", "attention_mask", "chinese_ref"]) 76 |   | File "/data/miniconda3/lib/python3.7/site-packages/datasets/load.py", line 747, in load_dataset 77 |   | use_auth_token=use_auth_token, 78 |   | File "/data/miniconda3/lib/python3.7/site-packages/datasets/builder.py", line 513, in download_and_prepare 79 |   | self.download_post_processing_resources(dl_manager) 80 |   | File "/data/miniconda3/lib/python3.7/site-packages/datasets/builder.py", line 673, in download_post_processing_resources 81 |   | for split in self.info.splits: 82 |   | TypeError: 'NoneType' object is not iterable 83 |   | WARNING:datasets.builder:Reusing dataset csv (/usr/local/app/.cache/huggingface/datasets/csv/default-1c257ebd48e225e7/0.0.0/2960f95a26e85d40ca41a230ac88787f715ee3003edaacb8b1f0891e9f04dda2) 84 |   | Traceback (most recent call last): 85 |   | File "/data/miniconda3/lib/python3.7/runpy.py", line 193, in _run_module_as_main 86 |   | "__main__", mod_spec) 87 |   | File "/data/miniconda3/lib/python3.7/runpy.py", line 85, in _run_code 88 |   | exec(code, run_globals) 89 |   | File "/data/miniconda3/lib/python3.7/site-packages/torch/distributed/launch.py", line 340, in <module> 90 |   | main() 91 |   | File "/data/miniconda3/lib/python3.7/site-packages/torch/distributed/launch.py", line 326, in main 92 |   | sigkill_handler(signal.SIGTERM, None) # not coming back 93 |   | File "/data/miniconda3/lib/python3.7/site-packages/torch/distributed/launch.py", line 301, in sigkill_handler 94 |   | raise subprocess.CalledProcessError(returncode=last_return_code, cmd=cmd) ``` On worker 1 it loads the dataset well, however on worker 2 will get this error. And I will meet this error from time to time, sometimes it just goes well.
false
843,111,936
https://api.github.com/repos/huggingface/datasets/issues/2130
https://github.com/huggingface/datasets/issues/2130
2,130
wikiann dataset is missing columns
closed
5
2021-03-29T08:23:00
2021-08-27T14:44:18
2021-08-27T14:44:18
dorost1234
[ "good first issue" ]
Hi Wikiann dataset needs to have "spans" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq
false
843,033,656
https://api.github.com/repos/huggingface/datasets/issues/2129
https://github.com/huggingface/datasets/issues/2129
2,129
How to train BERT model with next sentence prediction?
closed
4
2021-03-29T06:48:03
2021-04-01T04:58:40
2021-04-01T04:58:40
jnishi
[]
Hello. I'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction like ` TextDatasetForNextSentencePrediction` of `huggingface/transformers` ?
false
843,023,910
https://api.github.com/repos/huggingface/datasets/issues/2128
https://github.com/huggingface/datasets/issues/2128
2,128
Dialogue action slot name and value are reversed in MultiWoZ 2.2
closed
1
2021-03-29T06:34:02
2021-03-31T12:48:01
2021-03-31T12:48:01
adamlin120
[ "dataset bug" ]
Hi @yjernite, thank you for adding MultiWoZ 2.2 in the huggingface datasets platform. It is beneficial! I spot an error that the order of Dialogue action slot names and values are reversed. https://github.com/huggingface/datasets/blob/649b2c469779bc4221e1b6969aa2496d63eb5953/datasets/multi_woz_v22/multi_woz_v22.py#L251-L262
false
843,017,199
https://api.github.com/repos/huggingface/datasets/issues/2127
https://github.com/huggingface/datasets/pull/2127
2,127
make documentation more clear to use different cloud storage
closed
0
2021-03-29T06:24:06
2021-03-29T12:16:24
2021-03-29T12:16:24
philschmid
[]
This PR extends the cloud storage documentation. To show you can use a different `fsspec` implementation.
true
842,779,966
https://api.github.com/repos/huggingface/datasets/issues/2126
https://github.com/huggingface/datasets/pull/2126
2,126
Replace legacy torch.Tensor constructor with torch.tensor
closed
0
2021-03-28T16:57:30
2021-03-29T09:27:14
2021-03-29T09:27:13
mariosasko
[]
The title says it all (motivated by [this issue](https://github.com/pytorch/pytorch/issues/53146) in the pytorch repo).
true
842,690,570
https://api.github.com/repos/huggingface/datasets/issues/2125
https://github.com/huggingface/datasets/issues/2125
2,125
Is dataset timit_asr broken?
closed
2
2021-03-28T08:30:18
2021-03-28T12:29:25
2021-03-28T12:29:25
kosuke-kitahara
[]
Using `timit_asr` dataset, I saw all records are the same. ``` python from datasets import load_dataset, load_metric timit = load_dataset("timit_asr") from datasets import ClassLabel import random import pandas as pd from IPython.display import display, HTML def show_random_elements(dataset, num_examples=10): assert num_examples <= len(dataset), "Can't pick more elements than there are in the dataset." picks = [] for _ in range(num_examples): pick = random.randint(0, len(dataset)-1) while pick in picks: pick = random.randint(0, len(dataset)-1) picks.append(pick) df = pd.DataFrame(dataset[picks]) display(HTML(df.to_html())) show_random_elements(timit['train'].remove_columns(["file", "phonetic_detail", "word_detail", "dialect_region", "id", "sentence_type", "speaker_id"]), num_examples=20) ``` `output` <img width="312" alt="Screen Shot 2021-03-28 at 17 29 04" src="https://user-images.githubusercontent.com/42398050/112746646-21acee80-8feb-11eb-84f3-dbb5d4269724.png"> I double-checked it [here](https://huggingface.co/datasets/viewer/), and met the same problem. <img width="1374" alt="Screen Shot 2021-03-28 at 17 32 07" src="https://user-images.githubusercontent.com/42398050/112746698-9bdd7300-8feb-11eb-97ed-5babead385f4.png">
false
842,627,729
https://api.github.com/repos/huggingface/datasets/issues/2124
https://github.com/huggingface/datasets/issues/2124
2,124
Adding ScaNN library to do MIPS?
open
1
2021-03-28T00:07:00
2021-03-29T13:23:43
null
shamanez
[]
@lhoestq Hi I am thinking of adding this new google library to do the MIPS similar to **add_faiss_idex**. As the paper suggests, it is really fast when it comes to retrieving the nearest neighbors. https://github.com/google-research/google-research/tree/master/scann ![image](https://user-images.githubusercontent.com/16892570/112738294-78ec9800-8fc6-11eb-9a5f-3d7ee5818e76.png)
false
842,577,285
https://api.github.com/repos/huggingface/datasets/issues/2123
https://github.com/huggingface/datasets/issues/2123
2,123
Problem downloading GEM wiki_auto_asset_turk dataset
closed
5
2021-03-27T18:41:28
2021-05-12T16:15:18
2021-05-12T16:15:17
mille-s
[]
@yjernite ### Summary I am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code. ### Steps to reproduce Code snippet: from datasets import load_dataset #dataset = load_dataset('gem', 'web_nlg_en') dataset = load_dataset('gem', 'wiki_auto_asset_turk') ``` **Expected behavior:** I expect the dataset to start downloading (download bar appears and progresses toward 100%) **Actual behavior:** Instead of seeing the download bar appearing, nothing happens; the following appears in the console as expected, but nothing more: Downloading: 36.6kB [00:00, 37.2MB/s] Downloading: 41.7kB [00:00, ?B/s] Downloading and preparing dataset gem/wiki_auto_asset_turk (download: 121.37 MiB, generated: 145.69 MiB, post-processed: Unknown size, total: 267.07 MiB) to C:\Users\sfmil\.cache\huggingface\datasets\gem\wiki_auto_asset_turk\1.0.0\f252756d7f1b8f019aac71a1623b2950acfe10d25d956668ac4eae4e93c58b8d... ### Is this a regression? No, it was the first time I was trying to download this dataset (same for the other ones). ### Debug info - Python version: Python 3.8.2 - OS version: Windows 10 Family
false
842,194,588
https://api.github.com/repos/huggingface/datasets/issues/2122
https://github.com/huggingface/datasets/pull/2122
2,122
Fast table queries with interpolation search
closed
0
2021-03-26T18:09:20
2021-08-04T18:11:59
2021-04-06T14:33:01
lhoestq
[]
## Intro This should fix issue #1803 Currently querying examples in a dataset is O(n) because of the underlying pyarrow ChunkedArrays implementation. To fix this I implemented interpolation search that is pretty effective since datasets usually verifies the condition of evenly distributed chunks (the default chunk size is fixed). ## Benchmark Here is a [benchmark](https://pastebin.com/utEXUqsR) I did on bookcorpus (74M rows): for the current implementation ```python >>> python speed.py Loaded dataset 'bookcorpus', len=74004228, nbytes=4835358766 ========================= Querying unshuffled bookcorpus ========================= Avg access time key=1 : 0.018ms Avg access time key=74004227 : 0.215ms Avg access time key=range(74003204, 74004228) : 1.416ms Avg access time key=RandIter(low=0, high=74004228, size=1024, seed=42): 92.532ms ========================== Querying shuffled bookcorpus ========================== Avg access time key=1 : 0.187ms Avg access time key=74004227 : 6.642ms Avg access time key=range(74003204, 74004228) : 90.941ms Avg access time key=RandIter(low=0, high=74004228, size=1024, seed=42): 3448.456ms ``` for the new one using interpolation search: ```python >>> python speed.py Loaded dataset 'bookcorpus', len=74004228, nbytes=4835358766 ========================= Querying unshuffled bookcorpus ========================= Avg access time key=1 : 0.076ms Avg access time key=74004227 : 0.056ms Avg access time key=range(74003204, 74004228) : 1.807ms Avg access time key=RandIter(low=0, high=74004228, size=1024, seed=42): 24.028ms ========================== Querying shuffled bookcorpus ========================== Avg access time key=1 : 0.061ms Avg access time key=74004227 : 0.058ms Avg access time key=range(74003204, 74004228) : 22.166ms Avg access time key=RandIter(low=0, high=74004228, size=1024, seed=42): 42.757ms ``` The RandIter class is just an iterable of 1024 random indices from 0 to 74004228. Here is also a plot showing the speed improvement depending on the dataset size: ![image](https://user-images.githubusercontent.com/42851186/112673587-32335c80-8e65-11eb-9a0c-58ad774abaec.png) ## Implementation details: - `datasets.table.Table` objects implement interpolation search for the `slice` method - The interpolation search requires to store the offsets of all the chunks of a table. The offsets are stored when the `Table` is initialized. - `datasets.table.Table.slice` returns a `datasets.table.Table` using interpolation search - `datasets.table.Table.fast_slice` returns a `pyarrow.Table` object using interpolation search. This is useful to get a part of a dataset if we don't need the indexing structure for future computations. For example it's used when querying an example as a dictionary. - Now a `Dataset` object is always backed by a `datasets.table.Table` object. If one passes a `pyarrow.Table` to initialize a `Dataset`, then it's converted to a `datasets.table.Table` ## Checklist: - [x] implement interpolation search - [x] use `datasets.table.Table` in `Dataset` objects - [x] update current tests - [x] add tests for interpolation search - [x] comments and docstring - [x] add the benchmark to the CI Fix #1803.
true
842,148,633
https://api.github.com/repos/huggingface/datasets/issues/2121
https://github.com/huggingface/datasets/pull/2121
2,121
Add Validation For README
closed
7
2021-03-26T17:02:17
2021-05-10T13:17:18
2021-05-10T09:41:41
gchhablani
[]
Hi @lhoestq, @yjernite This is a simple Readme parser. All classes specific to different sections can inherit `Section` class, and we can define more attributes in each. Let me know if this is going in the right direction :) Currently the output looks like this, for `to_dict()` on `FashionMNIST` `README.md`: ```json { "name": "./datasets/fashion_mnist/README.md", "attributes": "", "subsections": [ { "name": "Dataset Card for FashionMNIST", "attributes": "", "subsections": [ { "name": "Table of Contents", "attributes": "- [Dataset Description](#dataset-description)\n - [Dataset Summary](#dataset-summary)\n - [Supported Tasks](#supported-tasks-and-leaderboards)\n - [Languages](#languages)\n- [Dataset Structure](#dataset-structure)\n - [Data Instances](#data-instances)\n - [Data Fields](#data-instances)\n - [Data Splits](#data-instances)\n- [Dataset Creation](#dataset-creation)\n - [Curation Rationale](#curation-rationale)\n - [Source Data](#source-data)\n - [Annotations](#annotations)\n - [Personal and Sensitive Information](#personal-and-sensitive-information)\n- [Considerations for Using the Data](#considerations-for-using-the-data)\n - [Social Impact of Dataset](#social-impact-of-dataset)\n - [Discussion of Biases](#discussion-of-biases)\n - [Other Known Limitations](#other-known-limitations)\n- [Additional Information](#additional-information)\n - [Dataset Curators](#dataset-curators)\n - [Licensing Information](#licensing-information)\n - [Citation Information](#citation-information)\n - [Contributions](#contributions)", "subsections": [] }, { "name": "Dataset Description", "attributes": "- **Homepage:** [GitHub](https://github.com/zalandoresearch/fashion-mnist)\n- **Repository:** [GitHub](https://github.com/zalandoresearch/fashion-mnist)\n- **Paper:** [arXiv](https://arxiv.org/pdf/1708.07747.pdf)\n- **Leaderboard:**\n- **Point of Contact:**", "subsections": [ { "name": "Dataset Summary", "attributes": "Fashion-MNIST is a dataset of Zalando's article images\u2014consisting of a training set of 60,000 examples and a test set of 10,000 examples. Each example is a 28x28 grayscale image, associated with a label from 10 classes. We intend Fashion-MNIST to serve as a direct drop-in replacement for the original MNIST dataset for benchmarking machine learning algorithms. It shares the same image size and structure of training and testing splits.", "subsections": [] }, { "name": "Supported Tasks and Leaderboards", "attributes": "[More Information Needed]", "subsections": [] }, { "name": "Languages", "attributes": "[More Information Needed]", "subsections": [] } ] }, { "name": "Dataset Structure", "attributes": "", "subsections": [ { "name": "Data Instances", "attributes": "A data point comprises an image and its label.", "subsections": [] }, { "name": "Data Fields", "attributes": "- `image`: a 2d array of integers representing the 28x28 image.\n- `label`: an integer between 0 and 9 representing the classes with the following mapping:\n | Label | Description |\n | --- | --- |\n | 0 | T-shirt/top |\n | 1 | Trouser |\n | 2 | Pullover |\n | 3 | Dress |\n | 4 | Coat |\n | 5 | Sandal |\n | 6 | Shirt |\n | 7 | Sneaker |\n | 8 | Bag |\n | 9 | Ankle boot |", "subsections": [] }, { "name": "Data Splits", "attributes": "The data is split into training and test set. The training set contains 60,000 images and the test set 10,000 images.", "subsections": [] } ] }, { "name": "Dataset Creation", "attributes": "", "subsections": [ { "name": "Curation Rationale", "attributes": "**From the arXiv paper:**\nThe original MNIST dataset contains a lot of handwritten digits. Members of the AI/ML/Data Science community love this dataset and use it as a benchmark to validate their algorithms. In fact, MNIST is often the first dataset researchers try. \"If it doesn't work on MNIST, it won't work at all\", they said. \"Well, if it does work on MNIST, it may still fail on others.\"\nHere are some good reasons:\n- MNIST is too easy. Convolutional nets can achieve 99.7% on MNIST. Classic machine learning algorithms can also achieve 97% easily. Check out our side-by-side benchmark for Fashion-MNIST vs. MNIST, and read \"Most pairs of MNIST digits can be distinguished pretty well by just one pixel.\"\n- MNIST is overused. In this April 2017 Twitter thread, Google Brain research scientist and deep learning expert Ian Goodfellow calls for people to move away from MNIST.\n- MNIST can not represent modern CV tasks, as noted in this April 2017 Twitter thread, deep learning expert/Keras author Fran\u00e7ois Chollet.", "subsections": [] }, { "name": "Source Data", "attributes": "", "subsections": [ { "name": "Initial Data Collection and Normalization", "attributes": "**From the arXiv paper:**\nFashion-MNIST is based on the assortment on Zalando\u2019s website. Every fashion product on Zalando has a set of pictures shot by professional photographers, demonstrating different aspects of the product, i.e. front and back looks, details, looks with model and in an outfit. The original picture has a light-gray background (hexadecimal color: #fdfdfd) and stored in 762 \u00d7 1000 JPEG format. For efficiently serving different frontend components, the original picture is resampled with multiple resolutions, e.g. large, medium, small, thumbnail and tiny.\nWe use the front look thumbnail images of 70,000 unique products to build Fashion-MNIST. Those products come from different gender groups: men, women, kids and neutral. In particular, whitecolor products are not included in the dataset as they have low contrast to the background. The thumbnails (51 \u00d7 73) are then fed into the following conversion pipeline:\n1. Converting the input to a PNG image.\n2. Trimming any edges that are close to the color of the corner pixels. The \u201ccloseness\u201d is defined by the distance within 5% of the maximum possible intensity in RGB space.\n3. Resizing the longest edge of the image to 28 by subsampling the pixels, i.e. some rows and columns are skipped over.\n4. Sharpening pixels using a Gaussian operator of the radius and standard deviation of 1.0, with increasing effect near outlines.\n5. Extending the shortest edge to 28 and put the image to the center of the canvas.\n6. Negating the intensities of the image.\n7. Converting the image to 8-bit grayscale pixels.", "subsections": [] }, { "name": "Who are the source image producers?", "attributes": "**From the arXiv paper:**\nEvery fashion product on Zalando has a set of pictures shot by professional photographers, demonstrating different aspects of the product, i.e. front and back looks, details, looks with model and in an outfit.", "subsections": [] } ] }, { "name": "Annotations", "attributes": "", "subsections": [ { "name": "Annotation process", "attributes": "**From the arXiv paper:**\nFor the class labels, they use the silhouette code of the product. The silhouette code is manually labeled by the in-house fashion experts and reviewed by a separate team at Zalando. Each product Zalando is the Europe\u2019s largest online fashion platform. Each product contains only one silhouette code.", "subsections": [] }, { "name": "Who are the annotators?", "attributes": "**From the arXiv paper:**\nThe silhouette code is manually labeled by the in-house fashion experts and reviewed by a separate team at Zalando.", "subsections": [] } ] }, { "name": "Personal and Sensitive Information", "attributes": "[More Information Needed]", "subsections": [] } ] }, { "name": "Considerations for Using the Data", "attributes": "", "subsections": [ { "name": "Social Impact of Dataset", "attributes": "[More Information Needed]", "subsections": [] }, { "name": "Discussion of Biases", "attributes": "[More Information Needed]", "subsections": [] }, { "name": "Other Known Limitations", "attributes": "[More Information Needed]", "subsections": [] } ] }, { "name": "Additional Information", "attributes": "", "subsections": [ { "name": "Dataset Curators", "attributes": "Han Xiao and Kashif Rasul and Roland Vollgraf", "subsections": [] }, { "name": "Licensing Information", "attributes": "MIT Licence", "subsections": [] }, { "name": "Citation Information", "attributes": "@article{DBLP:journals/corr/abs-1708-07747,\n author = {Han Xiao and\n Kashif Rasul and\n Roland Vollgraf},\n title = {Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning\n Algorithms},\n journal = {CoRR},\n volume = {abs/1708.07747},\n year = {2017},\n url = {http://arxiv.org/abs/1708.07747},\n archivePrefix = {arXiv},\n eprint = {1708.07747},\n timestamp = {Mon, 13 Aug 2018 16:47:27 +0200},\n biburl = {https://dblp.org/rec/bib/journals/corr/abs-1708-07747},\n bibsource = {dblp computer science bibliography, https://dblp.org}\n}", "subsections": [] }, { "name": "Contributions", "attributes": "Thanks to [@gchhablani](https://github.com/gchablani) for adding this dataset.", "subsections": [] } ] } ] } ] } ``` Thanks, Gunjan
true
841,954,521
https://api.github.com/repos/huggingface/datasets/issues/2120
https://github.com/huggingface/datasets/issues/2120
2,120
dataset viewer does not work anymore
closed
2
2021-03-26T13:22:13
2021-03-26T15:52:22
2021-03-26T15:52:22
dorost1234
[ "nlp-viewer" ]
Hi I normally use this link to see all datasets and how I can load them https://huggingface.co/datasets/viewer/ Now I am getting 502 Bad Gateway nginx/1.18.0 (Ubuntu) could you bring this webpage back ? this was very helpful @lhoestq thanks for your help
false
841,567,199
https://api.github.com/repos/huggingface/datasets/issues/2119
https://github.com/huggingface/datasets/pull/2119
2,119
copy.deepcopy os.environ instead of copy
closed
0
2021-03-26T03:58:38
2021-03-26T15:13:52
2021-03-26T15:13:52
NihalHarish
[]
Fixes: https://github.com/huggingface/datasets/issues/2115 - bug fix: using envrion.copy() returns a dict. - using deepcopy(environ) returns an `_environ` object - Changing the datatype of the _environ object can break code, if subsequent libraries perform operations using apis exclusive to the environ object, like `environ.getenv()` for example. Testing: Tested the change on my terminal: ``` >>> import os >>> x = deepcopy(os.environ) >>> y = os.environ >>> x is y False >>> isinstance(x, type(os.environ)) True >>> z = os.environ.copy() >>> isinstance(z, type(os.environ)) False >>> isinstance(z, dict) True ```
true
841,563,329
https://api.github.com/repos/huggingface/datasets/issues/2118
https://github.com/huggingface/datasets/pull/2118
2,118
Remove os.environ.copy in Dataset.map
closed
1
2021-03-26T03:48:17
2021-03-26T12:03:23
2021-03-26T12:00:05
mariosasko
[]
Replace `os.environ.copy` with in-place modification Fixes #2115
true
841,535,283
https://api.github.com/repos/huggingface/datasets/issues/2117
https://github.com/huggingface/datasets/issues/2117
2,117
load_metric from local "glue.py" meet error 'NoneType' object is not callable
closed
3
2021-03-26T02:35:22
2021-08-25T21:44:05
2021-03-26T02:40:26
Frankie123421
[]
actual_task = "mnli" if task == "mnli-mm" else task dataset = load_dataset(path='/home/glue.py', name=actual_task) metric = load_metric(path='/home/glue.py', name=actual_task) --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-8-7ab77a465d81> in <module> 1 actual_task = "mnli" if task == "mnli-mm" else task 2 dataset = load_dataset(path='/home/jcli/glue.py', name=actual_task) ----> 3 metric = load_metric(path='/home/jcli/glue.py', name=actual_task) ~/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs) 508 keep_in_memory=keep_in_memory, 509 experiment_id=experiment_id, --> 510 **metric_init_kwargs, 511 ) 512 TypeError: 'NoneType' object is not callable Please help
false
841,481,292
https://api.github.com/repos/huggingface/datasets/issues/2116
https://github.com/huggingface/datasets/issues/2116
2,116
Creating custom dataset results in error while calling the map() function
closed
1
2021-03-26T00:37:46
2021-03-31T14:30:32
2021-03-31T14:30:32
GeetDsa
[]
calling `map()` of `datasets` library results into an error while defining a Custom dataset. Reproducible example: ``` import datasets class MyDataset(datasets.Dataset): def __init__(self, sentences): "Initialization" self.samples = sentences def __len__(self): "Denotes the total number of samples" return len(self.samples) def __getitem__(self, index): "Generates one sample of data" # Select sample # Load data and get label samples = self.samples[index] return samples def preprocess_function_train(examples): inputs = examples labels = [example+tokenizer.eos_token for example in examples ] inputs = tokenizer(inputs, max_length=30, padding=True, truncation=True) labels = tokenizer(labels, max_length=30, padding=True, truncation=True) model_inputs = inputs model_inputs["labels"] = labels["input_ids"] print("about to return") return model_inputs ##train["sentence"] is dataframe column train_dataset = MyDataset(train['sentence'].values.tolist()) train_dataset = train_dataset.map( preprocess_function, batched = True, batch_size=32 ) ``` Stack trace of error: ``` Traceback (most recent call last): File "dir/train_generate.py", line 362, in <module> main() File "dir/train_generate.py", line 245, in main train_dataset = train_dataset.map( File "anaconda_dir/anaconda3/envs/env1/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 1244, in map return self._map_single( File "anaconda_dir/anaconda3/envs/env1/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 149, in wrapper unformatted_columns = set(self.column_names) - set(self._format_columns or []) File "anaconda_dir/anaconda3/envs/env1/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 526, in column_names return self._data.column_names AttributeError: 'MyDataset' object has no attribute '_data' ```
false
841,283,974
https://api.github.com/repos/huggingface/datasets/issues/2115
https://github.com/huggingface/datasets/issues/2115
2,115
The datasets.map() implementation modifies the datatype of os.environ object
closed
0
2021-03-25T20:29:19
2021-03-26T15:13:52
2021-03-26T15:13:52
leleamol
[]
In our testing, we noticed that the datasets.map() implementation is modifying the datatype of python os.environ object from '_Environ' to 'dict'. This causes following function calls to fail as follows: ` x = os.environ.get("TEST_ENV_VARIABLE_AFTER_dataset_map", default=None) TypeError: get() takes no keyword arguments ` It looks like the following line in datasets.map implementation introduced this functionality. https://github.com/huggingface/datasets/blob/0cb1ac06acb0df44a1cf4128d03a01865faa2504/src/datasets/arrow_dataset.py#L1421 Here is the test script to reproduce this error. ``` from datasets import load_dataset from transformers import AutoTokenizer import os def test_train(): model_checkpoint = "distilgpt2" datasets = load_dataset('wikitext', 'wikitext-2-raw-v1') tokenizer = AutoTokenizer.from_pretrained(model_checkpoint, use_fast=True) tokenizer.pad_token = tokenizer.eos_token def tokenize_function(examples): y = tokenizer(examples['text'], truncation=True, max_length=64) return y x = os.environ.get("TEST_ENV_VARIABLE_BEFORE_dataset_map", default=None) print(f"Testing environment variable: TEST_ENV_VARIABLE_BEFORE_dataset_map {x}") print(f"Data type of os.environ before datasets.map = {os.environ.__class__.__name__}") datasets.map(tokenize_function, batched=True, num_proc=2, remove_columns=["text"]) print(f"Data type of os.environ after datasets.map = {os.environ.__class__.__name__}") x = os.environ.get("TEST_ENV_VARIABLE_AFTER_dataset_map", default=None) print(f"Testing environment variable: TEST_ENV_VARIABLE_AFTER_dataset_map {x}") if __name__ == "__main__": test_train() ```
false
841,207,878
https://api.github.com/repos/huggingface/datasets/issues/2114
https://github.com/huggingface/datasets/pull/2114
2,114
Support for legal NLP datasets (EURLEX, ECtHR cases and EU-REG-IR)
closed
2
2021-03-25T18:40:17
2021-03-31T10:38:50
2021-03-31T10:38:50
iliaschalkidis
[]
Add support for two legal NLP datasets: - EURLEX (https://www.aclweb.org/anthology/P19-1636/) - ECtHR cases (https://arxiv.org/abs/2103.13084) - EU-REG-IR (https://arxiv.org/abs/2101.10726)
true
841,191,303
https://api.github.com/repos/huggingface/datasets/issues/2113
https://github.com/huggingface/datasets/pull/2113
2,113
Implement Dataset as context manager
closed
0
2021-03-25T18:18:30
2021-03-31T11:30:14
2021-03-31T08:30:11
albertvillanova
[]
When used as context manager, it would be safely deleted if some exception is raised. This will avoid > During handling of the above exception, another exception occurred:
true
841,098,008
https://api.github.com/repos/huggingface/datasets/issues/2112
https://github.com/huggingface/datasets/pull/2112
2,112
Support for legal NLP datasets (EURLEX and ECtHR cases)
closed
0
2021-03-25T16:24:17
2021-03-25T18:39:31
2021-03-25T18:34:31
iliaschalkidis
[]
Add support for two legal NLP datasets: - EURLEX (https://www.aclweb.org/anthology/P19-1636/) - ECtHR cases (https://arxiv.org/abs/2103.13084)
true
841,082,087
https://api.github.com/repos/huggingface/datasets/issues/2111
https://github.com/huggingface/datasets/pull/2111
2,111
Compute WER metric iteratively
closed
7
2021-03-25T16:06:48
2021-04-06T07:20:43
2021-04-06T07:20:43
albertvillanova
[]
Compute WER metric iteratively to avoid MemoryError. Fix #2078.
true
840,794,995
https://api.github.com/repos/huggingface/datasets/issues/2110
https://github.com/huggingface/datasets/pull/2110
2,110
Fix incorrect assertion in builder.py
closed
2
2021-03-25T10:39:20
2021-04-12T13:33:03
2021-04-12T13:33:03
dreamgonfly
[]
Fix incorrect num_examples comparison assertion in builder.py
true
840,746,598
https://api.github.com/repos/huggingface/datasets/issues/2109
https://github.com/huggingface/datasets/pull/2109
2,109
Add more issue templates and customize issue template chooser
closed
2
2021-03-25T09:41:53
2021-04-19T06:20:11
2021-04-19T06:20:11
albertvillanova
[]
When opening an issue, it is not evident for the users how to choose a blank issue template. There is a link at the bottom of all the other issue templates (`Don’t see your issue here? Open a blank issue.`), but this is not very visible for users. This is the reason why many users finally chose the `add-dataset` template instead (this is more visible) for issues that indeed are not requesting the addition of a new dataset. ~~With this PR, the default blank issue template would be as visible as the other templates (as the `add-dataset` template), thus making easier for the users to choose it.~~ With this PR: - more issue templates, besides `add-dataset`, are added: `bug-report` and `feature-request` - the issue template chooser is customized, so that it now includes a link to `Discussions` for questions
true
840,181,055
https://api.github.com/repos/huggingface/datasets/issues/2108
https://github.com/huggingface/datasets/issues/2108
2,108
Is there a way to use a GPU only when training an Index in the process of add_faisis_index?
open
0
2021-03-24T21:32:16
2021-03-25T06:31:43
null
shamanez
[ "question" ]
Motivation - Some FAISS indexes like IVF consist of the training step that clusters the dataset into a given number of indexes. It would be nice if we can use a GPU to do the training step and covert the index back to CPU as mention in [this faiss example](https://gist.github.com/mdouze/46d6bbbaabca0b9778fca37ed2bcccf6).
false
839,495,825
https://api.github.com/repos/huggingface/datasets/issues/2107
https://github.com/huggingface/datasets/pull/2107
2,107
Metadata validation
closed
5
2021-03-24T08:52:41
2021-04-26T08:27:14
2021-04-26T08:27:13
theo-m
[]
- `pydantic` metadata schema with dedicated validators against our taxonomy - ci script to validate new changes against this schema and start a vertuous loop - soft validation on tasks ids since we expect the taxonomy to undergo some changes in the near future for reference with the current validation we have ~365~ 378 datasets with invalid metadata! full error report [_here_.](https://gist.github.com/theo-m/61b3c0c47fc6121d08d3174bd4c2a26b)
true
839,084,264
https://api.github.com/repos/huggingface/datasets/issues/2106
https://github.com/huggingface/datasets/issues/2106
2,106
WMT19 Dataset for Kazakh-English is not formatted correctly
open
1
2021-03-23T20:14:47
2021-03-25T21:36:20
null
trina731
[ "dataset bug" ]
In addition to the bug of languages being switched from Issue @415, there are incorrect translations in the dataset because the English-Kazakh translations have a one off formatting error. The News Commentary v14 parallel data set for kk-en from http://www.statmt.org/wmt19/translation-task.html has a bug here: > Line 94. The Swiss National Bank, for its part, has been battling with the deflationary effects of the franc’s dramatic appreciation over the past few years. Швейцарияның Ұлттық банкі өз тарапынан, соңғы бірнеше жыл ішінде франк құнының қатты өсуінің дефляциялық әсерімен күресіп келеді. > > Line 95. Дефляциялық күштер 2008 жылы терең және ұзаққа созылған жаһандық дағдарысқа байланысты орын алған ірі экономикалық және қаржылық орын алмасулардың арқасында босатылды. Жеке қарыз қаражаты үлесінің қысқаруы орталық банктің рефляцияға жұмсалған күш-жігеріне тұрақты соққан қарсы желдей болды. > > Line 96. The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008. Private deleveraging became a steady headwind to central bank efforts to reflate. 2009 жылы, алдыңғы қатарлы экономикалардың шамамен үштен бірі бағаның төмендеуін көрсетті, бұл соғыстан кейінгі жоғары деңгей болды. As you can see, line 95 has only the Kazakh translation which should be part of line 96. This causes all of the following English-Kazakh translation pairs to be one off rendering ALL of those translations incorrect. This issue was not fixed when the dataset was imported to Huggingface. By running this code ``` import datasets from datasets import load_dataset dataset = load_dataset('wmt19', 'kk-en') for key in dataset['train']['translation']: if 'The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008.' in key['kk']: print(key['en']) print(key['kk']) break ``` we get: > 2009 жылы, алдыңғы қатарлы экономикалардың шамамен үштен бірі бағаның төмендеуін көрсетті, бұл соғыстан кейінгі жоғары деңгей болды. > The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008. Private deleveraging became a steady headwind to central bank efforts to reflate. which shows that the issue still persists in the Huggingface dataset. The Kazakh sentence matches up to the next English sentence in the dataset instead of the current one. Please let me know if there's you have any ideas to fix this one-off error from the dataset or if this can be fixed by Huggingface.
false
839,059,226
https://api.github.com/repos/huggingface/datasets/issues/2105
https://github.com/huggingface/datasets/issues/2105
2,105
Request to remove S2ORC dataset
open
3
2021-03-23T19:43:06
2021-08-04T19:18:02
null
kyleclo
[]
Hi! I was wondering if it's possible to remove [S2ORC](https://huggingface.co/datasets/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!
false
839,027,834
https://api.github.com/repos/huggingface/datasets/issues/2104
https://github.com/huggingface/datasets/issues/2104
2,104
Trouble loading wiki_movies
closed
2
2021-03-23T18:59:54
2022-03-30T08:22:58
2022-03-30T08:22:58
adityaarunsinghal
[]
Hello, I am trying to load_dataset("wiki_movies") and it gives me this error - `FileNotFoundError: Couldn't find file locally at wiki_movies/wiki_movies.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.2/datasets/wiki_movies/wiki_movies.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/wiki_movies/wiki_movies.py` Trying to do `python run_mlm.py \ --model_name_or_path roberta-base \ --dataset_name wiki_movies \` also gives the same error. Is this something on my end? From what I can tell, this dataset was re-added by @lhoestq a few months ago. Thank you!
false
838,946,916
https://api.github.com/repos/huggingface/datasets/issues/2103
https://github.com/huggingface/datasets/issues/2103
2,103
citation, homepage, and license fields of `dataset_info.json` are duplicated many times
closed
1
2021-03-23T17:18:09
2021-04-06T14:39:59
2021-04-06T14:39:59
samsontmr
[ "enhancement", "good first issue" ]
This happens after a `map` operation when `num_proc` is set to `>1`. I tested this by cleaning up the json before running the `map` op on the dataset so it's unlikely it's coming from an earlier concatenation. Example result: ``` "citation": "@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n ``` @lhoestq and I believe this is happening due to the fields being concatenated `num_proc` times.
false
838,794,090
https://api.github.com/repos/huggingface/datasets/issues/2102
https://github.com/huggingface/datasets/pull/2102
2,102
Move Dataset.to_csv to csv module
closed
0
2021-03-23T14:35:46
2021-03-24T14:07:35
2021-03-24T14:07:34
albertvillanova
[ "refactoring" ]
Move the implementation of `Dataset.to_csv` to module `datasets.io.csv`.
true
838,586,184
https://api.github.com/repos/huggingface/datasets/issues/2101
https://github.com/huggingface/datasets/pull/2101
2,101
MIAM dataset - new citation details
closed
2
2021-03-23T10:41:23
2021-03-23T18:08:10
2021-03-23T18:08:10
eusip
[]
Hi @lhoestq, I have updated the citations to reference an OpenReview preprint.
true
838,574,631
https://api.github.com/repos/huggingface/datasets/issues/2100
https://github.com/huggingface/datasets/pull/2100
2,100
Fix deprecated warning message and docstring
closed
3
2021-03-23T10:27:52
2021-03-24T08:19:41
2021-03-23T18:03:49
albertvillanova
[ "documentation" ]
Fix deprecated warnings: - Use deprecated Sphinx directive in docstring - Fix format of deprecated message - Raise FutureWarning
true
838,523,819
https://api.github.com/repos/huggingface/datasets/issues/2099
https://github.com/huggingface/datasets/issues/2099
2,099
load_from_disk takes a long time to load local dataset
closed
8
2021-03-23T09:28:37
2021-03-23T17:12:16
2021-03-23T17:12:16
samsontmr
[]
I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though). Does anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers? Tagging @lhoestq since you seem to be working on these issues and PRs :)
false
838,447,959
https://api.github.com/repos/huggingface/datasets/issues/2098
https://github.com/huggingface/datasets/issues/2098
2,098
SQuAD version
closed
2
2021-03-23T07:47:54
2021-03-26T09:48:54
2021-03-26T09:48:54
h-peng17
[]
Hi~ I want train on squad dataset. What's the version of the squad? Is it 1.1 or 1.0? I'm new in QA, I don't find some descriptions about it.
false
838,105,289
https://api.github.com/repos/huggingface/datasets/issues/2097
https://github.com/huggingface/datasets/pull/2097
2,097
fixes issue #1110 by descending further if `obj["_type"]` is a dict
closed
0
2021-03-22T21:00:55
2021-03-22T21:01:11
2021-03-22T21:01:11
dcfidalgo
[]
Check metrics
true
838,038,379
https://api.github.com/repos/huggingface/datasets/issues/2096
https://github.com/huggingface/datasets/issues/2096
2,096
CoNLL 2003 dataset not including German
closed
2
2021-03-22T19:23:56
2023-07-25T16:49:07
2023-07-25T16:49:07
rxian
[ "dataset request" ]
Hello, thanks for all the work on developing and maintaining this amazing platform, which I am enjoying working with! I was wondering if there is a reason why the German CoNLL 2003 dataset is not included in the [repository](https://github.com/huggingface/datasets/tree/master/datasets/conll2003), since a copy of it could be found in some places on the internet such as GitHub? I could help adding the German data to the hub, unless there are some copyright issues that I am unaware of... This is considering that many work use the union of CoNLL 2002 and 2003 datasets for comparing cross-lingual NER transfer performance in `en`, `de`, `es`, and `nl`. E.g., [XLM-R](https://www.aclweb.org/anthology/2020.acl-main.747.pdf). ## Adding a Dataset - **Name:** CoNLL 2003 German - **Paper:** https://www.aclweb.org/anthology/W03-0419/ - **Data:** https://github.com/huggingface/datasets/tree/master/datasets/conll2003
false
837,209,211
https://api.github.com/repos/huggingface/datasets/issues/2093
https://github.com/huggingface/datasets/pull/2093
2,093
Fix: Allows a feature to be named "_type"
closed
4
2021-03-21T23:21:57
2021-03-25T14:35:54
2021-03-25T14:35:54
dcfidalgo
[]
This PR tries to fix issue #1110. Sorry for taking so long to come back to this. It's a simple fix, but i am not sure if it works for all possible types of `obj`. Let me know what you think @lhoestq
true
836,984,043
https://api.github.com/repos/huggingface/datasets/issues/2092
https://github.com/huggingface/datasets/issues/2092
2,092
How to disable making arrow tables in load_dataset ?
closed
7
2021-03-21T04:50:07
2022-06-01T16:49:52
2022-06-01T16:49:52
Jeevesh8
[]
Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?
false
836,831,403
https://api.github.com/repos/huggingface/datasets/issues/2091
https://github.com/huggingface/datasets/pull/2091
2,091
Fix copy snippet in docs
closed
0
2021-03-20T15:08:22
2021-03-24T08:20:50
2021-03-23T17:18:31
mariosasko
[ "documentation" ]
With this change the lines starting with `...` in the code blocks can be properly copied to clipboard.
true
836,807,498
https://api.github.com/repos/huggingface/datasets/issues/2090
https://github.com/huggingface/datasets/pull/2090
2,090
Add machine translated multilingual STS benchmark dataset
closed
6
2021-03-20T13:28:07
2021-03-29T13:24:42
2021-03-29T13:00:15
PhilipMay
[]
also see here https://github.com/PhilipMay/stsb-multi-mt
true
836,788,019
https://api.github.com/repos/huggingface/datasets/issues/2089
https://github.com/huggingface/datasets/issues/2089
2,089
Add documentaton for dataset README.md files
closed
8
2021-03-20T11:44:38
2023-07-25T16:45:38
2023-07-25T16:45:37
PhilipMay
[]
Hi, the dataset README files have special headers. Somehow a documenation of the allowed values and tags is missing. Could you add that? Just to give some concrete questions that should be answered imo: - which values can be passted to multilinguality? - what should be passed to language_creators? - which values should licenses have? What do I say when it is a custom license? Should I add a link? - how should I choose size_categories ? What are valid ranges? - what are valid task_categories? Thanks Philip
false
836,763,733
https://api.github.com/repos/huggingface/datasets/issues/2088
https://github.com/huggingface/datasets/pull/2088
2,088
change bibtex template to author instead of authors
closed
1
2021-03-20T09:23:44
2021-03-23T15:40:12
2021-03-23T15:40:12
PhilipMay
[]
Hi, IMO when using BibTex Author should be used instead of Authors. See here: http://www.bibtex.org/Using/de/ Thanks Philip
true
836,587,392
https://api.github.com/repos/huggingface/datasets/issues/2087
https://github.com/huggingface/datasets/pull/2087
2,087
Update metadata if dataset features are modified
closed
4
2021-03-20T02:05:23
2021-04-09T09:25:33
2021-04-09T09:25:33
mariosasko
[]
This PR adds a decorator that updates the dataset metadata if a previously executed transform modifies its features. Fixes #2083
true
836,249,587
https://api.github.com/repos/huggingface/datasets/issues/2086
https://github.com/huggingface/datasets/pull/2086
2,086
change user permissions to -rw-r--r--
closed
1
2021-03-19T18:14:56
2021-03-24T13:59:04
2021-03-24T13:59:04
bhavitvyamalik
[]
Fix for #2065
true
835,870,994
https://api.github.com/repos/huggingface/datasets/issues/2085
https://github.com/huggingface/datasets/pull/2085
2,085
Fix max_wait_time in requests
closed
0
2021-03-19T11:22:26
2021-03-23T15:36:38
2021-03-23T15:36:37
lhoestq
[]
it was handled as a min time, not max cc @SBrandeis
true
835,750,671
https://api.github.com/repos/huggingface/datasets/issues/2084
https://github.com/huggingface/datasets/issues/2084
2,084
CUAD - Contract Understanding Atticus Dataset
closed
1
2021-03-19T09:27:43
2021-04-16T08:50:44
2021-04-16T08:50:44
theo-m
[ "dataset request" ]
## Adding a Dataset - **Name:** CUAD - Contract Understanding Atticus Dataset - **Description:** As one of the only large, specialized NLP benchmarks annotated by experts, CUAD can serve as a challenging research benchmark for the broader NLP community. - **Paper:** https://arxiv.org/abs/2103.06268 - **Data:** https://github.com/TheAtticusProject/cuad/ - **Motivation:** good domain specific datasets are valuable Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
false
835,695,425
https://api.github.com/repos/huggingface/datasets/issues/2083
https://github.com/huggingface/datasets/issues/2083
2,083
`concatenate_datasets` throws error when changing the order of datasets to concatenate
closed
1
2021-03-19T08:29:48
2021-04-09T09:25:33
2021-04-09T09:25:33
patrickvonplaten
[]
Hey, I played around with the `concatenate_datasets(...)` function: https://huggingface.co/docs/datasets/package_reference/main_classes.html?highlight=concatenate_datasets#datasets.concatenate_datasets and noticed that when the order in which the datasets are concatenated changes an error is thrown where it should not IMO. Here is a google colab to reproduce the error: https://colab.research.google.com/drive/17VTFU4KQ735-waWZJjeOHS6yDTfV5ekK?usp=sharing
false
835,401,555
https://api.github.com/repos/huggingface/datasets/issues/2082
https://github.com/huggingface/datasets/pull/2082
2,082
Updated card using information from data statement and datasheet
closed
0
2021-03-19T00:39:38
2021-03-19T14:29:09
2021-03-19T14:29:09
mcmillanmajora
[]
I updated and clarified the REFreSD [data card](https://github.com/mcmillanmajora/datasets/blob/refresd_card/datasets/refresd/README.md) with information from the Eleftheria's [website](https://elbria.github.io/post/refresd/). I added brief descriptions where the initial card referred to the paper, and I also recreated some of the tables in the paper to show relevant dataset statistics. I'll email Eleftheria to see if she has any comments on the card.
true
835,112,968
https://api.github.com/repos/huggingface/datasets/issues/2081
https://github.com/huggingface/datasets/pull/2081
2,081
Fix docstrings issues
closed
0
2021-03-18T18:11:01
2021-04-07T14:37:43
2021-04-07T14:37:43
albertvillanova
[ "documentation" ]
Fix docstring issues.
true
835,023,000
https://api.github.com/repos/huggingface/datasets/issues/2080
https://github.com/huggingface/datasets/issues/2080
2,080
Multidimensional arrays in a Dataset
closed
2
2021-03-18T16:29:14
2021-03-25T12:46:53
2021-03-25T12:46:53
vermouthmjl
[]
Hi, I'm trying to put together a `datasets.Dataset` to be used with LayoutLM which is available in `transformers`. This model requires as input the bounding boxes of each of the token of a sequence. This is when I realized that `Dataset` does not support multi-dimensional arrays as a value for a column in a row. The following code results in conversion error in pyarrow (`pyarrow.lib.ArrowInvalid: ('Can only convert 1-dimensional array values', 'Conversion failed for column bbox with type object')`) ``` from datasets import Dataset import pandas as pd import numpy as np dataset = pd.DataFrame({ 'bbox': [ np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]), np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]), np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]), np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]) ], 'input_ids': [1, 2, 3, 4] }) dataset = Dataset.from_pandas(dataset) ``` Since I wanted to use pytorch for the downstream training task, I also tried a few ways to directly put in a column of 2-D pytorch tensor in a formatted dataset, but I can only have a list of 1-D tensors, or a list of arrays, or a list of lists. ``` import torch from datasets import Dataset import pandas as pd dataset = pd.DataFrame({ 'bbox': [ [[1,2,3,4],[1,2,3,4],[1,2,3,4]], [[1,2,3,4],[1,2,3,4],[1,2,3,4]], [[1,2,3,4],[1,2,3,4],[1,2,3,4]], [[1,2,3,4],[1,2,3,4],[1,2,3,4]] ], 'input_ids': [1, 2, 3, 4] }) dataset = Dataset.from_pandas(dataset) def test(examples): return {'bbbox': torch.Tensor(examples['bbox'])} dataset = dataset.map(test) print(dataset[0]['bbox']) print(dataset[0]['bbbox']) dataset.set_format(type='torch', columns=['input_ids', 'bbox'], output_all_columns=True) print(dataset[0]['bbox']) print(dataset[0]['bbbox']) def test2(examples): return {'bbbox': torch.stack(examples['bbox'])} dataset = dataset.map(test2) print(dataset[0]['bbox']) print(dataset[0]['bbbox']) ``` Is is possible to support n-D arrays/tensors in datasets? It seems that it can also be useful for this [feature request](https://github.com/huggingface/datasets/issues/263).
false
834,920,493
https://api.github.com/repos/huggingface/datasets/issues/2079
https://github.com/huggingface/datasets/pull/2079
2,079
Refactorize Metric.compute signature to force keyword arguments only
closed
0
2021-03-18T15:05:50
2021-03-23T15:31:44
2021-03-23T15:31:44
albertvillanova
[]
Minor refactoring of Metric.compute signature to force the use of keyword arguments, by using the single star syntax.
true
834,694,819
https://api.github.com/repos/huggingface/datasets/issues/2078
https://github.com/huggingface/datasets/issues/2078
2,078
MemoryError when computing WER metric
closed
11
2021-03-18T11:30:05
2021-05-01T08:31:49
2021-04-06T07:20:43
diego-fustes
[ "metric bug" ]
Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation: ``` wer = load_metric("wer") print(wer.compute(predictions=result["predicted"], references=result["target"])) ``` However, I receive the following exception: `Traceback (most recent call last): File "/home/diego/IpGlobal/wav2vec/test_wav2vec.py", line 51, in <module> print(wer.compute(predictions=result["predicted"], references=result["target"])) File "/home/diego/miniconda3/envs/wav2vec3.6/lib/python3.6/site-packages/datasets/metric.py", line 403, in compute output = self._compute(predictions=predictions, references=references, **kwargs) File "/home/diego/.cache/huggingface/modules/datasets_modules/metrics/wer/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281/wer.py", line 94, in _compute return wer(references, predictions) File "/home/diego/miniconda3/envs/wav2vec3.6/lib/python3.6/site-packages/jiwer/measures.py", line 81, in wer truth, hypothesis, truth_transform, hypothesis_transform, **kwargs File "/home/diego/miniconda3/envs/wav2vec3.6/lib/python3.6/site-packages/jiwer/measures.py", line 192, in compute_measures H, S, D, I = _get_operation_counts(truth, hypothesis) File "/home/diego/miniconda3/envs/wav2vec3.6/lib/python3.6/site-packages/jiwer/measures.py", line 273, in _get_operation_counts editops = Levenshtein.editops(source_string, destination_string) MemoryError` My system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.
false
834,649,536
https://api.github.com/repos/huggingface/datasets/issues/2077
https://github.com/huggingface/datasets/pull/2077
2,077
Bump huggingface_hub version
closed
1
2021-03-18T10:54:34
2021-03-18T11:33:26
2021-03-18T11:33:26
SBrandeis
[]
`0.0.2 => 0.0.6`
true
834,445,296
https://api.github.com/repos/huggingface/datasets/issues/2076
https://github.com/huggingface/datasets/issues/2076
2,076
Issue: Dataset download error
open
7
2021-03-18T06:36:06
2021-03-22T11:52:31
null
XuhuiZhou
[ "dataset bug" ]
The download link in `iwslt2017.py` file does not seem to work anymore. For example, `FileNotFoundError: Couldn't find file at https://wit3.fbk.eu/archive/2017-01-trnted/texts/zh/en/zh-en.tgz` Would be nice if we could modify it script and use the new downloadable link?
false
834,301,246
https://api.github.com/repos/huggingface/datasets/issues/2075
https://github.com/huggingface/datasets/issues/2075
2,075
ConnectionError: Couldn't reach common_voice.py
closed
2
2021-03-18T01:19:06
2021-03-20T10:29:41
2021-03-20T10:29:41
LifaSun
[]
When I run: from datasets import load_dataset, load_metric common_voice_train = load_dataset("common_voice", "zh-CN", split="train+validation") common_voice_test = load_dataset("common_voice", "zh-CN", split="test") Got: ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/master/datasets/common_voice/common_voice.py Version: 1.4.1 Thanks! @lhoestq @LysandreJik @thomwolf
false
834,268,463
https://api.github.com/repos/huggingface/datasets/issues/2074
https://github.com/huggingface/datasets/pull/2074
2,074
Fix size categories in YAML Tags
closed
9
2021-03-18T00:02:36
2021-03-23T17:11:10
2021-03-23T17:11:10
gchhablani
[]
This PR fixes several `size_categories` in YAML tags and makes them consistent. Additionally, I have added a few more categories after `1M`, up to `1T`. I would like to add that to the streamlit app also. This PR also adds a couple of infos that I found missing. The code for generating this: ```python for dataset in sorted(os.listdir('./datasets/')): if '.' not in dataset and dataset not in ['c4', 'csv', 'downloads', 'cc100', 'ccaligned_multilingual', 'celeb_a', 'chr_en', 'emea', 'glue']: infos = {} stats = {} st = '' with open(f'datasets/{dataset}/README.md') as f: d = f.read() start_dash = d.find('---') + 3 end_dash = d[start_dash:].find('---') + 3 rest_text = d[end_dash + 3:] try: full_yaml = OmegaConf.create(d[start_dash:end_dash]) readme = OmegaConf.to_container(full_yaml['size_categories'], resolve=True) except Exception as e: print(e) continue try: with open(f'datasets/{dataset}/dataset_infos.json') as f: data = json.load(f) except Exception as e: print(e) continue # Skip those without infos. done_set = set([]) num_keys = len(data.keys()) for keys in data: # dataset = load_dataset('opus100', f'{dirs}') total = 0 for split in data[keys]['splits']: total = total + data[keys]['splits'][split]['num_examples'] if total < 1000: st += "- n<1K" + '\n' infos[keys] = ["n<1K"] elif total >= 1000 and total < 10000: infos[keys] = ["1K<n<10K"] elif total >= 10000 and total < 100000: infos[keys] = ["10K<n<100K"] elif total >= 100000 and total < 1000000: infos[keys] = ["100K<n<1M"] elif total >= 1000000 and total < 10000000: infos[keys] = ["1M<n<10M"] elif total >= 10000000 and total < 100000000: infos[keys] = ["10M<n<100M"] elif total >= 100000000 and total < 1000000000: infos[keys] = ["100M<n<1B"] elif total >= 1000000000 and total < 10000000000: infos[keys] = ["1B<n<10B"] elif total >= 10000000000 and total < 100000000000: infos[keys] = ["10B<n<100B"] elif total >= 100000000000 and total < 1000000000000: infos[keys] = ["100B<n<1T"] else: infos[keys] = ["n>1T"] done_set = done_set.union(infos[keys]) if (isinstance(readme, list) and list(infos.values())[0] != readme) or (isinstance(readme, dict) and readme != infos): print('-' * 30) print(done_set) print(f"Changing Full YAML for {dataset}") print(OmegaConf.to_yaml(full_yaml)) if len(done_set) == 1: full_yaml['size_categories'] = list(done_set) else: full_yaml['size_categories'] = dict([(k, v) for k, v in sorted(infos.items(), key=lambda x: x[0])]) full_yaml_string = OmegaConf.to_yaml(full_yaml) print('-' * 30) print(full_yaml_string) inp = input('Do you wish to continue?(Y/N)') if inp == 'Y': with open(f'./datasets/{dataset}/README.md', 'w') as f: f.write('---\n') f.write(full_yaml_string) f.write('---') f.write(rest_text) else: break ``` Note that the lower-bound is inclusive. I'm unsure if this is how it is done in the tagging app. EDIT: It would be great if there was a way to make the task categories consistent too. For this, the streamlit app can look into all the datasets and check for existing categories and show them in the list. This may add some consistency. EDIT: I understand this will not work for cases where only the infos for some of the configs are present, for example: `ccaligned_multingual` has only 5 out of several configs present, and infos has only information about them. Hence, I have skipped a few datasets in the code, if there are more such datasets, then I'll ignore them too.
true
834,192,501
https://api.github.com/repos/huggingface/datasets/issues/2073
https://github.com/huggingface/datasets/pull/2073
2,073
Fixes check of TF_AVAILABLE and TORCH_AVAILABLE
closed
0
2021-03-17T21:28:53
2021-03-18T09:09:25
2021-03-18T09:09:24
philschmid
[]
# What is this PR doing This PR implements the checks if `Tensorflow` and `Pytorch` are available the same way as `transformers` does it. I added the additional checks for the different `Tensorflow` and `torch` versions. #2068
true
834,054,837
https://api.github.com/repos/huggingface/datasets/issues/2072
https://github.com/huggingface/datasets/pull/2072
2,072
Fix docstring issues
closed
2
2021-03-17T18:13:44
2021-03-24T08:20:57
2021-03-18T12:41:21
albertvillanova
[ "documentation" ]
Fix docstring issues.
true
833,950,824
https://api.github.com/repos/huggingface/datasets/issues/2071
https://github.com/huggingface/datasets/issues/2071
2,071
Multiprocessing is slower than single process
closed
1
2021-03-17T16:08:58
2021-03-18T09:10:23
2021-03-18T09:10:23
theo-m
[ "bug" ]
```python # benchmark_filter.py import logging import sys import time from datasets import load_dataset, set_caching_enabled if __name__ == "__main__": set_caching_enabled(False) logging.basicConfig(level=logging.DEBUG) bc = load_dataset("bookcorpus") now = time.time() try: bc["train"].filter(lambda x: len(x["text"]) < 64, num_proc=int(sys.argv[1])) except Exception as e: print(f"cancelled: {e}") elapsed = time.time() - now print(elapsed) ``` Running `python benchmark_filter.py 1` (20min+) is faster than `python benchmark_filter.py 2` (2hrs+)
false