html_url
stringlengths
51
51
title
stringlengths
6
280
comments
stringlengths
67
24.7k
body
stringlengths
51
36.2k
__index_level_0__
int64
1
1.17k
comment_length
int64
16
1.45k
text
stringlengths
190
38.3k
embeddings
list
https://github.com/huggingface/datasets/issues/5354
Consider using "Sequence" instead of "List"
Hi! Linking a comment to provide more info on the issue: https://stackoverflow.com/a/39458225. This means we should replace all (most of) the occurrences of `List` with `Sequence` in function signatures. @tranhd95 Would you be interested in submitting a PR?
### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below. **How to reproduce** ```py ...
374
38
Consider using "Sequence" instead of "List" ### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors...
[ -1.1331499814987183, -0.8368464112281799, -0.7626137733459473, 1.4945868253707886, -0.16346971690654755, -1.2774741649627686, 0.18324048817157745, -1.0449070930480957, 1.7204946279525757, -0.7336402535438538, 0.3742302656173706, -1.6610338687896729, -0.021034322679042816, -0.64282089471817...
https://github.com/huggingface/datasets/issues/5354
Consider using "Sequence" instead of "List"
Hi all! I tried to reproduce this issue and didn't work for me. Also in your example i noticed that the variables have different names: `list_of_filenames` and `list_of_files`, could this be related to that? ```python #I found random data in parquet format: !wget "https://github.com/Teradata/kylo/raw/master/samples/...
### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below. **How to reproduce** ```py ...
374
121
Consider using "Sequence" instead of "List" ### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors...
[ -1.1363391876220703, -0.8300718069076538, -0.7608848214149475, 1.5081171989440918, -0.14507760107517242, -1.3009257316589355, 0.16242757439613342, -1.038397192955017, 1.6914905309677124, -0.746339738368988, 0.33476725220680237, -1.6492044925689697, -0.0681803748011589, -0.6372345685958862,...
https://github.com/huggingface/datasets/issues/5354
Consider using "Sequence" instead of "List"
@dantema There is indeed a typo in variable names. Nevertheless, I'm sorry if I was not clear but the output is from `mypy` type checker. You can run the code snippet without issues. The problem is with the type checking.
### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below. **How to reproduce** ```py ...
374
40
Consider using "Sequence" instead of "List" ### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors...
[ -1.1417784690856934, -0.8362669944763184, -0.7446849346160889, 1.523167371749878, -0.16777461767196655, -1.2624818086624146, 0.1661750078201294, -1.0644620656967163, 1.730597972869873, -0.7514219284057617, 0.38889002799987793, -1.6621534824371338, -0.03524775430560112, -0.6129856705665588,...
https://github.com/huggingface/datasets/issues/5354
Consider using "Sequence" instead of "List"
However, I found out that the type annotation is actually misleading. The [`from_parquet`](https://github.com/huggingface/datasets/blob/5ef1ab1cc06c2b7a574bf2df454cd9fcb071ccb2/src/datasets/arrow_dataset.py#L1039) method should also accept list of [`PathLike`](https://github.com/huggingface/datasets/blob/main/src/datas...
### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below. **How to reproduce** ```py ...
374
405
Consider using "Sequence" instead of "List" ### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors...
[ -1.1314377784729004, -0.8188692331314087, -0.7313487529754639, 1.4617829322814941, -0.12871532142162323, -1.2888085842132568, 0.1325380504131317, -1.0625556707382202, 1.6795439720153809, -0.7191594839096069, 0.34144148230552673, -1.6790275573730469, -0.08184440433979034, -0.619117379188537...
https://github.com/huggingface/datasets/issues/5354
Consider using "Sequence" instead of "List"
@avinashsai Hi, I've assigned you the issue. @tranhd95 Yes, feel free to report this in a new issue.
### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below. **How to reproduce** ```py ...
374
18
Consider using "Sequence" instead of "List" ### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors...
[ -1.1352909803390503, -0.841376006603241, -0.7508730292320251, 1.5250927209854126, -0.15977616608142853, -1.2457406520843506, 0.15540334582328796, -1.0274401903152466, 1.7066551446914673, -0.7475284934043884, 0.37051212787628174, -1.6746240854263306, -0.04259897395968437, -0.622329950332641...
https://github.com/huggingface/datasets/issues/5354
Consider using "Sequence" instead of "List"
@avinashsai Are you still working on this? If not I would like to give it a try.
### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below. **How to reproduce** ```py ...
374
17
Consider using "Sequence" instead of "List" ### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors...
[ -1.140164852142334, -0.8359280824661255, -0.7365889549255371, 1.528045892715454, -0.14484205842018127, -1.2515233755111694, 0.16015976667404175, -1.0420674085617065, 1.7080374956130981, -0.7429264187812805, 0.3787233829498291, -1.659406065940857, -0.0451342798769474, -0.6294645667076111, ...
https://github.com/huggingface/datasets/issues/5352
__init__() got an unexpected keyword argument 'input_size'
Hi @J-shel, thanks for reporting. I think the issue comes from your call to `load_dataset`. As first argument, you should pass: - either the name of your dataset ("mrf") if this is already published on the Hub - or the path to the loading script of your dataset ("path/to/your/local/mrf.py").
### Describe the bug I try to define a custom configuration with a input_size attribute following the instructions by "Specifying several dataset configurations" in https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html But when I load the dataset, I got an error "__init__() got an unexpected keyword argument...
376
49
__init__() got an unexpected keyword argument 'input_size' ### Describe the bug I try to define a custom configuration with a input_size attribute following the instructions by "Specifying several dataset configurations" in https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html But when I load the dataset, ...
[ -1.3179267644882202, -0.8770233988761902, -0.6781187057495117, 1.4217880964279175, -0.23399394750595093, -1.1265684366226196, 0.21573123335838318, -1.0965675115585327, 1.5839473009109497, -0.7800825238227844, 0.32759523391723633, -1.5724815130233765, -0.059767208993434906, -0.5339155197143...
https://github.com/huggingface/datasets/issues/5352
__init__() got an unexpected keyword argument 'input_size'
Hi, following your suggestion, I changed my call to load_dataset. Below is the latest: reader = load_dataset('data/mrf.py',"default", input_size=1024, split=split, streaming=True, keep_in_memory=None) However, I still got the same error. I have one question that is if I only define input_size=2048 in BUILDER_CONFIGS...
### Describe the bug I try to define a custom configuration with a input_size attribute following the instructions by "Specifying several dataset configurations" in https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html But when I load the dataset, I got an error "__init__() got an unexpected keyword argument...
376
65
__init__() got an unexpected keyword argument 'input_size' ### Describe the bug I try to define a custom configuration with a input_size attribute following the instructions by "Specifying several dataset configurations" in https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html But when I load the dataset, ...
[ -1.319419503211975, -0.8907701373100281, -0.6713951230049133, 1.427407145500183, -0.23848478496074677, -1.1148940324783325, 0.22346144914627075, -1.092186450958252, 1.556174397468567, -0.7811620235443115, 0.31490352749824524, -1.5549242496490479, -0.07593421638011932, -0.5233897566795349, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
Hi! `DatasetBuilder` is a parent class for concrete builders: `GeneratorBasedBuilder`, `ArrowBasedBuilder` and `BeamBasedBuilder`. When writing a builder script, these classes are the ones you should inherit from. And since all of them implement `_prepare_split`, you only have to implement the three methods mentioned a...
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
44
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
Thanks so much @mariosasko for the fast response! I've been referencing [this page in the docs](https://huggingface.co/docs/datasets/v2.4.0/en/about_dataset_load) because it it pretty comprehensive in terms of what we have to do and I figured since we subclass the `BuilderConfig` the same pattern would hold, but I've a...
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
63
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
cc @stevhliu who may have some ideas on how to improve this part of the docs.
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
16
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
one more question for my understanding @mariosasko. the requirement of a loading script has always seemed counterintuitive to me. if i have to provide a script with every dataset, what is the point of using `datasets` if we're doing all the work of loading it, I can just do that in my code and skip the datasets integra...
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
274
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
> the requirement of a loading script has always seemed counterintuitive to me This is a requirement only for datasets not stored in standard formats such as CSV, JSON, SQL, Parquet, ImageFolder, etc. > if i have to provide a script with every dataset, what is the point of using datasets if we're doing all the...
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
163
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
ok, that makes sense. thank you @mariosasko. I realized i'd never looked on the hub at any of the files associated with any datasets. just did that now and it appears that i'll need to have a script regardless _but_ that will just contain my custom config and builder classes, so without realizing it I was already makin...
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
117
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
`download_and_prepare` downloads and prepares the arrow files. You need to call `as_dataset` on the builder to get the dataset.
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
19
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
ok, I think I was assigning the output of `builder.download_and_prepare` but it's an inplace op, so that explains the `NoneType` i was getting back. Now I'm getting: ``` ArrowInvalid Traceback (most recent call last) <ipython-input-7-3ed50fb87c70> in <module> ----> 1 ds = dataset_buil...
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
486
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
looks like my arrow files are all empty @mariosasko ![image](https://user-images.githubusercontent.com/7530947/208179977-9ae62c9a-866c-472b-9a09-25d1191188fb.png) i also see the `incomplete_info.lock` file a level up too. seems like the data isn't being persisted to disk when I call `download_and_prepare`. is ...
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
44
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
quick update @mariosasko. i got it working! i had to downgrade to `datasets==2.4.0`. testing other versions now and will let you know the results.
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
24
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5351
Do we need to implement `_prepare_split`?
I've tested with every version of `datasets>2.4.0` and i get the same error with all of them.
### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im...
377
17
Do we need to implement `_prepare_split`? ### Describe the bug I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr...
[ -1.1564629077911377, -0.9498602747917175, -0.663374125957489, 1.4442986249923706, -0.10583754628896713, -1.330580234527588, 0.14666317403316498, -1.096006155014038, 1.5908504724502563, -0.7755786180496216, 0.2561314105987549, -1.64324951171875, 0.0009109014645218849, -0.5524017810821533, ...
https://github.com/huggingface/datasets/issues/5348
The data downloaded in the download folder of the cache does not respect `umask`
note, that `datasets` already did some of that umask fixing in the past and also at the hub - the recent work on the hub about the same: https://github.com/huggingface/huggingface_hub/pull/1220 Also I noticed that each file has a .json counterpart and the latter always has the correct perms: ``` -rw------- 1 uue...
### Describe the bug For a project on a cluster we are several users to share the same cache for the datasets library. And we have a problem with the permissions on the data downloaded in the cache. Indeed, it seems that the data is downloaded by giving read and write permissions only to the user launching the com...
378
83
The data downloaded in the download folder of the cache does not respect `umask` ### Describe the bug For a project on a cluster we are several users to share the same cache for the datasets library. And we have a problem with the permissions on the data downloaded in the cache. Indeed, it seems that the data is...
[ -1.1730643510818481, -0.8724946975708008, -0.6110600233078003, 1.3566467761993408, -0.0034409910440444946, -1.4475382566452026, 0.09687133878469467, -0.9708036780357361, 1.5548968315124512, -0.8790223598480225, 0.3033035695552826, -1.7146375179290771, 0.0685807541012764, -0.622688114643096...
https://github.com/huggingface/datasets/issues/5345
Wrong dtype for array in audio features
After some more investigation, this is due to [this line of code](https://github.com/huggingface/datasets/blob/main/src/datasets/features/audio.py#L279). The function `sf.read(file)` should be updated to `sf.read(file, dtype="float32")` Indeed, the default value in soundfile is `float64` ([see here](https://pysoundf...
### Describe the bug When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datasets cannot be merged. ### Steps to repro...
380
31
Wrong dtype for array in audio features ### Describe the bug When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datase...
[ -1.3297818899154663, -1.0398335456848145, -0.7567360997200012, 1.6267650127410889, -0.26365041732788086, -1.0593035221099854, 0.12767237424850464, -1.0891356468200684, 1.538980484008789, -0.7270705103874207, 0.23913443088531494, -1.5932620763778687, 0.00082396250218153, -0.7497702240943909...
https://github.com/huggingface/datasets/issues/5345
Wrong dtype for array in audio features
@qmeeus I agree, decoding of different audio formats should return the same dtypes indeed! But note that here you are concatenating datasets with different sampling rates: 48000 for CommonVoice and 16000 for Voxpopuli. So you should cast them to the same sampling rate value before interleaving, for example: ``` cv...
### Describe the bug When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datasets cannot be merged. ### Steps to repro...
380
115
Wrong dtype for array in audio features ### Describe the bug When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datase...
[ -1.3297818899154663, -1.0398335456848145, -0.7567360997200012, 1.6267650127410889, -0.26365041732788086, -1.0593035221099854, 0.12767237424850464, -1.0891356468200684, 1.538980484008789, -0.7270705103874207, 0.23913443088531494, -1.5932620763778687, 0.00082396250218153, -0.7497702240943909...
https://github.com/huggingface/datasets/issues/5345
Wrong dtype for array in audio features
Hi Polina, Sorry for the late answer It is possible that the issue was due to a bug that is now fixed. I installed an editable version of datasets from github, but I don't recall whether I had updated it at the time of the issue. My research led me to other directions so I did not follow through on the interleave dat...
### Describe the bug When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datasets cannot be merged. ### Steps to repro...
380
63
Wrong dtype for array in audio features ### Describe the bug When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datase...
[ -1.3297818899154663, -1.0398335456848145, -0.7567360997200012, 1.6267650127410889, -0.26365041732788086, -1.0593035221099854, 0.12767237424850464, -1.0891356468200684, 1.538980484008789, -0.7270705103874207, 0.23913443088531494, -1.5932620763778687, 0.00082396250218153, -0.7497702240943909...
https://github.com/huggingface/datasets/issues/5342
Emotion dataset cannot be downloaded
Hi @cbarond there's already an open issue at https://github.com/dair-ai/emotion_dataset/issues/5, as the data seems to be missing now, so check that issue instead 👍🏻
### Describe the bug The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`. It was working yesterday (December 7, 2022), but stopped working today (December 8, 2022). ### Steps to reproduce the bug ...
381
23
Emotion dataset cannot be downloaded ### Describe the bug The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`. It was working yesterday (December 7, 2022), but stopped working today (December 8, 202...
[ -1.0933345556259155, -0.8987973928451538, -0.8178911209106445, 1.440356731414795, -0.18660634756088257, -1.2340970039367676, 0.13042962551116943, -0.977516770362854, 1.568128228187561, -0.7103350162506104, 0.25876522064208984, -1.6854135990142822, -0.0371609665453434, -0.6266356706619263, ...
https://github.com/huggingface/datasets/issues/5342
Emotion dataset cannot be downloaded
Thanks @cbarond for reporting and @alvarobartt for pointing to the issue we opened in the author's repo. Indeed, this issue was first raised in the "emotion" dataset Community tab: https://huggingface.co/datasets/emotion/discussions/3 I'm closing this issue and leave the issue above for the subsequent updates. ...
### Describe the bug The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`. It was working yesterday (December 7, 2022), but stopped working today (December 8, 2022). ### Steps to reproduce the bug ...
381
46
Emotion dataset cannot be downloaded ### Describe the bug The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`. It was working yesterday (December 7, 2022), but stopped working today (December 8, 202...
[ -1.1503911018371582, -0.9479205012321472, -0.8195348978042603, 1.4430569410324097, -0.15113148093223572, -1.2766895294189453, 0.11261900514364243, -0.955824613571167, 1.629323124885559, -0.6967946290969849, 0.22675691545009613, -1.6891469955444336, -0.0682714432477951, -0.551869809627533, ...
https://github.com/huggingface/datasets/issues/5342
Emotion dataset cannot be downloaded
> try using "SetFit/emotion" instead I' replaced "emotion" with "SetFit/Emotion", but the code is getting stuck at `emotions = load_dataset("SetFit/emotion")` I pause execution using the debugger, and it takes me to filelock.py:226 `with self._thread_lock:` Do you know a way to get past this issue?
### Describe the bug The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`. It was working yesterday (December 7, 2022), but stopped working today (December 8, 2022). ### Steps to reproduce the bug ...
381
44
Emotion dataset cannot be downloaded ### Describe the bug The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`. It was working yesterday (December 7, 2022), but stopped working today (December 8, 202...
[ -1.1264564990997314, -0.8865939974784851, -0.8132458329200745, 1.4566452503204346, -0.19419091939926147, -1.2872220277786255, 0.17019471526145935, -1.0140031576156616, 1.598790168762207, -0.740466296672821, 0.2705131471157074, -1.6355057954788208, 0.005785159766674042, -0.5599542260169983,...
https://github.com/huggingface/datasets/issues/5338
`map()` stops every 1000 steps
Hi ! > It starts using all the cores (I am not sure why because I did not pass num_proc) The tokenizer uses Rust code that is multithreaded. And maybe the `feature_extractor` might run some things in parallel as well - but I'm not super familiar with its internals. > then progress bar stops at every 1k steps. ...
### Describe the bug I am passing the following `prepare_dataset` function to `Dataset.map` (code is inspired from [here](https://github.com/huggingface/community-events/blob/main/whisper-fine-tuning-event/run_speech_recognition_seq2seq_streaming.py#L454)) ```python3 def prepare_dataset(batch): # load and res...
382
108
`map()` stops every 1000 steps ### Describe the bug I am passing the following `prepare_dataset` function to `Dataset.map` (code is inspired from [here](https://github.com/huggingface/community-events/blob/main/whisper-fine-tuning-event/run_speech_recognition_seq2seq_streaming.py#L454)) ```python3 def prepare_da...
[ -1.2085660696029663, -0.8807992935180664, -0.6695201992988586, 1.4731371402740479, -0.08829861879348755, -1.3140623569488525, 0.18715448677539825, -1.0919111967086792, 1.6589032411575317, -0.7869752645492554, 0.3509739637374878, -1.6116012334823608, 0.056128792464733124, -0.595148921012878...
https://github.com/huggingface/datasets/issues/5338
`map()` stops every 1000 steps
Hi @lhoestq Thanks for the explanation! it was so helpful! Let me check why `feature_extractor` is running on multiple cpus.
### Describe the bug I am passing the following `prepare_dataset` function to `Dataset.map` (code is inspired from [here](https://github.com/huggingface/community-events/blob/main/whisper-fine-tuning-event/run_speech_recognition_seq2seq_streaming.py#L454)) ```python3 def prepare_dataset(batch): # load and res...
382
20
`map()` stops every 1000 steps ### Describe the bug I am passing the following `prepare_dataset` function to `Dataset.map` (code is inspired from [here](https://github.com/huggingface/community-events/blob/main/whisper-fine-tuning-event/run_speech_recognition_seq2seq_streaming.py#L454)) ```python3 def prepare_da...
[ -1.2085660696029663, -0.8807992935180664, -0.6695201992988586, 1.4731371402740479, -0.08829861879348755, -1.3140623569488525, 0.18715448677539825, -1.0919111967086792, 1.6589032411575317, -0.7869752645492554, 0.3509739637374878, -1.6116012334823608, 0.056128792464733124, -0.595148921012878...
https://github.com/huggingface/datasets/issues/5337
Support webdataset format
I like the idea of having `webdataset` as an optional dependency to ensure our loader generates web datasets the same way as the main project.
Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234. In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local directory, or from a dataset o...
383
25
Support webdataset format Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234. In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local di...
[ -1.1585583686828613, -0.9291451573371887, -0.9457713961601257, 1.511222243309021, -0.28723976016044617, -1.4082257747650146, 0.09250763058662415, -1.1086831092834473, 1.6967360973358154, -0.8459455370903015, 0.39014437794685364, -1.738334059715271, 0.11118140071630478, -0.4907057583332062,...
https://github.com/huggingface/datasets/issues/5337
Support webdataset format
Webdataset is the one of the most popular dataset formats for large scale computer vision tasks. Upvote for this issue.
Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234. In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local directory, or from a dataset o...
383
20
Support webdataset format Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234. In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local di...
[ -1.1678622961044312, -0.9333143830299377, -0.9526174068450928, 1.5264627933502197, -0.2599416971206665, -1.4019267559051514, 0.1297958940267563, -1.1392103433609009, 1.6976109743118286, -0.7957373261451721, 0.3557114601135254, -1.7333347797393799, 0.1317697912454605, -0.48345568776130676, ...
https://github.com/huggingface/datasets/issues/5337
Support webdataset format
We haven't had the bandwidth to implement it so far, but if someone wants to give it a shot please don't hesitate ^^
Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234. In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local directory, or from a dataset o...
383
23
Support webdataset format Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234. In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local di...
[ -1.1931383609771729, -0.9310227632522583, -0.9216461777687073, 1.5134222507476807, -0.2293902337551117, -1.3462190628051758, 0.0987570658326149, -1.08004629611969, 1.7124196290969849, -0.8611600399017334, 0.37386053800582886, -1.7161235809326172, 0.1091696247458458, -0.5106855630874634, ...
https://github.com/huggingface/datasets/issues/5332
Passing numpy array to ClassLabel names causes ValueError
Should `datasets` allow `ClassLabel` input parameter to be an `np.array` even though internally we need to cast it to a Python list? @lhoestq @mariosasko
### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: If I define my classes as: ``` my_classes = np.array(['on...
384
24
Passing numpy array to ClassLabel names causes ValueError ### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: ...
[ -1.2423409223556519, -0.9427435398101807, -0.6524389982223511, 1.6519964933395386, -0.1358901411294937, -1.1631975173950195, 0.15079466998577118, -0.955312967300415, 1.5839974880218506, -0.7523292303085327, 0.3405283987522125, -1.6442815065383911, 0.02535400539636612, -0.6284711360931396, ...
https://github.com/huggingface/datasets/issues/5332
Passing numpy array to ClassLabel names causes ValueError
Hi! No, I don't think so. The `names` parameter is [annotated](https://github.com/huggingface/datasets/blob/582236640b9109988e5f7a16a8353696ffa09a16/src/datasets/features/features.py#L892) as `List[str]` (**NumPy arrays are not lists**), and considering that type checking is not a common practice in Python, I think we ...
### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: If I define my classes as: ``` my_classes = np.array(['on...
384
38
Passing numpy array to ClassLabel names causes ValueError ### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: ...
[ -1.2423409223556519, -0.9427435398101807, -0.6524389982223511, 1.6519964933395386, -0.1358901411294937, -1.1631975173950195, 0.15079466998577118, -0.955312967300415, 1.5839974880218506, -0.7523292303085327, 0.3405283987522125, -1.6442815065383911, 0.02535400539636612, -0.6284711360931396, ...
https://github.com/huggingface/datasets/issues/5332
Passing numpy array to ClassLabel names causes ValueError
I appreciate it is the wrong type, and that type checking is not common, but I think there's a few circumstances that make it a good idea from a usability perspective. It's quite a difficult error to debug because it comes from a utility function (so it's not immediately obvious which parameter caused it). What make...
### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: If I define my classes as: ``` my_classes = np.array(['on...
384
234
Passing numpy array to ClassLabel names causes ValueError ### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: ...
[ -1.2423409223556519, -0.9427435398101807, -0.6524389982223511, 1.6519964933395386, -0.1358901411294937, -1.1631975173950195, 0.15079466998577118, -0.955312967300415, 1.5839974880218506, -0.7523292303085327, 0.3405283987522125, -1.6442815065383911, 0.02535400539636612, -0.6284711360931396, ...
https://github.com/huggingface/datasets/issues/5332
Passing numpy array to ClassLabel names causes ValueError
I agree that it's not easy to debug this issue, so perhaps we could add some basic type checking (e.g. `not isinstance(names, list)` -> error) to make debugging easier. Feel free to submit a PR. > Additionally, passing in a set instead of the list (again, relatively easy because people may run set(classes) to genera...
### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: If I define my classes as: ``` my_classes = np.array(['on...
384
100
Passing numpy array to ClassLabel names causes ValueError ### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: ...
[ -1.2423409223556519, -0.9427435398101807, -0.6524389982223511, 1.6519964933395386, -0.1358901411294937, -1.1631975173950195, 0.15079466998577118, -0.955312967300415, 1.5839974880218506, -0.7523292303085327, 0.3405283987522125, -1.6442815065383911, 0.02535400539636612, -0.6284711360931396, ...
https://github.com/huggingface/datasets/issues/5332
Passing numpy array to ClassLabel names causes ValueError
What about checking for `Sequence` instead? I think users can pass a list or a tuple as well.
### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: If I define my classes as: ``` my_classes = np.array(['on...
384
18
Passing numpy array to ClassLabel names causes ValueError ### Describe the bug If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error. ### Steps to reproduce the bug https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX TLDR: ...
[ -1.2423409223556519, -0.9427435398101807, -0.6524389982223511, 1.6519964933395386, -0.1358901411294937, -1.1631975173950195, 0.15079466998577118, -0.955312967300415, 1.5839974880218506, -0.7523292303085327, 0.3405283987522125, -1.6442815065383911, 0.02535400539636612, -0.6284711360931396, ...
https://github.com/huggingface/datasets/issues/5325
map(...batch_size=None) for IterableDataset
Hi! I agree it makes sense for `IterableDataset.map` to support the `batch_size=None` case. This should be super easy to fix.
### Feature request Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too. ### Motivation Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple of reasons why this might be nice. One is th...
385
20
map(...batch_size=None) for IterableDataset ### Feature request Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too. ### Motivation Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple o...
[ -1.2428652048110962, -1.0249131917953491, -0.705143392086029, 1.4132028818130493, -0.15908421576023102, -1.2323696613311768, 0.1825770139694214, -1.0238515138626099, 1.763264536857605, -0.8696856498718262, 0.3509354889392853, -1.7236558198928833, 0.08649272471666336, -0.44251522421836853, ...
https://github.com/huggingface/datasets/issues/5325
map(...batch_size=None) for IterableDataset
@mariosasko as this is something simple maybe I can include it as part of https://github.com/huggingface/datasets/pull/5311? Let me know :+1:
### Feature request Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too. ### Motivation Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple of reasons why this might be nice. One is th...
385
19
map(...batch_size=None) for IterableDataset ### Feature request Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too. ### Motivation Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple o...
[ -1.2332842350006104, -1.0495638847351074, -0.7255744934082031, 1.4376425743103027, -0.13801564276218414, -1.2274912595748901, 0.12961739301681519, -0.9991413354873657, 1.7146615982055664, -0.8204090595245361, 0.3283476233482361, -1.7594983577728271, 0.03388161212205887, -0.4589964747428894...
https://github.com/huggingface/datasets/issues/5325
map(...batch_size=None) for IterableDataset
Thanks again :) > For practical usages, an alternative to this would be to convert from an iterable dataset to a map-style dataset, but it is not obvious how to do this. This is interesting as well, if anyone wants to explore
### Feature request Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too. ### Motivation Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple of reasons why this might be nice. One is th...
385
42
map(...batch_size=None) for IterableDataset ### Feature request Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too. ### Motivation Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple o...
[ -1.2331047058105469, -1.01775324344635, -0.6940887570381165, 1.4135165214538574, -0.16815601289272308, -1.214682936668396, 0.1424940675497055, -1.0177501440048218, 1.7147804498672485, -0.8525761365890503, 0.329569548368454, -1.7453944683074951, 0.08093299716711044, -0.48796477913856506, ...
https://github.com/huggingface/datasets/issues/5324
Fix docstrings and types in documentation that appears on the website
Ok, I believe we've cleaned up most of the old syntax we were using for the user-facing docs! There are still a couple of `:obj:`'s and `:class:` floating around in the docstrings we don't expose that I'll track down :)
While I was working on https://github.com/huggingface/datasets/pull/5313 I've noticed that we have a mess in how we annotate types and format args and return values in the code. And some of it is displayed in the [Reference section](https://huggingface.co/docs/datasets/package_reference/builder_classes) of the document...
386
40
Fix docstrings and types in documentation that appears on the website While I was working on https://github.com/huggingface/datasets/pull/5313 I've noticed that we have a mess in how we annotate types and format args and return values in the code. And some of it is displayed in the [Reference section](https://hugging...
[ -1.3390648365020752, -0.9905195236206055, -0.8267769813537598, 1.5412458181381226, -0.1037326380610466, -1.3935750722885132, -0.0008664289489388466, -1.07442045211792, 1.6979824304580688, -0.8807410597801208, 0.293878436088562, -1.7097569704055786, 0.009175123646855354, -0.607541561126709,...
https://github.com/huggingface/datasets/issues/5317
`ImageFolder` performs poorly with large datasets
Hi ! ImageFolder is made for small scale datasets indeed. For large scale image datasets you better group your images in TAR archives or Arrow/Parquet files. This is true not just for ImageFolder loading performance, but also because having millions of files is not ideal for your filesystem or when moving the data arou...
### Describe the bug While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images. ## Setup * Nested directories (5 levels deep) * 3M+ images * 1 `metadata.jsonl` file ## Performance Degradation Point...
388
186
`ImageFolder` performs poorly with large datasets ### Describe the bug While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images. ## Setup * Nested directories (5 levels deep) * 3M+ images * 1 `metada...
[ -1.1987686157226562, -0.9523312449455261, -0.7291271686553955, 1.4155699014663696, -0.10784557461738586, -1.2535969018936157, 0.13453370332717896, -1.0448516607284546, 1.6751487255096436, -0.7176129817962646, 0.2097993642091751, -1.6848515272140503, -0.031679801642894745, -0.56804913282394...
https://github.com/huggingface/datasets/issues/5317
`ImageFolder` performs poorly with large datasets
@lhoestq Thanks! Perhaps it'd be worth adding a note on the documentation that `ImageFolder` is not intended for large datasets? This limitation is not intuitively obvious to someone who has not used it before, I think.
### Describe the bug While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images. ## Setup * Nested directories (5 levels deep) * 3M+ images * 1 `metadata.jsonl` file ## Performance Degradation Point...
388
36
`ImageFolder` performs poorly with large datasets ### Describe the bug While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images. ## Setup * Nested directories (5 levels deep) * 3M+ images * 1 `metada...
[ -1.1987686157226562, -0.9523312449455261, -0.7291271686553955, 1.4155699014663696, -0.10784557461738586, -1.2535969018936157, 0.13453370332717896, -1.0448516607284546, 1.6751487255096436, -0.7176129817962646, 0.2097993642091751, -1.6848515272140503, -0.031679801642894745, -0.56804913282394...
https://github.com/huggingface/datasets/issues/5317
`ImageFolder` performs poorly with large datasets
Thanks for the feedback @salieri! I opened #5329 to make it clear `ImageFolder` is not intended for large datasets. Please feel free to comment if you have any other feedback! 🙂
### Describe the bug While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images. ## Setup * Nested directories (5 levels deep) * 3M+ images * 1 `metadata.jsonl` file ## Performance Degradation Point...
388
31
`ImageFolder` performs poorly with large datasets ### Describe the bug While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images. ## Setup * Nested directories (5 levels deep) * 3M+ images * 1 `metada...
[ -1.1987686157226562, -0.9523312449455261, -0.7291271686553955, 1.4155699014663696, -0.10784557461738586, -1.2535969018936157, 0.13453370332717896, -1.0448516607284546, 1.6751487255096436, -0.7176129817962646, 0.2097993642091751, -1.6848515272140503, -0.031679801642894745, -0.56804913282394...
https://github.com/huggingface/datasets/issues/5315
Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails
EDIT: I think in this case, the metadata files (either README or JSON) should not be read (i.e. `self.info.splits` should be None). One idea: - I think ideally we should set this behavior when we pass `--save_info` to the CLI `test` - However, currently, the builder is unaware of this: `save_info` arg is not pas...
### Describe the bug If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails. That's what happened in https://huggingface.co/datasets/mrdbourke/food_vision_199_classes/discussions/2#6385f...
390
57
Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails ### Describe the bug If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails. That...
[ -1.113948941230774, -0.9091671705245972, -0.6788613200187683, 1.4712458848953247, -0.11289050430059433, -1.317491054534912, 0.1464281529188156, -1.0284781455993652, 1.6395132541656494, -0.7434434294700623, 0.2930738627910614, -1.7083076238632202, -0.017192382365465164, -0.595932126045227, ...
https://github.com/huggingface/datasets/issues/5315
Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails
> I think in this case @albertvillanova You mean in cases when the script was changed? I suggest that we: * add a check on the slice (like 'split_name[n%]) kind of format here: https://github.com/huggingface/datasets/blob/main/src/datasets/splits.py#L523 to catch things like this. * Error here happens before...
### Describe the bug If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails. That's what happened in https://huggingface.co/datasets/mrdbourke/food_vision_199_classes/discussions/2#6385f...
390
142
Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails ### Describe the bug If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails. That...
[ -1.113948941230774, -0.9091671705245972, -0.6788613200187683, 1.4712458848953247, -0.11289050430059433, -1.317491054534912, 0.1464281529188156, -1.0284781455993652, 1.6395132541656494, -0.7434434294700623, 0.2930738627910614, -1.7083076238632202, -0.017192382365465164, -0.595932126045227, ...
https://github.com/huggingface/datasets/issues/5315
Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails
I edited my previous comment: - First I proposed setting `self.info.splits` to None when `ignore_verifications=True` - I thought it was the easiest implementation because `ignore_verifications` is passed to `DatasetBuilder.download_and_prepare` - However, afterwards, I realized this might not be a good idea for ...
### Describe the bug If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails. That's what happened in https://huggingface.co/datasets/mrdbourke/food_vision_199_classes/discussions/2#6385f...
390
201
Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails ### Describe the bug If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails. That...
[ -1.113948941230774, -0.9091671705245972, -0.6788613200187683, 1.4712458848953247, -0.11289050430059433, -1.317491054534912, 0.1464281529188156, -1.0284781455993652, 1.6395132541656494, -0.7434434294700623, 0.2930738627910614, -1.7083076238632202, -0.017192382365465164, -0.595932126045227, ...
https://github.com/huggingface/datasets/issues/5314
Datasets: classification_report() got an unexpected keyword argument 'suffix'
@JonathanAlis also note that the metrics are deprecated in our `datasets` library. Please, use the new library 🤗 Evaluate instead: https://huggingface.co/docs/evaluate
https://github.com/huggingface/datasets/blob/main/metrics/seqeval/seqeval.py > import datasets predictions = [['O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-PER', 'I-PER', 'O']] references = [['O', 'O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-PER', 'I-PER', 'O']] seqeval = datasets.load_metri...
391
21
Datasets: classification_report() got an unexpected keyword argument 'suffix' https://github.com/huggingface/datasets/blob/main/metrics/seqeval/seqeval.py > import datasets predictions = [['O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-PER', 'I-PER', 'O']] references = [['O', 'O', 'O', 'B-MISC', '...
[ -1.3552393913269043, -0.9346601963043213, -0.8406415581703186, 1.6598427295684814, -0.20891276001930237, -1.05767822265625, -0.007367828860878944, -0.9546255469322205, 1.5658034086227417, -0.7757533192634583, 0.22902336716651917, -1.7130876779556274, -0.0316036082804203, -0.640803217887878...
https://github.com/huggingface/datasets/issues/5305
Dataset joelito/mc4_legal does not work with multiple files
Thanks for reporting @JoelNiklaus. Please note that since we moved all dataset loading scripts to the Hub, the issues and pull requests relative to specific datasets are directly handled on the Hub, in their Community tab. I'm transferring this issue there: https://huggingface.co/datasets/joelito/mc4_legal/discussio...
### Describe the bug The dataset https://huggingface.co/datasets/joelito/mc4_legal works for languages like bg with a single data file, but not for languages with multiple files like de. It shows zero rows for the de dataset. joelniklaus@Joels-MacBook-Pro ~/N/P/C/L/p/m/mc4_legal (main) [1]> python test_mc4_legal....
393
54
Dataset joelito/mc4_legal does not work with multiple files ### Describe the bug The dataset https://huggingface.co/datasets/joelito/mc4_legal works for languages like bg with a single data file, but not for languages with multiple files like de. It shows zero rows for the de dataset. joelniklaus@Joels-MacBook-...
[ -1.258362054824829, -0.9908857345581055, -0.795711100101471, 1.4838215112686157, -0.033149782568216324, -1.2504149675369263, 0.1168370246887207, -0.9932911396026611, 1.5552551746368408, -0.6639343500137329, 0.20511026680469513, -1.775362253189087, -0.1270586997270584, -0.5651912689208984, ...
https://github.com/huggingface/datasets/issues/5304
timit_asr doesn't load the test split.
The [timit_asr.py](https://huggingface.co/datasets/timit_asr/blob/main/timit_asr.py) script iterates over the WAV files per split directory using this: ```python wav_paths = sorted(Path(data_dir).glob(f"**/{split}/**/*.wav")) wav_paths = wav_paths if wav_paths else sorted(Path(data_dir).glob(f"**/{split.upper()}/**/...
### Describe the bug When I use the function ```timit = load_dataset('timit_asr', data_dir=data_dir)```, it only loads train split, not test split. I tried to change the directory and filename to lower case to upper case for the test split, but it does not work at all. ```python DatasetDict({ train: Datase...
394
42
timit_asr doesn't load the test split. ### Describe the bug When I use the function ```timit = load_dataset('timit_asr', data_dir=data_dir)```, it only loads train split, not test split. I tried to change the directory and filename to lower case to upper case for the test split, but it does not work at all. ``...
[ -1.1882445812225342, -1.1478807926177979, -0.7531515955924988, 1.6818410158157349, -0.19282428920269012, -1.1733320951461792, 0.07968270778656006, -0.9431273937225342, 1.548446536064148, -0.7097229361534119, 0.1579272747039795, -1.7373665571212769, -0.1452927589416504, -0.7916383147239685,...
https://github.com/huggingface/datasets/issues/5295
Extractions failed when .zip file located on read-only path (e.g., SageMaker FastFile mode)
Hi ! Thanks for reporting. Indeed the lock file should be placed in a directory with write permission (e.g. in the directory where the archive is extracted).
### Describe the bug Hi, `load_dataset()` does not work .zip files located on a read-only directory. Looks like it's because Dataset creates a lock file in the [same directory](https://github.com/huggingface/datasets/blob/df4bdd365f2abb695f113cbf8856a925bc70901b/src/datasets/utils/extract.py) as the .zip file. ...
395
27
Extractions failed when .zip file located on read-only path (e.g., SageMaker FastFile mode) ### Describe the bug Hi, `load_dataset()` does not work .zip files located on a read-only directory. Looks like it's because Dataset creates a lock file in the [same directory](https://github.com/huggingface/datasets/blo...
[ -1.2134716510772705, -0.9405736923217773, -0.6529114842414856, 1.45478355884552, -0.1730504184961319, -1.1634786128997803, 0.13532023131847382, -0.9901857972145081, 1.531798243522644, -0.7498309016227722, 0.31072452664375305, -1.6880139112472534, -0.04647236317396164, -0.52350252866745, ...
https://github.com/huggingface/datasets/issues/5295
Extractions failed when .zip file located on read-only path (e.g., SageMaker FastFile mode)
I opened https://github.com/huggingface/datasets/pull/5320 to fix this - it places the lock file in the cache directory instead of trying to put in next to the ZIP where it's read-only
### Describe the bug Hi, `load_dataset()` does not work .zip files located on a read-only directory. Looks like it's because Dataset creates a lock file in the [same directory](https://github.com/huggingface/datasets/blob/df4bdd365f2abb695f113cbf8856a925bc70901b/src/datasets/utils/extract.py) as the .zip file. ...
395
29
Extractions failed when .zip file located on read-only path (e.g., SageMaker FastFile mode) ### Describe the bug Hi, `load_dataset()` does not work .zip files located on a read-only directory. Looks like it's because Dataset creates a lock file in the [same directory](https://github.com/huggingface/datasets/blo...
[ -1.2134716510772705, -0.9405736923217773, -0.6529114842414856, 1.45478355884552, -0.1730504184961319, -1.1634786128997803, 0.13532023131847382, -0.9901857972145081, 1.531798243522644, -0.7498309016227722, 0.31072452664375305, -1.6880139112472534, -0.04647236317396164, -0.52350252866745, ...
https://github.com/huggingface/datasets/issues/5292
Missing documentation build for versions 2.7.1 and 2.6.2
- Build docs for 2.6.2: - Commit: a6a5a1cf4cdf1e0be65168aed5a327f543001fe8 - Build docs GH Action: https://github.com/huggingface/datasets/actions/runs/3539470622/jobs/5941404044 - Build docs for 2.7.1: - Commit: 5ef1ab1cc06c2b7a574bf2df454cd9fcb071ccb2 - Build docs GH Action: https://github.com/huggingfac...
After the patch releases [2.7.1](https://github.com/huggingface/datasets/releases/tag/2.7.1) and [2.6.2](https://github.com/huggingface/datasets/releases/tag/2.6.2), the online docs were not properly built (the build_documentation workflow was not triggered). There was a fix by: - #5291 However, both documentati...
396
28
Missing documentation build for versions 2.7.1 and 2.6.2 After the patch releases [2.7.1](https://github.com/huggingface/datasets/releases/tag/2.7.1) and [2.6.2](https://github.com/huggingface/datasets/releases/tag/2.6.2), the online docs were not properly built (the build_documentation workflow was not triggered). ...
[ -1.233512282371521, -0.8936444520950317, -0.8072652220726013, 1.3956623077392578, -0.07910232990980148, -1.240429162979126, 0.026858747005462646, -1.0523921251296997, 1.5705183744430542, -0.635078489780426, 0.31448063254356384, -1.733963131904602, -0.09406106919050217, -0.4470330476760864,...
https://github.com/huggingface/datasets/issues/5288
Lossy json serialization - deserialization of dataset info
Hi ! JSON is a lossy format indeed. If you want to keep the feature types or other metadata I'd encourage you to store them as well. For example you can use `dataset.info.write_to_directory` and `DatasetInfo.from_directory` to store the feature types, split info, description, license etc.
### Describe the bug Saving a dataset to disk as json (using `to_json`) and then loading it again (using `load_dataset`) results in features whose labels are not type-cast correctly. In the code snippet below, `features.label` should have a label of type `ClassLabel` but has type `Value` instead. ### Steps to re...
397
45
Lossy json serialization - deserialization of dataset info ### Describe the bug Saving a dataset to disk as json (using `to_json`) and then loading it again (using `load_dataset`) results in features whose labels are not type-cast correctly. In the code snippet below, `features.label` should have a label of type ...
[ -1.1897344589233398, -0.9814062714576721, -0.6852878928184509, 1.635185718536377, -0.21063145995140076, -1.1490113735198975, 0.1617978811264038, -1.009643316268921, 1.7089990377426147, -0.745701789855957, 0.26563555002212524, -1.6730543375015259, 0.027536578476428986, -0.6985669732093811, ...
https://github.com/huggingface/datasets/issues/5286
FileNotFoundError: Couldn't find file at https://dumps.wikimedia.org/enwiki/20220301/dumpstatus.json
I found a solution If you specifically install datasets==1.18 and then run import datasets wiki = datasets.load_dataset('wikipedia', '20200501.en') then this should work (it worked for me.)
### Describe the bug I follow the steps provided on the website [https://huggingface.co/datasets/wikipedia](https://huggingface.co/datasets/wikipedia) $ pip install apache_beam mwparserfromhell >>> from datasets import load_dataset >>> load_dataset("wikipedia", "20220301.en") however this results in the follo...
398
26
FileNotFoundError: Couldn't find file at https://dumps.wikimedia.org/enwiki/20220301/dumpstatus.json ### Describe the bug I follow the steps provided on the website [https://huggingface.co/datasets/wikipedia](https://huggingface.co/datasets/wikipedia) $ pip install apache_beam mwparserfromhell >>> from datasets...
[ -1.2083998918533325, -0.9795846939086914, -0.7561137080192566, 1.441972255706787, -0.188189297914505, -1.1714004278182983, 0.05929543077945709, -0.9904125332832336, 1.6878396272659302, -0.6838863492012024, 0.20123139023780823, -1.6519056558609009, -0.08784256130456924, -0.5095714330673218,...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
Thanks @lhoestq and @alvarobartt! This would be extremely helpful to have working for the Whisper fine-tuning event - we're **only** training using streaming mode, so it'll be quite important to have this feature working to make training as easy as possible! _c.f._ https://twitter.com/sanchitgandhi99/status/15921...
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
43
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
> Thanks @lhoestq and @alvarobartt! > > > > This would be extremely helpful to have working for the Whisper fine-tuning event - we're **only** training using streaming mode, so it'll be quite important to have this feature working to make training as easy as possible! > > > > _c.f._ https://twitter.com/sanchitgan...
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
100
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
@sanchit-gandhi PR is ready and open for review at #5287, but there's still one issue I may need @lhoestq's input :hugs:
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
21
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
Let us know @sanchit-gandhi if you need a new release of `datasets` soon with this fix included :)
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
18
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
Thanks for the fix guys! We can direct people to install `datasets` from main if that's easier!
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
17
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
Hi @asennoussi so this was already fixed and released as part of https://github.com/huggingface/datasets/releases/tag/2.8.0, so you should be able to install it as `pip install datasets==2.8.0` or just to use `pip install datasets --upgrade` to get the latest version, as of now, the https://github.com/huggingface/datas...
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
47
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
Still facing the same issue though: ``` from datasets import IterableDatasetDict, load_dataset raw_datasets = vectorized_datasets = IterableDatasetDict() raw_datasets["train"] = load_dataset("asennoussi/private", split="train", use_auth_token=True, streaming=True) raw_datasets["test"] = load_dataset("asenno...
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
120
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
Hmm weird, could you try to print ```python print("Processed features: ", vectorized_datasets['train'].features) ``` again after iterating over the `vectorized_datasets`? In the code above, should be last line :)
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
28
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
Didn't seem to fix it: ``` Original features: dict_keys(['path', 'audio', 'sentence']) Processed features: None Processed features: None ```
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
18
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
Actually the culprit looks to be this one: `vectorized_datasets = raw_datasets.map(prepare_dataset).with_format("torch")` When I remove this line: `vectorized_datasets = vectorized_datasets.remove_columns(['input_length', 'labels_length']+list(next(iter(raw_datasets.values())).features))` I still get ``` Proce...
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
28
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
> The culprit is definitely `.map` Just validated it. Any idea please? Yes, indeed `.map` losses the features, because AFAIK pre-fetching the data to infer the features is expensive and not ideal, that's part of this issue https://github.com/huggingface/datasets/issues/3888 Anyway, now you can pass the `features`...
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
101
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5284
Features of IterableDataset set to None by remove column
@asennoussi alternatively you can just call `._resolve_features()` from your `IterableDataset` and it will pre-fetch the data to resolve the features, but note that feature-inference is not as accurate as if you manually specify which features and feature-types the `IterableDataset` has, as mentioned in the comment abo...
### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) ...
399
56
Features of IterableDataset set to None by remove column ### Describe the bug The `remove_column` method of the IterableDataset sets the dataset features to None. ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset # load LS in streaming mode dataset = load_dataset("librisp...
[ -1.1740634441375732, -0.986666202545166, -0.724602222442627, 1.5060685873031616, -0.16580796241760254, -1.1591317653656006, 0.15898366272449493, -0.9965685606002808, 1.6621525287628174, -0.7151062488555908, 0.23000434041023254, -1.7126665115356445, -0.02927146479487419, -0.7022446990013123...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
Or for example an archive on GitHub releases! Before I added support for JXL (locally only, PR still pending) I was considering hosting my files on GitHub instead...
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
28
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1481211185455322, -0.8668554425239563, -0.7009403109550476, 1.5864934921264648, -0.11684641242027283, -1.2785141468048096, 0.226686030626297, -0.9447305202484131, 1.6145557165145874, -0.7712887525558472, 0.4066356420516968, -1.7100402116775513, 0.02905506268143654, -0.5384873747825623, ...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
+1 to this. I would like to use 'audiofolder' with a data_dir that's on S3, for example. I don't want to upload my dataset to the Hub, but I would find all the fingerprinting/caching features useful.
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
36
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1301863193511963, -0.887853741645813, -0.7564798593521118, 1.556461215019226, -0.11461750417947769, -1.2745490074157715, 0.21984700858592987, -0.9521751403808594, 1.6449880599975586, -0.7799682021141052, 0.39993906021118164, -1.7132654190063477, 0.010704532265663147, -0.5761101841926575...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
Adding to the conversation, Dask also uses `fsspec` for this feature. [Dask: How to connect to remote data](https://docs.dask.org/en/stable/how-to/connect-to-remote-data.html) Happy to help on this feature :D
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
25
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1261876821517944, -0.865461528301239, -0.7369795441627502, 1.5711097717285156, -0.07206477969884872, -1.266869306564331, 0.1847299188375473, -0.9479480981826782, 1.653198003768921, -0.7508422136306763, 0.4013937711715698, -1.6984187364578247, -0.013095520436763763, -0.5696078538894653, ...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
+1 to this feature request since I think it also tackles my use-case. I am collaborating with a team, working with a loading script which takes some time to generate the dataset artifacts. It would be very handy to use this as a cloud cache to avoid duplicating the effort. Currently we could use `builder.download_a...
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
78
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1604654788970947, -0.8987336158752441, -0.6765923500061035, 1.4956525564193726, -0.12263771891593933, -1.270355463027954, 0.20320464670658112, -1.019025206565857, 1.624469518661499, -0.7517934441566467, 0.37090015411376953, -1.6710683107376099, -0.02991263009607792, -0.5710994601249695,...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
Makes sense ! If you want to load locally a dataset that you download_and_prepared on a cloud storage, you would use `load_dataset(path_to_cloud_storage)` indeed. It would download the data from the cloud storage, cache them locally, and return a `Dataset`.
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
39
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1247591972351074, -0.8659507632255554, -0.7190048694610596, 1.5625264644622803, -0.10728877782821655, -1.3009120225906372, 0.19700269401073456, -0.9690713286399841, 1.6566994190216064, -0.806896448135376, 0.38287004828453064, -1.6959503889083862, -0.002525700256228447, -0.58512496948242...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
It seems currently the `cached_path` function handles all URLs by `get_from_cache` that only supports `ftp` and `http(s)` here: https://github.com/huggingface/datasets/blob/b5672a956d5de864e6f5550e493527d962d6ae55/src/datasets/utils/file_utils.py#L181 I guess one can add another condition that handles `s3://` or `g...
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
35
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1119447946548462, -0.8206652402877808, -0.7333627343177795, 1.5866179466247559, -0.06604381650686264, -1.2926174402236938, 0.2216891348361969, -0.9680379033088684, 1.628993034362793, -0.7275192737579346, 0.40322235226631165, -1.7049504518508911, -0.05162134766578674, -0.5773789286613464...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
I could use this functionality, so I put together a PR using @kyamagu's suggestion to use `fsspec` in `datasets.utils.file_utils` https://github.com/huggingface/datasets/pull/5580
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
20
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.0952835083007812, -0.8498707413673401, -0.7236981987953186, 1.5619399547576904, -0.08013491332530975, -1.2700855731964111, 0.22350482642650604, -0.9411389231681824, 1.6429386138916016, -0.7555069923400879, 0.39713624119758606, -1.700418472290039, -0.03550271317362785, -0.575844228267669...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
Thanks @dwyatte for adding support for fsspec urls Let me just reopen this since the original issue is not resolved
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
20
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1162397861480713, -0.8388081789016724, -0.6944715976715088, 1.6116211414337158, -0.1191871166229248, -1.2930155992507935, 0.19488219916820526, -0.9393096566200256, 1.6099915504455566, -0.7943482995033264, 0.4006102979183197, -1.6975775957107544, 0.011448396369814873, -0.5964275598526001...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
I'm not yet understanding how to use https://github.com/huggingface/datasets/pull/5580 in order to use `load_dataset(data_files="s3://...")`. Any help/example would be much appreciated :) thanks!
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
21
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.0874907970428467, -0.874066174030304, -0.7081629037857056, 1.5715142488479614, -0.09644901752471924, -1.2539818286895752, 0.1999494582414627, -0.9408275485038757, 1.6259489059448242, -0.7350842952728271, 0.38461852073669434, -1.6943416595458984, -0.04317580536007881, -0.5396792888641357...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
It's still not officially supported x) But you can try to update `request_etag` in `file_utils.py` to use `fsspec_head` instead of `http_head`. It is responsible of getting the ETags of the remote files for caching. This change may do the trick for S3 urls
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
43
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.0978657007217407, -0.8224558234214783, -0.7115296721458435, 1.6014676094055176, -0.09623211622238159, -1.293861985206604, 0.23310768604278564, -0.9644601941108704, 1.6769380569458008, -0.8170479536056519, 0.4120827317237854, -1.6851418018341064, 0.04258618503808975, -0.5588087439537048,...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
Thank you for your guys help on this and merging in #5580. I manually pulled the changes to my local datasets package (datasets.utils.file_utils.py) since it only seemed to be this file that was changed in the PR and I'm getting the error: InvalidSchema: No connection adapters were found for 's3://bucket/folder/'. I'...
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
92
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1877988576889038, -0.9315599799156189, -0.7374240756034851, 1.5009020566940308, -0.11799561232328415, -1.2380927801132202, 0.1705128252506256, -1.006676197052002, 1.680005431175232, -0.7602040767669678, 0.33744391798973083, -1.6719021797180176, -0.020392054691910744, -0.5125157833099365...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
The current implementation depends on gcsfs/s3fs being able to authenticate through some other means e.g., environmental variables. For AWS, it looks like you can set `AWS_ACCESS_KEY_ID`, `AWS_SECRET_ACCESS_KEY`, and `AWS_SESSION_TOKEN` Note that while testing this just now, I did note a discrepancy between gcsfs an...
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
130
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1259617805480957, -0.8903080821037292, -0.7191109657287598, 1.5024460554122925, -0.09601156413555145, -1.2793055772781372, 0.1703445017337799, -1.0465589761734009, 1.6725752353668213, -0.7534469366073608, 0.30338990688323975, -1.6468414068222046, -0.00529266195371747, -0.590070068836212...
https://github.com/huggingface/datasets/issues/5281
Support cloud storage in load_dataset
> Note that while testing this just now, I did note a discrepancy between gcsfs and s3fs that we might want to address where gcsfs passes the timeout from storage_options [here](https://github.com/huggingface/datasets/blob/3e6269979fc80ae8939294d26298897f0db5b84d/src/datasets/utils/file_utils.py#L333) down into the aio...
Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`. This has...
400
109
Support cloud storage in load_dataset Would be nice to be able to do ```python load_dataset("s3://...") ``` or even ```python data_files=["gs://..."] storage_options = {...} load_dataset(..., data_files=data_files, storage_options=storage_options) ``` The idea would be to use `fsspec` as in `download_and_...
[ -1.1931078433990479, -0.8865266442298889, -0.7406773567199707, 1.4943337440490723, -0.12763677537441254, -1.2567174434661865, 0.1296108514070511, -1.0605189800262451, 1.6105358600616455, -0.7359844446182251, 0.3050369322299957, -1.6380168199539185, -0.007218174636363983, -0.578545749187469...
https://github.com/huggingface/datasets/issues/5280
Import error
Hi, 3.8.13 Get Outlook for Android<https://aka.ms/AAb9ysg> ________________________________ From: Quentin Lhoest ***@***.***> Sent: Tuesday, November 22, 2022 2:37:02 PM To: huggingface/datasets ***@***.***> Cc: feketedavid1012 ***@***.***>; Author ***@***.***> Subject: Re: [huggingface/datasets] Import error (Issue ...
https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28 Hy, I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message wrong)
401
69
Import error https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28 Hy, I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message w...
[ -1.2245451211929321, -0.9519438147544861, -0.5376725792884827, 1.4114515781402588, -0.18234844505786896, -1.2721543312072754, 0.1501534879207611, -1.0091004371643066, 1.595414638519287, -0.5358529090881348, 0.1921544373035431, -1.6243865489959717, -0.10301874577999115, -0.5325057506561279,...
https://github.com/huggingface/datasets/issues/5280
Import error
Then it should work as expected if you use the same python when using `datasets` Please make sure you're running your code in the right environment
https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28 Hy, I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message wrong)
401
26
Import error https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28 Hy, I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message w...
[ -1.2907732725143433, -0.8051122426986694, -0.8636984825134277, 1.4544912576675415, -0.15099065005779266, -1.2534384727478027, 0.03319337218999863, -1.0669690370559692, 1.4942463636398315, -0.7183879017829895, 0.19982896745204926, -1.6829895973205566, -0.06631534546613693, -0.39326384663581...
https://github.com/huggingface/datasets/issues/5280
Import error
It's the right environment. But in if statement I have "3.8.13" < 3.7 And in the error message is Python>=3.7 which is true in my case (3.8.13 is greater then 3.7), so I don't understand my python should be below the 3.7 which case the if statement is right, but the message is wrong, or above 3.7 which case if statemen...
https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28 Hy, I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message wrong)
401
147
Import error https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28 Hy, I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message w...
[ -1.2348964214324951, -0.9628428816795349, -0.5841107368469238, 1.4090332984924316, -0.2059527337551117, -1.2893599271774292, 0.14153769612312317, -1.034083604812622, 1.5843340158462524, -0.6413482427597046, 0.1616053581237793, -1.6310269832611084, -0.13474372029304504, -0.5262941718101501,...
https://github.com/huggingface/datasets/issues/5278
load_dataset does not read jsonl metadata file properly
Can you try to remove "drop_labels=false" ? It may force the loader to infer the labels instead of reading the metadata
### Describe the bug Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column. It just spits out "image" and "label" as features. B...
402
21
load_dataset does not read jsonl metadata file properly ### Describe the bug Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column...
[ -1.1722538471221924, -0.9874001741409302, -0.759548544883728, 1.489896535873413, -0.1571158766746521, -1.2294549942016602, 0.1569814831018448, -1.0602964162826538, 1.5801401138305664, -0.8204648494720459, 0.28715646266937256, -1.6882760524749756, -0.015584716573357582, -0.5398836731910706,...
https://github.com/huggingface/datasets/issues/5278
load_dataset does not read jsonl metadata file properly
Okay, now it seems to actually load the metadata and create the train_split, but it still says only returns "image" and "label", which is always 0 since all images are from same folder
### Describe the bug Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column. It just spits out "image" and "label" as features. B...
402
33
load_dataset does not read jsonl metadata file properly ### Describe the bug Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column...
[ -1.1759647130966187, -0.9833332896232605, -0.7441419959068298, 1.4840329885482788, -0.15907064080238342, -1.2307698726654053, 0.1437048763036728, -1.065576195716858, 1.5801584720611572, -0.8430256247520447, 0.2931825816631317, -1.6919947862625122, -0.019899562001228333, -0.5299819707870483...
https://github.com/huggingface/datasets/issues/5278
load_dataset does not read jsonl metadata file properly
> Can you try updating `datasets` ? Metadata support was added in `datasets` 2.4 Update: This was the issue.
### Describe the bug Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column. It just spits out "image" and "label" as features. B...
402
19
load_dataset does not read jsonl metadata file properly ### Describe the bug Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column...
[ -1.1760004758834839, -0.9716751575469971, -0.7537791728973389, 1.4831427335739136, -0.15419664978981018, -1.2533154487609863, 0.1560819000005722, -1.051365613937378, 1.6201120615005493, -0.7968655824661255, 0.27176833152770996, -1.7120777368545532, -0.039096683263778687, -0.544991195201873...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
Sounds like one of the file is not a valid one, can you make sure you uploaded valid mp3 files ?
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
21
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2023125886917114, -1.0082038640975952, -0.7943263053894043, 1.4618934392929077, -0.19647520780563354, -1.2047638893127441, 0.1335814893245697, -1.0326474905014038, 1.6041899919509888, -0.8130604028701782, 0.2604953646659851, -1.6522417068481445, 0.10070887953042984, -0.5118575096130371,...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
Well I just sharded the original commonVoice dataset and pushed a small chunk of it in a private rep What did go wrong? Holen Sie sich Outlook für iOS<https://aka.ms/o0ukef> ________________________________ Von: Quentin Lhoest ***@***.***> Gesendet: Tuesday, November 22, 2022 3:03:40 PM An: huggingface/datasets ***@*...
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
112
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2165149450302124, -1.038955569267273, -0.6129095554351807, 1.4359979629516602, -0.19989962875843048, -1.2486038208007812, 0.18720325827598572, -1.0107054710388184, 1.6350244283676147, -0.6475505232810974, 0.18775707483291626, -1.6260625123977661, -0.013724486343562603, -0.56718891859054...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
It should be all good then ! Could you share a link to your repository for me to investigate what went wrong ?
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
23
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2314778566360474, -0.997174084186554, -0.7661173939704895, 1.4328819513320923, -0.18563005328178406, -1.1878058910369873, 0.1385161578655243, -1.0476622581481934, 1.5914175510406494, -0.8040013313293457, 0.24579352140426636, -1.6527099609375, 0.09208837896585464, -0.5104434490203857, ...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
https://huggingface.co/datasets/DTU54DL/common-voice-test16k Am Di., 22. Nov. 2022 um 16:43 Uhr schrieb Quentin Lhoest < ***@***.***>: > It should be all good then ! > Could you share a link to your repository for me to investigate what went > wrong ? > > — > Reply to this email directly, view it on GitHub > <https:/...
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
76
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.1809501647949219, -0.9750272035598755, -0.6727975010871887, 1.3768119812011719, -0.1401805430650711, -1.2024718523025513, 0.16715672612190247, -1.0679795742034912, 1.6264891624450684, -0.629643976688385, 0.22449806332588196, -1.691903829574585, -0.04451072961091995, -0.5420190095901489,...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
I see ! This is a bug with MP3 files. When we store audio data in parquet, we store the bytes and the file name. From the file name extension we know if it's a WAV, an MP3 or else. But here it looks like the paths are all None. It looks like it comes from here: https://github.com/huggingface/datasets/blob/7fee...
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
73
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2237224578857422, -0.9301946759223938, -0.7563335299491882, 1.41595458984375, -0.14739854633808136, -1.2214552164077759, 0.12228212505578995, -1.0183627605438232, 1.614269733428955, -0.7715045809745789, 0.2577328085899353, -1.6719316244125366, 0.040651775896549225, -0.49856555461883545,...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
@lhoestq I remember we wanted to avoid storing redundant data but maybe it's not that crucial indeed to store one more string value. Or we can store paths only for mp3s, considering that for other formats we don't have such a problem with reading from bytes without format specified.
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
49
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.265916347503662, -0.9846614003181458, -0.7658290266990662, 1.4486602544784546, -0.16831177473068237, -1.2107670307159424, 0.14646321535110474, -1.081231713294983, 1.5822337865829468, -0.7880792617797852, 0.27461543679237366, -1.6464152336120605, 0.07322409749031067, -0.4983493387699127,...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
thanks for the help! can I do anything on my side? we are doing a DL project and we need the data really quick. thanks bogdan > Message ID: ***@***.***> >
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
31
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2627248764038086, -0.9977168440818787, -0.7200328707695007, 1.393822431564331, -0.23218956589698792, -1.2249743938446045, 0.15750783681869507, -1.0767273902893066, 1.6008732318878174, -0.7729817032814026, 0.20860140025615692, -1.655421257019043, 0.019205667078495026, -0.4889303743839264...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
I opened a pull requests here: https://github.com/huggingface/datasets/pull/5285, we'll do a new release soon with this fix. Otherwise if you're really in a hurry you can install `datasets` from this PR
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
30
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2496193647384644, -0.9765840768814087, -0.786060094833374, 1.4415929317474365, -0.17721234261989594, -1.2027170658111572, 0.13879597187042236, -1.0286753177642822, 1.6046327352523804, -0.7760149240493774, 0.25483062863349915, -1.6858975887298584, 0.014647134579718113, -0.526800751686096...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
any idea on what's going wrong here? thanks Am So., 27. Nov. 2022 um 13:53 Uhr schrieb Bogdan Capsa < ***@***.***>: > [image: image.png] > >> Message ID: ***@***.***> >> >
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
31
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.236569881439209, -1.0269745588302612, -0.6960256695747375, 1.3576411008834839, -0.2544994354248047, -1.2162744998931885, 0.20918342471122742, -1.0425646305084229, 1.6523768901824951, -0.7125768661499023, 0.2279398888349533, -1.6549357175827026, -0.024651262909173965, -0.496742844581604,...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
I have attached this screenshot above . can u pls help? So can not pip from pull request ![image](https://user-images.githubusercontent.com/48530104/204354027-6173e6d1-e3d4-4085-a363-e924cfe1a7f4.png)
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
19
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2575498819351196, -0.9444321990013123, -0.7481180429458618, 1.4291225671768188, -0.15634191036224365, -1.1953096389770508, 0.13540786504745483, -1.0080151557922363, 1.5977168083190918, -0.7740605473518372, 0.29220643639564514, -1.628374457359314, 0.033249884843826294, -0.506635308265686...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
The pull request has been merged on `main`. You can install `datasets` from `main` using ``` pip install git+https://github.com/huggingface/datasets.git ```
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
20
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.213504433631897, -0.9395443797111511, -0.7641224265098572, 1.4931994676589966, -0.1710723638534546, -1.2629755735397339, 0.16276653110980988, -1.0390715599060059, 1.6553657054901123, -0.7732201814651489, 0.27239540219306946, -1.6769499778747559, 0.03538985922932625, -0.5351313948631287,...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
I've tried to load this dataset DTU54DL/common-voice-test16k, but am getting the same error. So the bug fix will fix only if I upload a new dataset, or also loading previously uploaded datasets? thanks Am Mo., 28. Nov. 2022 um 19:51 Uhr schrieb Quentin Lhoest < ***@***.***>: > The pull request has been merged on ma...
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
103
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.227881669998169, -1.0120747089385986, -0.6498963832855225, 1.3768483400344849, -0.21116602420806885, -1.2106375694274902, 0.16845373809337616, -1.0729312896728516, 1.633545994758606, -0.6536146998405457, 0.20996902883052826, -1.6974087953567505, -0.05560732260346413, -0.525266706943512,...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
> So the bug fix will fix only if I upload a new dataset, or also loading previously uploaded datasets? You have to reupload the dataset, sorry for the inconvenience
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
30
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2553917169570923, -0.9863103032112122, -0.7905780673027039, 1.447359561920166, -0.21210473775863647, -1.207574725151062, 0.12582971155643463, -1.0731642246246338, 1.5865862369537354, -0.8383920192718506, 0.26923415064811707, -1.6797409057617188, 0.05695488676428795, -0.5417887568473816,...
https://github.com/huggingface/datasets/issues/5276
Bug in downloading common_voice data and snall chunk of it to one's own hub
thank you so much for the help! works like a charm! Am Di., 29. Nov. 2022 um 12:15 Uhr schrieb Quentin Lhoest < ***@***.***>: > So the bug fix will fix only if I upload a new dataset, or also loading > previously uploaded datasets? > > You have to reupload the dataset, sorry for the inconvenience > > — > Reply to thi...
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4...
403
92
Bug in downloading common_voice data and snall chunk of it to one's own hub ### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image...
[ -1.2216267585754395, -1.0219714641571045, -0.6812902688980103, 1.4104489088058472, -0.1973455250263214, -1.2055637836456299, 0.16963109374046326, -1.08225417137146, 1.6632575988769531, -0.6676566004753113, 0.23157230019569397, -1.6840180158615112, -0.03684938699007034, -0.5427006483078003,...
https://github.com/huggingface/datasets/issues/5275
YAML integer keys are not preserved Hub server-side
@huggingface/datasets if you agree, I can make the bulk edit on the Hub to fix integer keys into strings.
After an internal discussion (https://github.com/huggingface/moon-landing/issues/4563): - YAML integer keys are not preserved server-side: they are transformed to strings - See for example this Hub PR: https://huggingface.co/datasets/acronym_identification/discussions/1/files - Original: ```yaml ...
404
19
YAML integer keys are not preserved Hub server-side After an internal discussion (https://github.com/huggingface/moon-landing/issues/4563): - YAML integer keys are not preserved server-side: they are transformed to strings - See for example this Hub PR: https://huggingface.co/datasets/acronym_identification/discu...
[ -1.0408730506896973, -0.7701073288917542, -0.7022444009780884, 1.6244213581085205, -0.0046040830202400684, -1.3531185388565063, 0.038567349314689636, -0.9089091420173645, 1.5970324277877808, -0.737040638923645, 0.4560614228248596, -1.7285661697387695, -0.056928135454654694, -0.701209604740...