html_url stringlengths 51 51 | title stringlengths 6 280 | comments stringlengths 67 24.7k | body stringlengths 51 36.2k | __index_level_0__ int64 1 1.17k | comment_length int64 16 1.45k | text stringlengths 190 38.3k | embeddings list |
|---|---|---|---|---|---|---|---|
https://github.com/huggingface/datasets/issues/5354 | Consider using "Sequence" instead of "List" | Hi! Linking a comment to provide more info on the issue: https://stackoverflow.com/a/39458225. This means we should replace all (most of) the occurrences of `List` with `Sequence` in function signatures.
@tranhd95 Would you be interested in submitting a PR? | ### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below.
**How to reproduce**
```py
... | 374 | 38 | Consider using "Sequence" instead of "List"
### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors... | [
-1.1331499814987183,
-0.8368464112281799,
-0.7626137733459473,
1.4945868253707886,
-0.16346971690654755,
-1.2774741649627686,
0.18324048817157745,
-1.0449070930480957,
1.7204946279525757,
-0.7336402535438538,
0.3742302656173706,
-1.6610338687896729,
-0.021034322679042816,
-0.64282089471817... |
https://github.com/huggingface/datasets/issues/5354 | Consider using "Sequence" instead of "List" | Hi all! I tried to reproduce this issue and didn't work for me. Also in your example i noticed that the variables have different names: `list_of_filenames` and `list_of_files`, could this be related to that?
```python
#I found random data in parquet format:
!wget "https://github.com/Teradata/kylo/raw/master/samples/... | ### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below.
**How to reproduce**
```py
... | 374 | 121 | Consider using "Sequence" instead of "List"
### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors... | [
-1.1363391876220703,
-0.8300718069076538,
-0.7608848214149475,
1.5081171989440918,
-0.14507760107517242,
-1.3009257316589355,
0.16242757439613342,
-1.038397192955017,
1.6914905309677124,
-0.746339738368988,
0.33476725220680237,
-1.6492044925689697,
-0.0681803748011589,
-0.6372345685958862,... |
https://github.com/huggingface/datasets/issues/5354 | Consider using "Sequence" instead of "List" | @dantema There is indeed a typo in variable names. Nevertheless, I'm sorry if I was not clear but the output is from `mypy` type checker. You can run the code snippet without issues. The problem is with the type checking. | ### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below.
**How to reproduce**
```py
... | 374 | 40 | Consider using "Sequence" instead of "List"
### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors... | [
-1.1417784690856934,
-0.8362669944763184,
-0.7446849346160889,
1.523167371749878,
-0.16777461767196655,
-1.2624818086624146,
0.1661750078201294,
-1.0644620656967163,
1.730597972869873,
-0.7514219284057617,
0.38889002799987793,
-1.6621534824371338,
-0.03524775430560112,
-0.6129856705665588,... |
https://github.com/huggingface/datasets/issues/5354 | Consider using "Sequence" instead of "List" | However, I found out that the type annotation is actually misleading. The [`from_parquet`](https://github.com/huggingface/datasets/blob/5ef1ab1cc06c2b7a574bf2df454cd9fcb071ccb2/src/datasets/arrow_dataset.py#L1039) method should also accept list of [`PathLike`](https://github.com/huggingface/datasets/blob/main/src/datas... | ### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below.
**How to reproduce**
```py
... | 374 | 405 | Consider using "Sequence" instead of "List"
### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors... | [
-1.1314377784729004,
-0.8188692331314087,
-0.7313487529754639,
1.4617829322814941,
-0.12871532142162323,
-1.2888085842132568,
0.1325380504131317,
-1.0625556707382202,
1.6795439720153809,
-0.7191594839096069,
0.34144148230552673,
-1.6790275573730469,
-0.08184440433979034,
-0.619117379188537... |
https://github.com/huggingface/datasets/issues/5354 | Consider using "Sequence" instead of "List" | @avinashsai Hi, I've assigned you the issue.
@tranhd95 Yes, feel free to report this in a new issue. | ### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below.
**How to reproduce**
```py
... | 374 | 18 | Consider using "Sequence" instead of "List"
### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors... | [
-1.1352909803390503,
-0.841376006603241,
-0.7508730292320251,
1.5250927209854126,
-0.15977616608142853,
-1.2457406520843506,
0.15540334582328796,
-1.0274401903152466,
1.7066551446914673,
-0.7475284934043884,
0.37051212787628174,
-1.6746240854263306,
-0.04259897395968437,
-0.622329950332641... |
https://github.com/huggingface/datasets/issues/5354 | Consider using "Sequence" instead of "List" | @avinashsai Are you still working on this? If not I would like to give it a try. | ### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below.
**How to reproduce**
```py
... | 374 | 17 | Consider using "Sequence" instead of "List"
### Feature request
Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors... | [
-1.140164852142334,
-0.8359280824661255,
-0.7365889549255371,
1.528045892715454,
-0.14484205842018127,
-1.2515233755111694,
0.16015976667404175,
-1.0420674085617065,
1.7080374956130981,
-0.7429264187812805,
0.3787233829498291,
-1.659406065940857,
-0.0451342798769474,
-0.6294645667076111,
... |
https://github.com/huggingface/datasets/issues/5352 | __init__() got an unexpected keyword argument 'input_size' | Hi @J-shel, thanks for reporting.
I think the issue comes from your call to `load_dataset`. As first argument, you should pass:
- either the name of your dataset ("mrf") if this is already published on the Hub
- or the path to the loading script of your dataset ("path/to/your/local/mrf.py"). | ### Describe the bug
I try to define a custom configuration with a input_size attribute following the instructions by "Specifying several dataset configurations" in https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html
But when I load the dataset, I got an error "__init__() got an unexpected keyword argument... | 376 | 49 | __init__() got an unexpected keyword argument 'input_size'
### Describe the bug
I try to define a custom configuration with a input_size attribute following the instructions by "Specifying several dataset configurations" in https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html
But when I load the dataset, ... | [
-1.3179267644882202,
-0.8770233988761902,
-0.6781187057495117,
1.4217880964279175,
-0.23399394750595093,
-1.1265684366226196,
0.21573123335838318,
-1.0965675115585327,
1.5839473009109497,
-0.7800825238227844,
0.32759523391723633,
-1.5724815130233765,
-0.059767208993434906,
-0.5339155197143... |
https://github.com/huggingface/datasets/issues/5352 | __init__() got an unexpected keyword argument 'input_size' | Hi, following your suggestion, I changed my call to load_dataset. Below is the latest:
reader = load_dataset('data/mrf.py',"default", input_size=1024, split=split, streaming=True, keep_in_memory=None)
However, I still got the same error.
I have one question that is if I only define input_size=2048 in BUILDER_CONFIGS... | ### Describe the bug
I try to define a custom configuration with a input_size attribute following the instructions by "Specifying several dataset configurations" in https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html
But when I load the dataset, I got an error "__init__() got an unexpected keyword argument... | 376 | 65 | __init__() got an unexpected keyword argument 'input_size'
### Describe the bug
I try to define a custom configuration with a input_size attribute following the instructions by "Specifying several dataset configurations" in https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html
But when I load the dataset, ... | [
-1.319419503211975,
-0.8907701373100281,
-0.6713951230049133,
1.427407145500183,
-0.23848478496074677,
-1.1148940324783325,
0.22346144914627075,
-1.092186450958252,
1.556174397468567,
-0.7811620235443115,
0.31490352749824524,
-1.5549242496490479,
-0.07593421638011932,
-0.5233897566795349,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | Hi! `DatasetBuilder` is a parent class for concrete builders: `GeneratorBasedBuilder`, `ArrowBasedBuilder` and `BeamBasedBuilder`. When writing a builder script, these classes are the ones you should inherit from. And since all of them implement `_prepare_split`, you only have to implement the three methods mentioned a... | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 44 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | Thanks so much @mariosasko for the fast response! I've been referencing [this page in the docs](https://huggingface.co/docs/datasets/v2.4.0/en/about_dataset_load) because it it pretty comprehensive in terms of what we have to do and I figured since we subclass the `BuilderConfig` the same pattern would hold, but I've a... | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 63 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | cc @stevhliu who may have some ideas on how to improve this part of the docs. | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 16 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | one more question for my understanding @mariosasko. the requirement of a loading script has always seemed counterintuitive to me. if i have to provide a script with every dataset, what is the point of using `datasets` if we're doing all the work of loading it, I can just do that in my code and skip the datasets integra... | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 274 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | > the requirement of a loading script has always seemed counterintuitive to me
This is a requirement only for datasets not stored in standard formats such as CSV, JSON, SQL, Parquet, ImageFolder, etc.
> if i have to provide a script with every dataset, what is the point of using datasets if we're doing all the... | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 163 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | ok, that makes sense. thank you @mariosasko. I realized i'd never looked on the hub at any of the files associated with any datasets. just did that now and it appears that i'll need to have a script regardless _but_ that will just contain my custom config and builder classes, so without realizing it I was already makin... | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 117 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | `download_and_prepare` downloads and prepares the arrow files. You need to call `as_dataset` on the builder to get the dataset. | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 19 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | ok, I think I was assigning the output of `builder.download_and_prepare` but it's an inplace op, so that explains the `NoneType` i was getting back. Now I'm getting:
```
ArrowInvalid Traceback (most recent call last)
<ipython-input-7-3ed50fb87c70> in <module>
----> 1 ds = dataset_buil... | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 486 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | looks like my arrow files are all empty @mariosasko

i also see the `incomplete_info.lock` file a level up too. seems like the data isn't being persisted to disk when I call `download_and_prepare`. is ... | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 44 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | quick update @mariosasko. i got it working! i had to downgrade to `datasets==2.4.0`. testing other versions now and will let you know the results. | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 24 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5351 | Do we need to implement `_prepare_split`? | I've tested with every version of `datasets>2.4.0` and i get the same error with all of them. | ### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstract (as are the others we are required to im... | 377 | 17 | Do we need to implement `_prepare_split`?
### Describe the bug
I'm not sure this is a bug or if it's just missing in the documentation, or i'm not doing something correctly, but I'm subclassing `DatasetBuilder` and getting the following error because on the `DatasetBuilder` class the `_prepare_split` method is abstr... | [
-1.1564629077911377,
-0.9498602747917175,
-0.663374125957489,
1.4442986249923706,
-0.10583754628896713,
-1.330580234527588,
0.14666317403316498,
-1.096006155014038,
1.5908504724502563,
-0.7755786180496216,
0.2561314105987549,
-1.64324951171875,
0.0009109014645218849,
-0.5524017810821533,
... |
https://github.com/huggingface/datasets/issues/5348 | The data downloaded in the download folder of the cache does not respect `umask` | note, that `datasets` already did some of that umask fixing in the past and also at the hub - the recent work on the hub about the same: https://github.com/huggingface/huggingface_hub/pull/1220
Also I noticed that each file has a .json counterpart and the latter always has the correct perms:
```
-rw------- 1 uue... | ### Describe the bug
For a project on a cluster we are several users to share the same cache for the datasets library. And we have a problem with the permissions on the data downloaded in the cache.
Indeed, it seems that the data is downloaded by giving read and write permissions only to the user launching the com... | 378 | 83 | The data downloaded in the download folder of the cache does not respect `umask`
### Describe the bug
For a project on a cluster we are several users to share the same cache for the datasets library. And we have a problem with the permissions on the data downloaded in the cache.
Indeed, it seems that the data is... | [
-1.1730643510818481,
-0.8724946975708008,
-0.6110600233078003,
1.3566467761993408,
-0.0034409910440444946,
-1.4475382566452026,
0.09687133878469467,
-0.9708036780357361,
1.5548968315124512,
-0.8790223598480225,
0.3033035695552826,
-1.7146375179290771,
0.0685807541012764,
-0.622688114643096... |
https://github.com/huggingface/datasets/issues/5345 | Wrong dtype for array in audio features | After some more investigation, this is due to [this line of code](https://github.com/huggingface/datasets/blob/main/src/datasets/features/audio.py#L279). The function `sf.read(file)` should be updated to `sf.read(file, dtype="float32")`
Indeed, the default value in soundfile is `float64` ([see here](https://pysoundf... | ### Describe the bug
When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datasets cannot be merged.
### Steps to repro... | 380 | 31 | Wrong dtype for array in audio features
### Describe the bug
When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datase... | [
-1.3297818899154663,
-1.0398335456848145,
-0.7567360997200012,
1.6267650127410889,
-0.26365041732788086,
-1.0593035221099854,
0.12767237424850464,
-1.0891356468200684,
1.538980484008789,
-0.7270705103874207,
0.23913443088531494,
-1.5932620763778687,
0.00082396250218153,
-0.7497702240943909... |
https://github.com/huggingface/datasets/issues/5345 | Wrong dtype for array in audio features | @qmeeus I agree, decoding of different audio formats should return the same dtypes indeed!
But note that here you are concatenating datasets with different sampling rates: 48000 for CommonVoice and 16000 for Voxpopuli. So you should cast them to the same sampling rate value before interleaving, for example:
```
cv... | ### Describe the bug
When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datasets cannot be merged.
### Steps to repro... | 380 | 115 | Wrong dtype for array in audio features
### Describe the bug
When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datase... | [
-1.3297818899154663,
-1.0398335456848145,
-0.7567360997200012,
1.6267650127410889,
-0.26365041732788086,
-1.0593035221099854,
0.12767237424850464,
-1.0891356468200684,
1.538980484008789,
-0.7270705103874207,
0.23913443088531494,
-1.5932620763778687,
0.00082396250218153,
-0.7497702240943909... |
https://github.com/huggingface/datasets/issues/5345 | Wrong dtype for array in audio features | Hi Polina,
Sorry for the late answer
It is possible that the issue was due to a bug that is now fixed. I installed an editable version of datasets from github, but I don't recall whether I had updated it at the time of the issue. My research led me to other directions so I did not follow through on the interleave dat... | ### Describe the bug
When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datasets cannot be merged.
### Steps to repro... | 380 | 63 | Wrong dtype for array in audio features
### Describe the bug
When concatenating/interleaving different datasets, I stumble into an error because the features can't be aligned. After some investigation, I understood that the audio arrays had different dtypes, namely `float32` and `float64`. Consequently, the datase... | [
-1.3297818899154663,
-1.0398335456848145,
-0.7567360997200012,
1.6267650127410889,
-0.26365041732788086,
-1.0593035221099854,
0.12767237424850464,
-1.0891356468200684,
1.538980484008789,
-0.7270705103874207,
0.23913443088531494,
-1.5932620763778687,
0.00082396250218153,
-0.7497702240943909... |
https://github.com/huggingface/datasets/issues/5342 | Emotion dataset cannot be downloaded | Hi @cbarond there's already an open issue at https://github.com/dair-ai/emotion_dataset/issues/5, as the data seems to be missing now, so check that issue instead 👍🏻 | ### Describe the bug
The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`.
It was working yesterday (December 7, 2022), but stopped working today (December 8, 2022).
### Steps to reproduce the bug
... | 381 | 23 | Emotion dataset cannot be downloaded
### Describe the bug
The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`.
It was working yesterday (December 7, 2022), but stopped working today (December 8, 202... | [
-1.0933345556259155,
-0.8987973928451538,
-0.8178911209106445,
1.440356731414795,
-0.18660634756088257,
-1.2340970039367676,
0.13042962551116943,
-0.977516770362854,
1.568128228187561,
-0.7103350162506104,
0.25876522064208984,
-1.6854135990142822,
-0.0371609665453434,
-0.6266356706619263,
... |
https://github.com/huggingface/datasets/issues/5342 | Emotion dataset cannot be downloaded | Thanks @cbarond for reporting and @alvarobartt for pointing to the issue we opened in the author's repo.
Indeed, this issue was first raised in the "emotion" dataset Community tab: https://huggingface.co/datasets/emotion/discussions/3
I'm closing this issue and leave the issue above for the subsequent updates.
... | ### Describe the bug
The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`.
It was working yesterday (December 7, 2022), but stopped working today (December 8, 2022).
### Steps to reproduce the bug
... | 381 | 46 | Emotion dataset cannot be downloaded
### Describe the bug
The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`.
It was working yesterday (December 7, 2022), but stopped working today (December 8, 202... | [
-1.1503911018371582,
-0.9479205012321472,
-0.8195348978042603,
1.4430569410324097,
-0.15113148093223572,
-1.2766895294189453,
0.11261900514364243,
-0.955824613571167,
1.629323124885559,
-0.6967946290969849,
0.22675691545009613,
-1.6891469955444336,
-0.0682714432477951,
-0.551869809627533,
... |
https://github.com/huggingface/datasets/issues/5342 | Emotion dataset cannot be downloaded | > try using "SetFit/emotion" instead
I' replaced "emotion" with "SetFit/Emotion", but the code is getting stuck at
`emotions = load_dataset("SetFit/emotion")`
I pause execution using the debugger, and it takes me to filelock.py:226
`with self._thread_lock:`
Do you know a way to get past this issue? | ### Describe the bug
The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`.
It was working yesterday (December 7, 2022), but stopped working today (December 8, 2022).
### Steps to reproduce the bug
... | 381 | 44 | Emotion dataset cannot be downloaded
### Describe the bug
The emotion dataset gives a FileNotFoundError. The full error is: `FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1`.
It was working yesterday (December 7, 2022), but stopped working today (December 8, 202... | [
-1.1264564990997314,
-0.8865939974784851,
-0.8132458329200745,
1.4566452503204346,
-0.19419091939926147,
-1.2872220277786255,
0.17019471526145935,
-1.0140031576156616,
1.598790168762207,
-0.740466296672821,
0.2705131471157074,
-1.6355057954788208,
0.005785159766674042,
-0.5599542260169983,... |
https://github.com/huggingface/datasets/issues/5338 | `map()` stops every 1000 steps | Hi !
> It starts using all the cores (I am not sure why because I did not pass num_proc)
The tokenizer uses Rust code that is multithreaded. And maybe the `feature_extractor` might run some things in parallel as well - but I'm not super familiar with its internals.
> then progress bar stops at every 1k steps. ... | ### Describe the bug
I am passing the following `prepare_dataset` function to `Dataset.map` (code is inspired from [here](https://github.com/huggingface/community-events/blob/main/whisper-fine-tuning-event/run_speech_recognition_seq2seq_streaming.py#L454))
```python3
def prepare_dataset(batch):
# load and res... | 382 | 108 | `map()` stops every 1000 steps
### Describe the bug
I am passing the following `prepare_dataset` function to `Dataset.map` (code is inspired from [here](https://github.com/huggingface/community-events/blob/main/whisper-fine-tuning-event/run_speech_recognition_seq2seq_streaming.py#L454))
```python3
def prepare_da... | [
-1.2085660696029663,
-0.8807992935180664,
-0.6695201992988586,
1.4731371402740479,
-0.08829861879348755,
-1.3140623569488525,
0.18715448677539825,
-1.0919111967086792,
1.6589032411575317,
-0.7869752645492554,
0.3509739637374878,
-1.6116012334823608,
0.056128792464733124,
-0.595148921012878... |
https://github.com/huggingface/datasets/issues/5338 | `map()` stops every 1000 steps | Hi @lhoestq
Thanks for the explanation! it was so helpful! Let me check why `feature_extractor` is running on multiple cpus. | ### Describe the bug
I am passing the following `prepare_dataset` function to `Dataset.map` (code is inspired from [here](https://github.com/huggingface/community-events/blob/main/whisper-fine-tuning-event/run_speech_recognition_seq2seq_streaming.py#L454))
```python3
def prepare_dataset(batch):
# load and res... | 382 | 20 | `map()` stops every 1000 steps
### Describe the bug
I am passing the following `prepare_dataset` function to `Dataset.map` (code is inspired from [here](https://github.com/huggingface/community-events/blob/main/whisper-fine-tuning-event/run_speech_recognition_seq2seq_streaming.py#L454))
```python3
def prepare_da... | [
-1.2085660696029663,
-0.8807992935180664,
-0.6695201992988586,
1.4731371402740479,
-0.08829861879348755,
-1.3140623569488525,
0.18715448677539825,
-1.0919111967086792,
1.6589032411575317,
-0.7869752645492554,
0.3509739637374878,
-1.6116012334823608,
0.056128792464733124,
-0.595148921012878... |
https://github.com/huggingface/datasets/issues/5337 | Support webdataset format | I like the idea of having `webdataset` as an optional dependency to ensure our loader generates web datasets the same way as the main project. | Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234.
In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local directory, or from a dataset o... | 383 | 25 | Support webdataset format
Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234.
In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local di... | [
-1.1585583686828613,
-0.9291451573371887,
-0.9457713961601257,
1.511222243309021,
-0.28723976016044617,
-1.4082257747650146,
0.09250763058662415,
-1.1086831092834473,
1.6967360973358154,
-0.8459455370903015,
0.39014437794685364,
-1.738334059715271,
0.11118140071630478,
-0.4907057583332062,... |
https://github.com/huggingface/datasets/issues/5337 | Support webdataset format | Webdataset is the one of the most popular dataset formats for large scale computer vision tasks. Upvote for this issue. | Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234.
In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local directory, or from a dataset o... | 383 | 20 | Support webdataset format
Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234.
In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local di... | [
-1.1678622961044312,
-0.9333143830299377,
-0.9526174068450928,
1.5264627933502197,
-0.2599416971206665,
-1.4019267559051514,
0.1297958940267563,
-1.1392103433609009,
1.6976109743118286,
-0.7957373261451721,
0.3557114601135254,
-1.7333347797393799,
0.1317697912454605,
-0.48345568776130676,
... |
https://github.com/huggingface/datasets/issues/5337 | Support webdataset format | We haven't had the bandwidth to implement it so far, but if someone wants to give it a shot please don't hesitate ^^ | Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234.
In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local directory, or from a dataset o... | 383 | 23 | Support webdataset format
Webdataset is an efficient format for iterable datasets. It would be nice to support it in `datasets`, as discussed in https://github.com/rom1504/img2dataset/issues/234.
In particular it would be awesome to be able to load one using `load_dataset` in streaming mode (either from a local di... | [
-1.1931383609771729,
-0.9310227632522583,
-0.9216461777687073,
1.5134222507476807,
-0.2293902337551117,
-1.3462190628051758,
0.0987570658326149,
-1.08004629611969,
1.7124196290969849,
-0.8611600399017334,
0.37386053800582886,
-1.7161235809326172,
0.1091696247458458,
-0.5106855630874634,
... |
https://github.com/huggingface/datasets/issues/5332 | Passing numpy array to ClassLabel names causes ValueError | Should `datasets` allow `ClassLabel` input parameter to be an `np.array` even though internally we need to cast it to a Python list? @lhoestq @mariosasko | ### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
If I define my classes as:
```
my_classes = np.array(['on... | 384 | 24 | Passing numpy array to ClassLabel names causes ValueError
### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
... | [
-1.2423409223556519,
-0.9427435398101807,
-0.6524389982223511,
1.6519964933395386,
-0.1358901411294937,
-1.1631975173950195,
0.15079466998577118,
-0.955312967300415,
1.5839974880218506,
-0.7523292303085327,
0.3405283987522125,
-1.6442815065383911,
0.02535400539636612,
-0.6284711360931396,
... |
https://github.com/huggingface/datasets/issues/5332 | Passing numpy array to ClassLabel names causes ValueError | Hi! No, I don't think so. The `names` parameter is [annotated](https://github.com/huggingface/datasets/blob/582236640b9109988e5f7a16a8353696ffa09a16/src/datasets/features/features.py#L892) as `List[str]` (**NumPy arrays are not lists**), and considering that type checking is not a common practice in Python, I think we ... | ### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
If I define my classes as:
```
my_classes = np.array(['on... | 384 | 38 | Passing numpy array to ClassLabel names causes ValueError
### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
... | [
-1.2423409223556519,
-0.9427435398101807,
-0.6524389982223511,
1.6519964933395386,
-0.1358901411294937,
-1.1631975173950195,
0.15079466998577118,
-0.955312967300415,
1.5839974880218506,
-0.7523292303085327,
0.3405283987522125,
-1.6442815065383911,
0.02535400539636612,
-0.6284711360931396,
... |
https://github.com/huggingface/datasets/issues/5332 | Passing numpy array to ClassLabel names causes ValueError | I appreciate it is the wrong type, and that type checking is not common, but I think there's a few circumstances that make it a good idea from a usability perspective.
It's quite a difficult error to debug because it comes from a utility function (so it's not immediately obvious which parameter caused it). What make... | ### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
If I define my classes as:
```
my_classes = np.array(['on... | 384 | 234 | Passing numpy array to ClassLabel names causes ValueError
### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
... | [
-1.2423409223556519,
-0.9427435398101807,
-0.6524389982223511,
1.6519964933395386,
-0.1358901411294937,
-1.1631975173950195,
0.15079466998577118,
-0.955312967300415,
1.5839974880218506,
-0.7523292303085327,
0.3405283987522125,
-1.6442815065383911,
0.02535400539636612,
-0.6284711360931396,
... |
https://github.com/huggingface/datasets/issues/5332 | Passing numpy array to ClassLabel names causes ValueError | I agree that it's not easy to debug this issue, so perhaps we could add some basic type checking (e.g. `not isinstance(names, list)` -> error) to make debugging easier. Feel free to submit a PR.
> Additionally, passing in a set instead of the list (again, relatively easy because people may run set(classes) to genera... | ### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
If I define my classes as:
```
my_classes = np.array(['on... | 384 | 100 | Passing numpy array to ClassLabel names causes ValueError
### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
... | [
-1.2423409223556519,
-0.9427435398101807,
-0.6524389982223511,
1.6519964933395386,
-0.1358901411294937,
-1.1631975173950195,
0.15079466998577118,
-0.955312967300415,
1.5839974880218506,
-0.7523292303085327,
0.3405283987522125,
-1.6442815065383911,
0.02535400539636612,
-0.6284711360931396,
... |
https://github.com/huggingface/datasets/issues/5332 | Passing numpy array to ClassLabel names causes ValueError | What about checking for `Sequence` instead? I think users can pass a list or a tuple as well. | ### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
If I define my classes as:
```
my_classes = np.array(['on... | 384 | 18 | Passing numpy array to ClassLabel names causes ValueError
### Describe the bug
If a numpy array is passed to the names argument of ClassLabel, creating a dataset with those features causes an error.
### Steps to reproduce the bug
https://colab.research.google.com/drive/1cV_es1PWZiEuus17n-2C-w0KEoEZ68IX
TLDR:
... | [
-1.2423409223556519,
-0.9427435398101807,
-0.6524389982223511,
1.6519964933395386,
-0.1358901411294937,
-1.1631975173950195,
0.15079466998577118,
-0.955312967300415,
1.5839974880218506,
-0.7523292303085327,
0.3405283987522125,
-1.6442815065383911,
0.02535400539636612,
-0.6284711360931396,
... |
https://github.com/huggingface/datasets/issues/5325 | map(...batch_size=None) for IterableDataset | Hi! I agree it makes sense for `IterableDataset.map` to support the `batch_size=None` case. This should be super easy to fix. | ### Feature request
Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too.
### Motivation
Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple of reasons why this might be nice.
One is th... | 385 | 20 | map(...batch_size=None) for IterableDataset
### Feature request
Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too.
### Motivation
Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple o... | [
-1.2428652048110962,
-1.0249131917953491,
-0.705143392086029,
1.4132028818130493,
-0.15908421576023102,
-1.2323696613311768,
0.1825770139694214,
-1.0238515138626099,
1.763264536857605,
-0.8696856498718262,
0.3509354889392853,
-1.7236558198928833,
0.08649272471666336,
-0.44251522421836853,
... |
https://github.com/huggingface/datasets/issues/5325 | map(...batch_size=None) for IterableDataset | @mariosasko as this is something simple maybe I can include it as part of https://github.com/huggingface/datasets/pull/5311? Let me know :+1: | ### Feature request
Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too.
### Motivation
Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple of reasons why this might be nice.
One is th... | 385 | 19 | map(...batch_size=None) for IterableDataset
### Feature request
Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too.
### Motivation
Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple o... | [
-1.2332842350006104,
-1.0495638847351074,
-0.7255744934082031,
1.4376425743103027,
-0.13801564276218414,
-1.2274912595748901,
0.12961739301681519,
-0.9991413354873657,
1.7146615982055664,
-0.8204090595245361,
0.3283476233482361,
-1.7594983577728271,
0.03388161212205887,
-0.4589964747428894... |
https://github.com/huggingface/datasets/issues/5325 | map(...batch_size=None) for IterableDataset | Thanks again :)
> For practical usages, an alternative to this would be to convert from an iterable dataset to a map-style dataset, but it is not obvious how to do this.
This is interesting as well, if anyone wants to explore | ### Feature request
Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too.
### Motivation
Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple of reasons why this might be nice.
One is th... | 385 | 42 | map(...batch_size=None) for IterableDataset
### Feature request
Dataset.map(...) allows batch_size to be None. It would be nice if IterableDataset did too.
### Motivation
Although it may seem a bit of a spurious request given that `IterableDataset` is meant for larger than memory datasets, but there are a couple o... | [
-1.2331047058105469,
-1.01775324344635,
-0.6940887570381165,
1.4135165214538574,
-0.16815601289272308,
-1.214682936668396,
0.1424940675497055,
-1.0177501440048218,
1.7147804498672485,
-0.8525761365890503,
0.329569548368454,
-1.7453944683074951,
0.08093299716711044,
-0.48796477913856506,
... |
https://github.com/huggingface/datasets/issues/5324 | Fix docstrings and types in documentation that appears on the website | Ok, I believe we've cleaned up most of the old syntax we were using for the user-facing docs! There are still a couple of `:obj:`'s and `:class:` floating around in the docstrings we don't expose that I'll track down :) | While I was working on https://github.com/huggingface/datasets/pull/5313 I've noticed that we have a mess in how we annotate types and format args and return values in the code. And some of it is displayed in the [Reference section](https://huggingface.co/docs/datasets/package_reference/builder_classes) of the document... | 386 | 40 | Fix docstrings and types in documentation that appears on the website
While I was working on https://github.com/huggingface/datasets/pull/5313 I've noticed that we have a mess in how we annotate types and format args and return values in the code. And some of it is displayed in the [Reference section](https://hugging... | [
-1.3390648365020752,
-0.9905195236206055,
-0.8267769813537598,
1.5412458181381226,
-0.1037326380610466,
-1.3935750722885132,
-0.0008664289489388466,
-1.07442045211792,
1.6979824304580688,
-0.8807410597801208,
0.293878436088562,
-1.7097569704055786,
0.009175123646855354,
-0.607541561126709,... |
https://github.com/huggingface/datasets/issues/5317 | `ImageFolder` performs poorly with large datasets | Hi ! ImageFolder is made for small scale datasets indeed. For large scale image datasets you better group your images in TAR archives or Arrow/Parquet files. This is true not just for ImageFolder loading performance, but also because having millions of files is not ideal for your filesystem or when moving the data arou... | ### Describe the bug
While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images.
## Setup
* Nested directories (5 levels deep)
* 3M+ images
* 1 `metadata.jsonl` file
## Performance Degradation Point... | 388 | 186 | `ImageFolder` performs poorly with large datasets
### Describe the bug
While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images.
## Setup
* Nested directories (5 levels deep)
* 3M+ images
* 1 `metada... | [
-1.1987686157226562,
-0.9523312449455261,
-0.7291271686553955,
1.4155699014663696,
-0.10784557461738586,
-1.2535969018936157,
0.13453370332717896,
-1.0448516607284546,
1.6751487255096436,
-0.7176129817962646,
0.2097993642091751,
-1.6848515272140503,
-0.031679801642894745,
-0.56804913282394... |
https://github.com/huggingface/datasets/issues/5317 | `ImageFolder` performs poorly with large datasets | @lhoestq Thanks!
Perhaps it'd be worth adding a note on the documentation that `ImageFolder` is not intended for large datasets? This limitation is not intuitively obvious to someone who has not used it before, I think. | ### Describe the bug
While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images.
## Setup
* Nested directories (5 levels deep)
* 3M+ images
* 1 `metadata.jsonl` file
## Performance Degradation Point... | 388 | 36 | `ImageFolder` performs poorly with large datasets
### Describe the bug
While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images.
## Setup
* Nested directories (5 levels deep)
* 3M+ images
* 1 `metada... | [
-1.1987686157226562,
-0.9523312449455261,
-0.7291271686553955,
1.4155699014663696,
-0.10784557461738586,
-1.2535969018936157,
0.13453370332717896,
-1.0448516607284546,
1.6751487255096436,
-0.7176129817962646,
0.2097993642091751,
-1.6848515272140503,
-0.031679801642894745,
-0.56804913282394... |
https://github.com/huggingface/datasets/issues/5317 | `ImageFolder` performs poorly with large datasets | Thanks for the feedback @salieri! I opened #5329 to make it clear `ImageFolder` is not intended for large datasets. Please feel free to comment if you have any other feedback! 🙂 | ### Describe the bug
While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images.
## Setup
* Nested directories (5 levels deep)
* 3M+ images
* 1 `metadata.jsonl` file
## Performance Degradation Point... | 388 | 31 | `ImageFolder` performs poorly with large datasets
### Describe the bug
While testing image dataset creation, I'm seeing significant performance bottlenecks with imagefolders when scanning a directory structure with large number of images.
## Setup
* Nested directories (5 levels deep)
* 3M+ images
* 1 `metada... | [
-1.1987686157226562,
-0.9523312449455261,
-0.7291271686553955,
1.4155699014663696,
-0.10784557461738586,
-1.2535969018936157,
0.13453370332717896,
-1.0448516607284546,
1.6751487255096436,
-0.7176129817962646,
0.2097993642091751,
-1.6848515272140503,
-0.031679801642894745,
-0.56804913282394... |
https://github.com/huggingface/datasets/issues/5315 | Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails | EDIT:
I think in this case, the metadata files (either README or JSON) should not be read (i.e. `self.info.splits` should be None).
One idea:
- I think ideally we should set this behavior when we pass `--save_info` to the CLI `test`
- However, currently, the builder is unaware of this: `save_info` arg is not pas... | ### Describe the bug
If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails.
That's what happened in https://huggingface.co/datasets/mrdbourke/food_vision_199_classes/discussions/2#6385f... | 390 | 57 | Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails
### Describe the bug
If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails.
That... | [
-1.113948941230774,
-0.9091671705245972,
-0.6788613200187683,
1.4712458848953247,
-0.11289050430059433,
-1.317491054534912,
0.1464281529188156,
-1.0284781455993652,
1.6395132541656494,
-0.7434434294700623,
0.2930738627910614,
-1.7083076238632202,
-0.017192382365465164,
-0.595932126045227,
... |
https://github.com/huggingface/datasets/issues/5315 | Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails | > I think in this case
@albertvillanova You mean in cases when the script was changed?
I suggest that we:
* add a check on the slice (like 'split_name[n%]) kind of format here: https://github.com/huggingface/datasets/blob/main/src/datasets/splits.py#L523 to catch things like this.
* Error here happens before... | ### Describe the bug
If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails.
That's what happened in https://huggingface.co/datasets/mrdbourke/food_vision_199_classes/discussions/2#6385f... | 390 | 142 | Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails
### Describe the bug
If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails.
That... | [
-1.113948941230774,
-0.9091671705245972,
-0.6788613200187683,
1.4712458848953247,
-0.11289050430059433,
-1.317491054534912,
0.1464281529188156,
-1.0284781455993652,
1.6395132541656494,
-0.7434434294700623,
0.2930738627910614,
-1.7083076238632202,
-0.017192382365465164,
-0.595932126045227,
... |
https://github.com/huggingface/datasets/issues/5315 | Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails | I edited my previous comment:
- First I proposed setting `self.info.splits` to None when `ignore_verifications=True`
- I thought it was the easiest implementation because `ignore_verifications` is passed to `DatasetBuilder.download_and_prepare`
- However, afterwards, I realized this might not be a good idea for ... | ### Describe the bug
If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails.
That's what happened in https://huggingface.co/datasets/mrdbourke/food_vision_199_classes/discussions/2#6385f... | 390 | 201 | Adding new splits to a dataset script with existing old splits info in metadata's `dataset_info` fails
### Describe the bug
If you first create a custom dataset with a specific set of splits, generate metadata with `datasets-cli test ... --save_info`, then change your script to include more splits, it fails.
That... | [
-1.113948941230774,
-0.9091671705245972,
-0.6788613200187683,
1.4712458848953247,
-0.11289050430059433,
-1.317491054534912,
0.1464281529188156,
-1.0284781455993652,
1.6395132541656494,
-0.7434434294700623,
0.2930738627910614,
-1.7083076238632202,
-0.017192382365465164,
-0.595932126045227,
... |
https://github.com/huggingface/datasets/issues/5314 | Datasets: classification_report() got an unexpected keyword argument 'suffix' | @JonathanAlis also note that the metrics are deprecated in our `datasets` library.
Please, use the new library 🤗 Evaluate instead: https://huggingface.co/docs/evaluate | https://github.com/huggingface/datasets/blob/main/metrics/seqeval/seqeval.py
> import datasets
predictions = [['O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-PER', 'I-PER', 'O']]
references = [['O', 'O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-PER', 'I-PER', 'O']]
seqeval = datasets.load_metri... | 391 | 21 | Datasets: classification_report() got an unexpected keyword argument 'suffix'
https://github.com/huggingface/datasets/blob/main/metrics/seqeval/seqeval.py
> import datasets
predictions = [['O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-PER', 'I-PER', 'O']]
references = [['O', 'O', 'O', 'B-MISC', '... | [
-1.3552393913269043,
-0.9346601963043213,
-0.8406415581703186,
1.6598427295684814,
-0.20891276001930237,
-1.05767822265625,
-0.007367828860878944,
-0.9546255469322205,
1.5658034086227417,
-0.7757533192634583,
0.22902336716651917,
-1.7130876779556274,
-0.0316036082804203,
-0.640803217887878... |
https://github.com/huggingface/datasets/issues/5305 | Dataset joelito/mc4_legal does not work with multiple files | Thanks for reporting @JoelNiklaus.
Please note that since we moved all dataset loading scripts to the Hub, the issues and pull requests relative to specific datasets are directly handled on the Hub, in their Community tab. I'm transferring this issue there: https://huggingface.co/datasets/joelito/mc4_legal/discussio... | ### Describe the bug
The dataset https://huggingface.co/datasets/joelito/mc4_legal works for languages like bg with a single data file, but not for languages with multiple files like de. It shows zero rows for the de dataset.
joelniklaus@Joels-MacBook-Pro ~/N/P/C/L/p/m/mc4_legal (main) [1]> python test_mc4_legal.... | 393 | 54 | Dataset joelito/mc4_legal does not work with multiple files
### Describe the bug
The dataset https://huggingface.co/datasets/joelito/mc4_legal works for languages like bg with a single data file, but not for languages with multiple files like de. It shows zero rows for the de dataset.
joelniklaus@Joels-MacBook-... | [
-1.258362054824829,
-0.9908857345581055,
-0.795711100101471,
1.4838215112686157,
-0.033149782568216324,
-1.2504149675369263,
0.1168370246887207,
-0.9932911396026611,
1.5552551746368408,
-0.6639343500137329,
0.20511026680469513,
-1.775362253189087,
-0.1270586997270584,
-0.5651912689208984,
... |
https://github.com/huggingface/datasets/issues/5304 | timit_asr doesn't load the test split. | The [timit_asr.py](https://huggingface.co/datasets/timit_asr/blob/main/timit_asr.py) script iterates over the WAV files per split directory using this:
```python
wav_paths = sorted(Path(data_dir).glob(f"**/{split}/**/*.wav"))
wav_paths = wav_paths if wav_paths else sorted(Path(data_dir).glob(f"**/{split.upper()}/**/... | ### Describe the bug
When I use the function ```timit = load_dataset('timit_asr', data_dir=data_dir)```, it only loads train split, not test split.
I tried to change the directory and filename to lower case to upper case for the test split, but it does not work at all.
```python
DatasetDict({
train: Datase... | 394 | 42 | timit_asr doesn't load the test split.
### Describe the bug
When I use the function ```timit = load_dataset('timit_asr', data_dir=data_dir)```, it only loads train split, not test split.
I tried to change the directory and filename to lower case to upper case for the test split, but it does not work at all.
``... | [
-1.1882445812225342,
-1.1478807926177979,
-0.7531515955924988,
1.6818410158157349,
-0.19282428920269012,
-1.1733320951461792,
0.07968270778656006,
-0.9431273937225342,
1.548446536064148,
-0.7097229361534119,
0.1579272747039795,
-1.7373665571212769,
-0.1452927589416504,
-0.7916383147239685,... |
https://github.com/huggingface/datasets/issues/5295 | Extractions failed when .zip file located on read-only path (e.g., SageMaker FastFile mode) | Hi ! Thanks for reporting. Indeed the lock file should be placed in a directory with write permission (e.g. in the directory where the archive is extracted). | ### Describe the bug
Hi,
`load_dataset()` does not work .zip files located on a read-only directory. Looks like it's because Dataset creates a lock file in the [same directory](https://github.com/huggingface/datasets/blob/df4bdd365f2abb695f113cbf8856a925bc70901b/src/datasets/utils/extract.py) as the .zip file.
... | 395 | 27 | Extractions failed when .zip file located on read-only path (e.g., SageMaker FastFile mode)
### Describe the bug
Hi,
`load_dataset()` does not work .zip files located on a read-only directory. Looks like it's because Dataset creates a lock file in the [same directory](https://github.com/huggingface/datasets/blo... | [
-1.2134716510772705,
-0.9405736923217773,
-0.6529114842414856,
1.45478355884552,
-0.1730504184961319,
-1.1634786128997803,
0.13532023131847382,
-0.9901857972145081,
1.531798243522644,
-0.7498309016227722,
0.31072452664375305,
-1.6880139112472534,
-0.04647236317396164,
-0.52350252866745,
... |
https://github.com/huggingface/datasets/issues/5295 | Extractions failed when .zip file located on read-only path (e.g., SageMaker FastFile mode) | I opened https://github.com/huggingface/datasets/pull/5320 to fix this - it places the lock file in the cache directory instead of trying to put in next to the ZIP where it's read-only | ### Describe the bug
Hi,
`load_dataset()` does not work .zip files located on a read-only directory. Looks like it's because Dataset creates a lock file in the [same directory](https://github.com/huggingface/datasets/blob/df4bdd365f2abb695f113cbf8856a925bc70901b/src/datasets/utils/extract.py) as the .zip file.
... | 395 | 29 | Extractions failed when .zip file located on read-only path (e.g., SageMaker FastFile mode)
### Describe the bug
Hi,
`load_dataset()` does not work .zip files located on a read-only directory. Looks like it's because Dataset creates a lock file in the [same directory](https://github.com/huggingface/datasets/blo... | [
-1.2134716510772705,
-0.9405736923217773,
-0.6529114842414856,
1.45478355884552,
-0.1730504184961319,
-1.1634786128997803,
0.13532023131847382,
-0.9901857972145081,
1.531798243522644,
-0.7498309016227722,
0.31072452664375305,
-1.6880139112472534,
-0.04647236317396164,
-0.52350252866745,
... |
https://github.com/huggingface/datasets/issues/5292 | Missing documentation build for versions 2.7.1 and 2.6.2 | - Build docs for 2.6.2:
- Commit: a6a5a1cf4cdf1e0be65168aed5a327f543001fe8
- Build docs GH Action: https://github.com/huggingface/datasets/actions/runs/3539470622/jobs/5941404044
- Build docs for 2.7.1:
- Commit: 5ef1ab1cc06c2b7a574bf2df454cd9fcb071ccb2
- Build docs GH Action: https://github.com/huggingfac... | After the patch releases [2.7.1](https://github.com/huggingface/datasets/releases/tag/2.7.1) and [2.6.2](https://github.com/huggingface/datasets/releases/tag/2.6.2), the online docs were not properly built (the build_documentation workflow was not triggered).
There was a fix by:
- #5291
However, both documentati... | 396 | 28 | Missing documentation build for versions 2.7.1 and 2.6.2
After the patch releases [2.7.1](https://github.com/huggingface/datasets/releases/tag/2.7.1) and [2.6.2](https://github.com/huggingface/datasets/releases/tag/2.6.2), the online docs were not properly built (the build_documentation workflow was not triggered).
... | [
-1.233512282371521,
-0.8936444520950317,
-0.8072652220726013,
1.3956623077392578,
-0.07910232990980148,
-1.240429162979126,
0.026858747005462646,
-1.0523921251296997,
1.5705183744430542,
-0.635078489780426,
0.31448063254356384,
-1.733963131904602,
-0.09406106919050217,
-0.4470330476760864,... |
https://github.com/huggingface/datasets/issues/5288 | Lossy json serialization - deserialization of dataset info | Hi ! JSON is a lossy format indeed. If you want to keep the feature types or other metadata I'd encourage you to store them as well. For example you can use `dataset.info.write_to_directory` and `DatasetInfo.from_directory` to store the feature types, split info, description, license etc. | ### Describe the bug
Saving a dataset to disk as json (using `to_json`) and then loading it again (using `load_dataset`) results in features whose labels are not type-cast correctly. In the code snippet below, `features.label` should have a label of type `ClassLabel` but has type `Value` instead.
### Steps to re... | 397 | 45 | Lossy json serialization - deserialization of dataset info
### Describe the bug
Saving a dataset to disk as json (using `to_json`) and then loading it again (using `load_dataset`) results in features whose labels are not type-cast correctly. In the code snippet below, `features.label` should have a label of type ... | [
-1.1897344589233398,
-0.9814062714576721,
-0.6852878928184509,
1.635185718536377,
-0.21063145995140076,
-1.1490113735198975,
0.1617978811264038,
-1.009643316268921,
1.7089990377426147,
-0.745701789855957,
0.26563555002212524,
-1.6730543375015259,
0.027536578476428986,
-0.6985669732093811,
... |
https://github.com/huggingface/datasets/issues/5286 | FileNotFoundError: Couldn't find file at https://dumps.wikimedia.org/enwiki/20220301/dumpstatus.json | I found a solution
If you specifically install datasets==1.18 and then run
import datasets
wiki = datasets.load_dataset('wikipedia', '20200501.en')
then this should work (it worked for me.) | ### Describe the bug
I follow the steps provided on the website [https://huggingface.co/datasets/wikipedia](https://huggingface.co/datasets/wikipedia)
$ pip install apache_beam mwparserfromhell
>>> from datasets import load_dataset
>>> load_dataset("wikipedia", "20220301.en")
however this results in the follo... | 398 | 26 | FileNotFoundError: Couldn't find file at https://dumps.wikimedia.org/enwiki/20220301/dumpstatus.json
### Describe the bug
I follow the steps provided on the website [https://huggingface.co/datasets/wikipedia](https://huggingface.co/datasets/wikipedia)
$ pip install apache_beam mwparserfromhell
>>> from datasets... | [
-1.2083998918533325,
-0.9795846939086914,
-0.7561137080192566,
1.441972255706787,
-0.188189297914505,
-1.1714004278182983,
0.05929543077945709,
-0.9904125332832336,
1.6878396272659302,
-0.6838863492012024,
0.20123139023780823,
-1.6519056558609009,
-0.08784256130456924,
-0.5095714330673218,... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | Thanks @lhoestq and @alvarobartt!
This would be extremely helpful to have working for the Whisper fine-tuning event - we're **only** training using streaming mode, so it'll be quite important to have this feature working to make training as easy as possible!
_c.f._ https://twitter.com/sanchitgandhi99/status/15921... | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 43 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | > Thanks @lhoestq and @alvarobartt!
>
>
>
> This would be extremely helpful to have working for the Whisper fine-tuning event - we're **only** training using streaming mode, so it'll be quite important to have this feature working to make training as easy as possible!
>
>
>
> _c.f._ https://twitter.com/sanchitgan... | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 100 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | @sanchit-gandhi PR is ready and open for review at #5287, but there's still one issue I may need @lhoestq's input :hugs: | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 21 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | Let us know @sanchit-gandhi if you need a new release of `datasets` soon with this fix included :) | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 18 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | Thanks for the fix guys! We can direct people to install `datasets` from main if that's easier! | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 17 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | Hi @asennoussi so this was already fixed and released as part of https://github.com/huggingface/datasets/releases/tag/2.8.0, so you should be able to install it as `pip install datasets==2.8.0` or just to use `pip install datasets --upgrade` to get the latest version, as of now, the https://github.com/huggingface/datas... | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 47 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | Still facing the same issue though:
```
from datasets import IterableDatasetDict, load_dataset
raw_datasets = vectorized_datasets = IterableDatasetDict()
raw_datasets["train"] = load_dataset("asennoussi/private", split="train", use_auth_token=True, streaming=True)
raw_datasets["test"] = load_dataset("asenno... | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 120 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | Hmm weird, could you try to print
```python
print("Processed features: ", vectorized_datasets['train'].features)
```
again after iterating over the `vectorized_datasets`? In the code above, should be last line :) | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 28 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | Didn't seem to fix it:
```
Original features: dict_keys(['path', 'audio', 'sentence'])
Processed features: None
Processed features: None
``` | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 18 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | Actually the culprit looks to be this one:
`vectorized_datasets = raw_datasets.map(prepare_dataset).with_format("torch")`
When I remove this line: `vectorized_datasets = vectorized_datasets.remove_columns(['input_length', 'labels_length']+list(next(iter(raw_datasets.values())).features))`
I still get
```
Proce... | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 28 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | > The culprit is definitely `.map` Just validated it. Any idea please?
Yes, indeed `.map` losses the features, because AFAIK pre-fetching the data to infer the features is expensive and not ideal, that's part of this issue https://github.com/huggingface/datasets/issues/3888
Anyway, now you can pass the `features`... | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 101 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5284 | Features of IterableDataset set to None by remove column | @asennoussi alternatively you can just call `._resolve_features()` from your `IterableDataset` and it will pre-fetch the data to resolve the features, but note that feature-inference is not as accurate as if you manually specify which features and feature-types the `IterableDataset` has, as mentioned in the comment abo... | ### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True)
... | 399 | 56 | Features of IterableDataset set to None by remove column
### Describe the bug
The `remove_column` method of the IterableDataset sets the dataset features to None.
### Steps to reproduce the bug
```python
from datasets import Audio, load_dataset
# load LS in streaming mode
dataset = load_dataset("librisp... | [
-1.1740634441375732,
-0.986666202545166,
-0.724602222442627,
1.5060685873031616,
-0.16580796241760254,
-1.1591317653656006,
0.15898366272449493,
-0.9965685606002808,
1.6621525287628174,
-0.7151062488555908,
0.23000434041023254,
-1.7126665115356445,
-0.02927146479487419,
-0.7022446990013123... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | Or for example an archive on GitHub releases! Before I added support for JXL (locally only, PR still pending) I was considering hosting my files on GitHub instead... | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 28 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1481211185455322,
-0.8668554425239563,
-0.7009403109550476,
1.5864934921264648,
-0.11684641242027283,
-1.2785141468048096,
0.226686030626297,
-0.9447305202484131,
1.6145557165145874,
-0.7712887525558472,
0.4066356420516968,
-1.7100402116775513,
0.02905506268143654,
-0.5384873747825623,
... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | +1 to this. I would like to use 'audiofolder' with a data_dir that's on S3, for example. I don't want to upload my dataset to the Hub, but I would find all the fingerprinting/caching features useful. | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 36 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1301863193511963,
-0.887853741645813,
-0.7564798593521118,
1.556461215019226,
-0.11461750417947769,
-1.2745490074157715,
0.21984700858592987,
-0.9521751403808594,
1.6449880599975586,
-0.7799682021141052,
0.39993906021118164,
-1.7132654190063477,
0.010704532265663147,
-0.5761101841926575... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | Adding to the conversation, Dask also uses `fsspec` for this feature.
[Dask: How to connect to remote data](https://docs.dask.org/en/stable/how-to/connect-to-remote-data.html)
Happy to help on this feature :D | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 25 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1261876821517944,
-0.865461528301239,
-0.7369795441627502,
1.5711097717285156,
-0.07206477969884872,
-1.266869306564331,
0.1847299188375473,
-0.9479480981826782,
1.653198003768921,
-0.7508422136306763,
0.4013937711715698,
-1.6984187364578247,
-0.013095520436763763,
-0.5696078538894653,
... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | +1 to this feature request since I think it also tackles my use-case. I am collaborating with a team, working with a loading script which takes some time to generate the dataset artifacts. It would be very handy to use this as a cloud cache to avoid duplicating the effort.
Currently we could use `builder.download_a... | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 78 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1604654788970947,
-0.8987336158752441,
-0.6765923500061035,
1.4956525564193726,
-0.12263771891593933,
-1.270355463027954,
0.20320464670658112,
-1.019025206565857,
1.624469518661499,
-0.7517934441566467,
0.37090015411376953,
-1.6710683107376099,
-0.02991263009607792,
-0.5710994601249695,... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | Makes sense ! If you want to load locally a dataset that you download_and_prepared on a cloud storage, you would use `load_dataset(path_to_cloud_storage)` indeed. It would download the data from the cloud storage, cache them locally, and return a `Dataset`. | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 39 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1247591972351074,
-0.8659507632255554,
-0.7190048694610596,
1.5625264644622803,
-0.10728877782821655,
-1.3009120225906372,
0.19700269401073456,
-0.9690713286399841,
1.6566994190216064,
-0.806896448135376,
0.38287004828453064,
-1.6959503889083862,
-0.002525700256228447,
-0.58512496948242... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | It seems currently the `cached_path` function handles all URLs by `get_from_cache` that only supports `ftp` and `http(s)` here:
https://github.com/huggingface/datasets/blob/b5672a956d5de864e6f5550e493527d962d6ae55/src/datasets/utils/file_utils.py#L181
I guess one can add another condition that handles `s3://` or `g... | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 35 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1119447946548462,
-0.8206652402877808,
-0.7333627343177795,
1.5866179466247559,
-0.06604381650686264,
-1.2926174402236938,
0.2216891348361969,
-0.9680379033088684,
1.628993034362793,
-0.7275192737579346,
0.40322235226631165,
-1.7049504518508911,
-0.05162134766578674,
-0.5773789286613464... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | I could use this functionality, so I put together a PR using @kyamagu's suggestion to use `fsspec` in `datasets.utils.file_utils`
https://github.com/huggingface/datasets/pull/5580 | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 20 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.0952835083007812,
-0.8498707413673401,
-0.7236981987953186,
1.5619399547576904,
-0.08013491332530975,
-1.2700855731964111,
0.22350482642650604,
-0.9411389231681824,
1.6429386138916016,
-0.7555069923400879,
0.39713624119758606,
-1.700418472290039,
-0.03550271317362785,
-0.575844228267669... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | Thanks @dwyatte for adding support for fsspec urls
Let me just reopen this since the original issue is not resolved | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 20 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1162397861480713,
-0.8388081789016724,
-0.6944715976715088,
1.6116211414337158,
-0.1191871166229248,
-1.2930155992507935,
0.19488219916820526,
-0.9393096566200256,
1.6099915504455566,
-0.7943482995033264,
0.4006102979183197,
-1.6975775957107544,
0.011448396369814873,
-0.5964275598526001... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | I'm not yet understanding how to use https://github.com/huggingface/datasets/pull/5580 in order to use `load_dataset(data_files="s3://...")`. Any help/example would be much appreciated :) thanks! | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 21 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.0874907970428467,
-0.874066174030304,
-0.7081629037857056,
1.5715142488479614,
-0.09644901752471924,
-1.2539818286895752,
0.1999494582414627,
-0.9408275485038757,
1.6259489059448242,
-0.7350842952728271,
0.38461852073669434,
-1.6943416595458984,
-0.04317580536007881,
-0.5396792888641357... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | It's still not officially supported x) But you can try to update `request_etag` in `file_utils.py` to use `fsspec_head` instead of `http_head`. It is responsible of getting the ETags of the remote files for caching. This change may do the trick for S3 urls | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 43 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.0978657007217407,
-0.8224558234214783,
-0.7115296721458435,
1.6014676094055176,
-0.09623211622238159,
-1.293861985206604,
0.23310768604278564,
-0.9644601941108704,
1.6769380569458008,
-0.8170479536056519,
0.4120827317237854,
-1.6851418018341064,
0.04258618503808975,
-0.5588087439537048,... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | Thank you for your guys help on this and merging in #5580. I manually pulled the changes to my local datasets package (datasets.utils.file_utils.py) since it only seemed to be this file that was changed in the PR and I'm getting the error:
InvalidSchema: No connection adapters were found for 's3://bucket/folder/'. I'... | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 92 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1877988576889038,
-0.9315599799156189,
-0.7374240756034851,
1.5009020566940308,
-0.11799561232328415,
-1.2380927801132202,
0.1705128252506256,
-1.006676197052002,
1.680005431175232,
-0.7602040767669678,
0.33744391798973083,
-1.6719021797180176,
-0.020392054691910744,
-0.5125157833099365... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | The current implementation depends on gcsfs/s3fs being able to authenticate through some other means e.g., environmental variables. For AWS, it looks like you can set `AWS_ACCESS_KEY_ID`, `AWS_SECRET_ACCESS_KEY`, and `AWS_SESSION_TOKEN`
Note that while testing this just now, I did note a discrepancy between gcsfs an... | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 130 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1259617805480957,
-0.8903080821037292,
-0.7191109657287598,
1.5024460554122925,
-0.09601156413555145,
-1.2793055772781372,
0.1703445017337799,
-1.0465589761734009,
1.6725752353668213,
-0.7534469366073608,
0.30338990688323975,
-1.6468414068222046,
-0.00529266195371747,
-0.590070068836212... |
https://github.com/huggingface/datasets/issues/5281 | Support cloud storage in load_dataset | > Note that while testing this just now, I did note a discrepancy between gcsfs and s3fs that we might want to address where gcsfs passes the timeout from storage_options [here](https://github.com/huggingface/datasets/blob/3e6269979fc80ae8939294d26298897f0db5b84d/src/datasets/utils/file_utils.py#L333) down into the aio... | Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_prepare` and `save_to_disk`.
This has... | 400 | 109 | Support cloud storage in load_dataset
Would be nice to be able to do
```python
load_dataset("s3://...")
```
or even
```python
data_files=["gs://..."]
storage_options = {...}
load_dataset(..., data_files=data_files, storage_options=storage_options)
```
The idea would be to use `fsspec` as in `download_and_... | [
-1.1931078433990479,
-0.8865266442298889,
-0.7406773567199707,
1.4943337440490723,
-0.12763677537441254,
-1.2567174434661865,
0.1296108514070511,
-1.0605189800262451,
1.6105358600616455,
-0.7359844446182251,
0.3050369322299957,
-1.6380168199539185,
-0.007218174636363983,
-0.578545749187469... |
https://github.com/huggingface/datasets/issues/5280 | Import error | Hi,
3.8.13
Get Outlook for Android<https://aka.ms/AAb9ysg>
________________________________
From: Quentin Lhoest ***@***.***>
Sent: Tuesday, November 22, 2022 2:37:02 PM
To: huggingface/datasets ***@***.***>
Cc: feketedavid1012 ***@***.***>; Author ***@***.***>
Subject: Re: [huggingface/datasets] Import error (Issue ... | https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28
Hy,
I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message wrong) | 401 | 69 | Import error
https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28
Hy,
I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message w... | [
-1.2245451211929321,
-0.9519438147544861,
-0.5376725792884827,
1.4114515781402588,
-0.18234844505786896,
-1.2721543312072754,
0.1501534879207611,
-1.0091004371643066,
1.595414638519287,
-0.5358529090881348,
0.1921544373035431,
-1.6243865489959717,
-0.10301874577999115,
-0.5325057506561279,... |
https://github.com/huggingface/datasets/issues/5280 | Import error | Then it should work as expected if you use the same python when using `datasets`
Please make sure you're running your code in the right environment | https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28
Hy,
I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message wrong) | 401 | 26 | Import error
https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28
Hy,
I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message w... | [
-1.2907732725143433,
-0.8051122426986694,
-0.8636984825134277,
1.4544912576675415,
-0.15099065005779266,
-1.2534384727478027,
0.03319337218999863,
-1.0669690370559692,
1.4942463636398315,
-0.7183879017829895,
0.19982896745204926,
-1.6829895973205566,
-0.06631534546613693,
-0.39326384663581... |
https://github.com/huggingface/datasets/issues/5280 | Import error | It's the right environment. But in if statement I have
"3.8.13" < 3.7
And in the error message is Python>=3.7 which is true in my case (3.8.13 is greater then 3.7), so I don't understand my python should be below the 3.7 which case the if statement is right, but the message is wrong, or above 3.7 which case if statemen... | https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28
Hy,
I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message wrong) | 401 | 147 | Import error
https://github.com/huggingface/datasets/blob/cd3d8e637cfab62d352a3f4e5e60e96597b5f0e9/src/datasets/__init__.py#L28
Hy,
I have error at the above line. I have python version 3.8.13, the message says I need python>=3.7, which is True, but I think the if statement not working properly (or the message w... | [
-1.2348964214324951,
-0.9628428816795349,
-0.5841107368469238,
1.4090332984924316,
-0.2059527337551117,
-1.2893599271774292,
0.14153769612312317,
-1.034083604812622,
1.5843340158462524,
-0.6413482427597046,
0.1616053581237793,
-1.6310269832611084,
-0.13474372029304504,
-0.5262941718101501,... |
https://github.com/huggingface/datasets/issues/5278 | load_dataset does not read jsonl metadata file properly | Can you try to remove "drop_labels=false" ? It may force the loader to infer the labels instead of reading the metadata | ### Describe the bug
Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column. It just spits out "image" and "label" as features.
B... | 402 | 21 | load_dataset does not read jsonl metadata file properly
### Describe the bug
Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column... | [
-1.1722538471221924,
-0.9874001741409302,
-0.759548544883728,
1.489896535873413,
-0.1571158766746521,
-1.2294549942016602,
0.1569814831018448,
-1.0602964162826538,
1.5801401138305664,
-0.8204648494720459,
0.28715646266937256,
-1.6882760524749756,
-0.015584716573357582,
-0.5398836731910706,... |
https://github.com/huggingface/datasets/issues/5278 | load_dataset does not read jsonl metadata file properly | Okay, now it seems to actually load the metadata and create the train_split, but it still says only returns "image" and "label", which is always 0 since all images are from same folder | ### Describe the bug
Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column. It just spits out "image" and "label" as features.
B... | 402 | 33 | load_dataset does not read jsonl metadata file properly
### Describe the bug
Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column... | [
-1.1759647130966187,
-0.9833332896232605,
-0.7441419959068298,
1.4840329885482788,
-0.15907064080238342,
-1.2307698726654053,
0.1437048763036728,
-1.065576195716858,
1.5801584720611572,
-0.8430256247520447,
0.2931825816631317,
-1.6919947862625122,
-0.019899562001228333,
-0.5299819707870483... |
https://github.com/huggingface/datasets/issues/5278 | load_dataset does not read jsonl metadata file properly | > Can you try updating `datasets` ? Metadata support was added in `datasets` 2.4
Update: This was the issue. | ### Describe the bug
Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column. It just spits out "image" and "label" as features.
B... | 402 | 19 | load_dataset does not read jsonl metadata file properly
### Describe the bug
Hi, I'm following [this page](https://huggingface.co/docs/datasets/image_dataset) to create a dataset of images and captions via an image folder and a metadata.json file, but I can't seem to get the dataloader to recognize the "text" column... | [
-1.1760004758834839,
-0.9716751575469971,
-0.7537791728973389,
1.4831427335739136,
-0.15419664978981018,
-1.2533154487609863,
0.1560819000005722,
-1.051365613937378,
1.6201120615005493,
-0.7968655824661255,
0.27176833152770996,
-1.7120777368545532,
-0.039096683263778687,
-0.544991195201873... |
https://github.com/huggingface/datasets/issues/5276 | Bug in downloading common_voice data and snall chunk of it to one's own hub | Sounds like one of the file is not a valid one, can you make sure you uploaded valid mp3 files ? | ### Describe the bug
I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset
Help please?

| ### Describe the bug
I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset
Help please?
:
- YAML integer keys are not preserved server-side: they are transformed to strings
- See for example this Hub PR: https://huggingface.co/datasets/acronym_identification/discussions/1/files
- Original:
```yaml
... | 404 | 19 | YAML integer keys are not preserved Hub server-side
After an internal discussion (https://github.com/huggingface/moon-landing/issues/4563):
- YAML integer keys are not preserved server-side: they are transformed to strings
- See for example this Hub PR: https://huggingface.co/datasets/acronym_identification/discu... | [
-1.0408730506896973,
-0.7701073288917542,
-0.7022444009780884,
1.6244213581085205,
-0.0046040830202400684,
-1.3531185388565063,
0.038567349314689636,
-0.9089091420173645,
1.5970324277877808,
-0.737040638923645,
0.4560614228248596,
-1.7285661697387695,
-0.056928135454654694,
-0.701209604740... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.