html_url stringlengths 51 51 | title stringlengths 6 280 | comments stringlengths 67 24.7k | body stringlengths 51 36.2k | __index_level_0__ int64 1 1.17k | comment_length int64 16 1.45k | text stringlengths 190 38.3k | embeddings list |
|---|---|---|---|---|---|---|---|
https://github.com/huggingface/datasets/issues/5513 | Some functions use a param named `type` shouldn't that be avoided since it's a Python reserved name? | Hi @mariosasko, ok it makes sense. Anyway, don't you think it's worth it at some point to start a deprecation cycle e.g. `fs` in `load_from_disk`? It doesn't affect user experience but it's for sure a bad practice IMO, but's up to you 😄 Feel free to close this issue otherwise! | Hi @mariosasko, @lhoestq, or whoever reads this! :)
After going through `ArrowDataset.set_format` I found out that the `type` param is actually named `type` which is a Python reserved name as you may already know, shouldn't that be renamed to `format_type` before the 3.0.0 is released?
Just wanted to get your inp... | 304 | 50 | Some functions use a param named `type` shouldn't that be avoided since it's a Python reserved name?
Hi @mariosasko, @lhoestq, or whoever reads this! :)
After going through `ArrowDataset.set_format` I found out that the `type` param is actually named `type` which is a Python reserved name as you may already know, ... | [
-1.2074476480484009,
-0.9789834022521973,
-0.7793334722518921,
1.492044448852539,
-0.17461514472961426,
-1.2549453973770142,
0.07254193723201752,
-1.0410224199295044,
1.714010238647461,
-0.853183388710022,
0.3969155550003052,
-1.7728849649429321,
0.09019886702299118,
-0.6305654048919678,
... |
https://github.com/huggingface/datasets/issues/5513 | Some functions use a param named `type` shouldn't that be avoided since it's a Python reserved name? | I don't think deprecating a param name in this particular instance is worth the hassle, so I'm closing the issue 🙂. | Hi @mariosasko, @lhoestq, or whoever reads this! :)
After going through `ArrowDataset.set_format` I found out that the `type` param is actually named `type` which is a Python reserved name as you may already know, shouldn't that be renamed to `format_type` before the 3.0.0 is released?
Just wanted to get your inp... | 304 | 21 | Some functions use a param named `type` shouldn't that be avoided since it's a Python reserved name?
Hi @mariosasko, @lhoestq, or whoever reads this! :)
After going through `ArrowDataset.set_format` I found out that the `type` param is actually named `type` which is a Python reserved name as you may already know, ... | [
-1.2216826677322388,
-0.9274621605873108,
-0.8090871572494507,
1.4519572257995605,
-0.2148088812828064,
-1.2933354377746582,
0.16014038026332855,
-1.0748504400253296,
1.6663943529129028,
-0.8881301879882812,
0.4465438723564148,
-1.773697853088379,
0.0532158687710762,
-0.5506930947303772,
... |
https://github.com/huggingface/datasets/issues/5511 | Creating a dummy dataset from a bigger one | Update `datasets` or downgrade `huggingface-hub` ;)
The `huggingface-hub` lib did a breaking change a few months ago, and you're using an old version of `datasets` that does't support it | ### Describe the bug
I often want to create a dummy dataset from a bigger dataset for fast iteration when training. However, I'm having a hard time doing this especially when trying to upload the dataset to the Hub.
### Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset... | 305 | 29 | Creating a dummy dataset from a bigger one
### Describe the bug
I often want to create a dummy dataset from a bigger dataset for fast iteration when training. However, I'm having a hard time doing this especially when trying to upload the dataset to the Hub.
### Steps to reproduce the bug
```python
from datasets ... | [
-1.2701120376586914,
-0.9678511023521423,
-0.6017754077911377,
1.4185889959335327,
-0.156211256980896,
-1.290795922279358,
0.181573748588562,
-1.0857007503509521,
1.5226128101348877,
-0.8153998851776123,
0.2889370322227478,
-1.6111980676651,
0.009617511183023453,
-0.6312329769134521,
-0.... |
https://github.com/huggingface/datasets/issues/5511 | Creating a dummy dataset from a bigger one | Getting same error with latest versions.
```shell
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
Cell In[99], line 1
----> 1 dataset.push_to_hub("mirfan899/kids_phoneme_asr")
File /opt/conda/lib/python3.... | ### Describe the bug
I often want to create a dummy dataset from a bigger dataset for fast iteration when training. However, I'm having a hard time doing this especially when trying to upload the dataset to the Hub.
### Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset... | 305 | 237 | Creating a dummy dataset from a bigger one
### Describe the bug
I often want to create a dummy dataset from a bigger dataset for fast iteration when training. However, I'm having a hard time doing this especially when trying to upload the dataset to the Hub.
### Steps to reproduce the bug
```python
from datasets ... | [
-1.2701120376586914,
-0.9678511023521423,
-0.6017754077911377,
1.4185889959335327,
-0.156211256980896,
-1.290795922279358,
0.181573748588562,
-1.0857007503509521,
1.5226128101348877,
-0.8153998851776123,
0.2889370322227478,
-1.6111980676651,
0.009617511183023453,
-0.6312329769134521,
-0.... |
https://github.com/huggingface/datasets/issues/5508 | Saving a dataset after setting format to torch doesn't work, but only if filtering | Hey, I'm a research engineer working on language modelling wanting to contribute to open source. I was wondering if I could give it a shot? | ### Describe the bug
Saving a dataset after setting format to torch doesn't work, but only if filtering
### Steps to reproduce the bug
```
a = Dataset.from_dict({"b": [1, 2]})
a.set_format('torch')
a.save_to_disk("test_save") # saves successfully
a.filter(None).save_to_disk("test_save_filter") # does not
>> [..... | 306 | 25 | Saving a dataset after setting format to torch doesn't work, but only if filtering
### Describe the bug
Saving a dataset after setting format to torch doesn't work, but only if filtering
### Steps to reproduce the bug
```
a = Dataset.from_dict({"b": [1, 2]})
a.set_format('torch')
a.save_to_disk("test_save") # s... | [
-1.1304316520690918,
-0.8675511479377747,
-0.77523273229599,
1.534570336341858,
-0.15542897582054138,
-1.2044962644577026,
0.15054906904697418,
-1.0520758628845215,
1.714974045753479,
-0.8151836395263672,
0.31460586190223694,
-1.6270716190338135,
0.09192617982625961,
-0.6353017091751099,
... |
https://github.com/huggingface/datasets/issues/5506 | IterableDataset and Dataset return different batch sizes when using Trainer with multiple GPUs | Hi ! `datasets` doesn't do batching - the PyTorch DataLoader does and is created by the `Trainer`. Do you pass other arguments to training_args with respect to data loading ?
Also we recently released `.to_iterable_dataset` that does pretty much what you implemented, but using contiguous shards to get a better speed... | ### Describe the bug
I am training a Roberta model using 2 GPUs and the `Trainer` API with a batch size of 256.
Initially I used a standard `Dataset`, but had issues with slow data loading. After reading [this issue](https://github.com/huggingface/datasets/issues/2252), I swapped to loading my dataset as contiguous... | 307 | 61 | IterableDataset and Dataset return different batch sizes when using Trainer with multiple GPUs
### Describe the bug
I am training a Roberta model using 2 GPUs and the `Trainer` API with a batch size of 256.
Initially I used a standard `Dataset`, but had issues with slow data loading. After reading [this issue](ht... | [
-1.330300211906433,
-1.0070868730545044,
-0.7057075500488281,
1.541914939880371,
-0.20247779786586761,
-1.204946756362915,
0.18128500878810883,
-1.0793545246124268,
1.6051658391952515,
-0.857101559638977,
0.3794158399105072,
-1.6458162069320679,
0.10278196632862091,
-0.5253657102584839,
... |
https://github.com/huggingface/datasets/issues/5506 | IterableDataset and Dataset return different batch sizes when using Trainer with multiple GPUs | This is the full set of training args passed. No training args were changed when switching dataset types.
```python
training_args = TrainingArguments(
output_dir="./checkpoints",
overwrite_output_dir=True,
num_train_epochs=1,
per_device_train_batch_size=256,
save_steps=2000,
save_total... | ### Describe the bug
I am training a Roberta model using 2 GPUs and the `Trainer` API with a batch size of 256.
Initially I used a standard `Dataset`, but had issues with slow data loading. After reading [this issue](https://github.com/huggingface/datasets/issues/2252), I swapped to loading my dataset as contiguous... | 307 | 43 | IterableDataset and Dataset return different batch sizes when using Trainer with multiple GPUs
### Describe the bug
I am training a Roberta model using 2 GPUs and the `Trainer` API with a batch size of 256.
Initially I used a standard `Dataset`, but had issues with slow data loading. After reading [this issue](ht... | [
-1.330300211906433,
-1.0070868730545044,
-0.7057075500488281,
1.541914939880371,
-0.20247779786586761,
-1.204946756362915,
0.18128500878810883,
-1.0793545246124268,
1.6051658391952515,
-0.857101559638977,
0.3794158399105072,
-1.6458162069320679,
0.10278196632862091,
-0.5253657102584839,
... |
https://github.com/huggingface/datasets/issues/5506 | IterableDataset and Dataset return different batch sizes when using Trainer with multiple GPUs | Makes sense. Given that it's a `transformers` issue and already being tracked, I'll close this out. | ### Describe the bug
I am training a Roberta model using 2 GPUs and the `Trainer` API with a batch size of 256.
Initially I used a standard `Dataset`, but had issues with slow data loading. After reading [this issue](https://github.com/huggingface/datasets/issues/2252), I swapped to loading my dataset as contiguous... | 307 | 16 | IterableDataset and Dataset return different batch sizes when using Trainer with multiple GPUs
### Describe the bug
I am training a Roberta model using 2 GPUs and the `Trainer` API with a batch size of 256.
Initially I used a standard `Dataset`, but had issues with slow data loading. After reading [this issue](ht... | [
-1.330300211906433,
-1.0070868730545044,
-0.7057075500488281,
1.541914939880371,
-0.20247779786586761,
-1.204946756362915,
0.18128500878810883,
-1.0793545246124268,
1.6051658391952515,
-0.857101559638977,
0.3794158399105072,
-1.6458162069320679,
0.10278196632862091,
-0.5253657102584839,
... |
https://github.com/huggingface/datasets/issues/5505 | PyTorch BatchSampler still loads from Dataset one-by-one | This change seems to come from a few months ago in the PyTorch side. That's good news and it means we may not need to pass a batch_sampler as soon as we add `Dataset.__getitems__` to get the optimal speed :)
Thanks for reporting ! Would you like to open a PR to add `__getitems__` and remove this outdated documentati... | ### Describe the bug
In [the docs here](https://huggingface.co/docs/datasets/use_with_pytorch#use-a-batchsampler), it mentions the issue of the Dataset being read one-by-one, then states that using a BatchSampler resolves the issue.
I'm not sure if this is a mistake in the docs or the code, but it seems that the on... | 308 | 60 | PyTorch BatchSampler still loads from Dataset one-by-one
### Describe the bug
In [the docs here](https://huggingface.co/docs/datasets/use_with_pytorch#use-a-batchsampler), it mentions the issue of the Dataset being read one-by-one, then states that using a BatchSampler resolves the issue.
I'm not sure if this is ... | [
-1.2740635871887207,
-0.9522457718849182,
-0.8109618425369263,
1.4744322299957275,
-0.08666042983531952,
-1.3567854166030884,
0.14825250208377838,
-1.0554382801055908,
1.7112696170806885,
-0.830684244632721,
0.2660365104675293,
-1.6319947242736816,
0.12265457957983017,
-0.5992708206176758,... |
https://github.com/huggingface/datasets/issues/5505 | PyTorch BatchSampler still loads from Dataset one-by-one | Yeah I figured this was the sort of thing that probably once worked. I can confirm that you no longer need the batch sampler, just `batch_size=n` in the `DataLoader`.
I'll pass on the PR, I'm flat out right now, sorry. | ### Describe the bug
In [the docs here](https://huggingface.co/docs/datasets/use_with_pytorch#use-a-batchsampler), it mentions the issue of the Dataset being read one-by-one, then states that using a BatchSampler resolves the issue.
I'm not sure if this is a mistake in the docs or the code, but it seems that the on... | 308 | 40 | PyTorch BatchSampler still loads from Dataset one-by-one
### Describe the bug
In [the docs here](https://huggingface.co/docs/datasets/use_with_pytorch#use-a-batchsampler), it mentions the issue of the Dataset being read one-by-one, then states that using a BatchSampler resolves the issue.
I'm not sure if this is ... | [
-1.2740635871887207,
-0.9522457718849182,
-0.8109618425369263,
1.4744322299957275,
-0.08666042983531952,
-1.3567854166030884,
0.14825250208377838,
-1.0554382801055908,
1.7112696170806885,
-0.830684244632721,
0.2660365104675293,
-1.6319947242736816,
0.12265457957983017,
-0.5992708206176758,... |
https://github.com/huggingface/datasets/issues/5499 | `load_dataset` has ~4 seconds of overhead for cached data | Hi ! To skip the verification step that checks if newer data exist, you can enable offline mode with `HF_DATASETS_OFFLINE=1`.
Although I agree this step should be much faster for datasets hosted on the HF Hub - we could just compare the commit hash from the local data and the remote git repository. We're not been le... | ### Feature request
When loading a dataset that has been cached locally, the `load_dataset` function takes a lot longer than it should take to fetch the dataset from disk (or memory).
This is particularly noticeable for smaller datasets. For example, wikitext-2, comparing `load_data` (once cached) and `load_from_disk... | 310 | 77 | `load_dataset` has ~4 seconds of overhead for cached data
### Feature request
When loading a dataset that has been cached locally, the `load_dataset` function takes a lot longer than it should take to fetch the dataset from disk (or memory).
This is particularly noticeable for smaller datasets. For example, wikitex... | [
-1.2485748529434204,
-0.9309433698654175,
-0.7487165331840515,
1.4467313289642334,
-0.19418032467365265,
-1.2446805238723755,
0.14992734789848328,
-1.0850646495819092,
1.6498594284057617,
-0.8785093426704407,
0.2918948233127594,
-1.6871237754821777,
0.08295769989490509,
-0.5667169690132141... |
https://github.com/huggingface/datasets/issues/5498 | TypeError: 'bool' object is not iterable when filtering a datasets.arrow_dataset.Dataset | Hi! Instead of a single boolean, your filter function should return an iterable (of booleans) in the batched mode like so:
```python
train_dataset = train_dataset.filter(
function=lambda batch: [image is not None for image in batch["image"]],
batched=True,
batch_size=10)
```
... | ### Describe the bug
Hi,
Thanks for the amazing work on the library!
**Describe the bug**
I think I might have noticed a small bug in the filter method.
Having loaded a dataset using `load_dataset`, when I try to filter out empty entries with `batched=True`, I get a TypeError.
### Steps to reproduce the ... | 311 | 74 | TypeError: 'bool' object is not iterable when filtering a datasets.arrow_dataset.Dataset
### Describe the bug
Hi,
Thanks for the amazing work on the library!
**Describe the bug**
I think I might have noticed a small bug in the filter method.
Having loaded a dataset using `load_dataset`, when I try to fi... | [
-1.216828465461731,
-0.9283676147460938,
-0.6916117072105408,
1.3856755495071411,
-0.16936609148979187,
-1.228087306022644,
0.19421380758285522,
-1.0567854642868042,
1.6992367506027222,
-0.7671883702278137,
0.27118659019470215,
-1.6829215288162231,
0.00794027466326952,
-0.5832213759422302,... |
https://github.com/huggingface/datasets/issues/5496 | Add a `reduce` method | Hi! Sure, feel free to open a PR, so we can see the API you have in mind. | ### Feature request
Right now the `Dataset` class implements `map()` and `filter()`, but leaves out the third functional idiom popular among Python users: `reduce`.
### Motivation
A `reduce` method is often useful when calculating dataset statistics, for example, the occurrence of a particular n-gram or the average... | 312 | 18 | Add a `reduce` method
### Feature request
Right now the `Dataset` class implements `map()` and `filter()`, but leaves out the third functional idiom popular among Python users: `reduce`.
### Motivation
A `reduce` method is often useful when calculating dataset statistics, for example, the occurrence of a particul... | [
-1.073635220527649,
-0.8911888599395752,
-0.8345375061035156,
1.4596879482269287,
-0.22490829229354858,
-1.426038384437561,
0.18532468378543854,
-1.1373742818832397,
1.819708228111267,
-0.8317157626152039,
0.34981077909469604,
-1.7550373077392578,
0.06418803334236145,
-0.774227499961853,
... |
https://github.com/huggingface/datasets/issues/5495 | to_tf_dataset fails with datetime UTC columns even if not included in columns argument | Hi! This is indeed a bug in our zero-copy logic.
To fix it, instead of the line:
https://github.com/huggingface/datasets/blob/7cfac43b980ab9e4a69c2328f085770996323005/src/datasets/features/features.py#L702
we should have:
```python
return pa.types.is_primitive(pa_type) and not (pa.types.is_boolean(pa_type) or ... | ### Describe the bug
There appears to be some eager behavior in `to_tf_dataset` that runs against every column in a dataset even if they aren't included in the columns argument. This is problematic with datetime UTC columns due to them not working with zero copy. If I don't have UTC information in my datetime column... | 313 | 30 | to_tf_dataset fails with datetime UTC columns even if not included in columns argument
### Describe the bug
There appears to be some eager behavior in `to_tf_dataset` that runs against every column in a dataset even if they aren't included in the columns argument. This is problematic with datetime UTC columns due ... | [
-1.2224701642990112,
-0.8903769850730896,
-0.6871119737625122,
1.498617172241211,
-0.20761793851852417,
-1.198106288909912,
0.25135576725006104,
-1.1399424076080322,
1.5775479078292847,
-0.847285807132721,
0.34940415620803833,
-1.6227664947509766,
0.07213814556598663,
-0.6189757585525513,
... |
https://github.com/huggingface/datasets/issues/5494 | Update audio installation doc page | Totally agree, the docs should be in sync with our code.
Indeed to avoid confusing users, I think we should have updated the docs at the same time as this PR:
- #5167 | Our [installation documentation page](https://huggingface.co/docs/datasets/installation#audio) says that one can use Datasets for mp3 only with `torchaudio<0.12`. `torchaudio>0.12` is actually supported too but requires a specific version of ffmpeg which is not easily installed on all linux versions but there is a cust... | 314 | 33 | Update audio installation doc page
Our [installation documentation page](https://huggingface.co/docs/datasets/installation#audio) says that one can use Datasets for mp3 only with `torchaudio<0.12`. `torchaudio>0.12` is actually supported too but requires a specific version of ffmpeg which is not easily installed on a... | [
-1.1686840057373047,
-0.8682674765586853,
-0.8277369737625122,
1.3871928453445435,
-0.02029503881931305,
-1.3370705842971802,
0.0879618376493454,
-1.0775885581970215,
1.6758482456207275,
-0.8146716952323914,
0.37178805470466614,
-1.7299107313156128,
-0.059250444173812866,
-0.55361360311508... |
https://github.com/huggingface/datasets/issues/5494 | Update audio installation doc page | No, @polinaeterna, nothing to be sorry about.
My comment was for all of us datasets team, as a reminder: when making a PR, but also when reviewing some other's PR, we should not forget to update the corresponding docstring and doc pages. It is something we can improve if we help each other in reminding about it... :... | Our [installation documentation page](https://huggingface.co/docs/datasets/installation#audio) says that one can use Datasets for mp3 only with `torchaudio<0.12`. `torchaudio>0.12` is actually supported too but requires a specific version of ffmpeg which is not easily installed on all linux versions but there is a cust... | 314 | 58 | Update audio installation doc page
Our [installation documentation page](https://huggingface.co/docs/datasets/installation#audio) says that one can use Datasets for mp3 only with `torchaudio<0.12`. `torchaudio>0.12` is actually supported too but requires a specific version of ffmpeg which is not easily installed on a... | [
-1.161424160003662,
-0.8929922580718994,
-0.7799407839775085,
1.3978607654571533,
-0.04185804724693298,
-1.3138270378112793,
0.07529287785291672,
-1.076865792274475,
1.648038387298584,
-0.7778950333595276,
0.37946653366088867,
-1.7578752040863037,
-0.005315739661455154,
-0.5952066779136658... |
https://github.com/huggingface/datasets/issues/5494 | Update audio installation doc page | @polinaeterna I think we can close this issue now as we no longer use `torchaudio` for decoding. | Our [installation documentation page](https://huggingface.co/docs/datasets/installation#audio) says that one can use Datasets for mp3 only with `torchaudio<0.12`. `torchaudio>0.12` is actually supported too but requires a specific version of ffmpeg which is not easily installed on all linux versions but there is a cust... | 314 | 17 | Update audio installation doc page
Our [installation documentation page](https://huggingface.co/docs/datasets/installation#audio) says that one can use Datasets for mp3 only with `torchaudio<0.12`. `torchaudio>0.12` is actually supported too but requires a specific version of ffmpeg which is not easily installed on a... | [
-1.1674246788024902,
-0.8407456278800964,
-0.7989696860313416,
1.3934714794158936,
-0.024220071732997894,
-1.3427902460098267,
0.12882566452026367,
-1.08321213722229,
1.6402385234832764,
-0.8051636219024658,
0.40123072266578674,
-1.7056013345718384,
-0.048283278942108154,
-0.54516422748565... |
https://github.com/huggingface/datasets/issues/5492 | Push_to_hub in a pull request | Assigned to myself and will get to it in the next week, but if someone finds this issue annoying and wants to submit a PR before I do, just ping me here and I'll reassign :). | Right now `ds.push_to_hub()` can push a dataset on `main` or on a new branch with `branch=`, but there is no way to open a pull request. Even passing `branch=refs/pr/x` doesn't seem to work: it tries to create a branch with that name
cc @nateraw
It should be possible to tweak the use of `huggingface_hub` in `pus... | 315 | 36 | Push_to_hub in a pull request
Right now `ds.push_to_hub()` can push a dataset on `main` or on a new branch with `branch=`, but there is no way to open a pull request. Even passing `branch=refs/pr/x` doesn't seem to work: it tries to create a branch with that name
cc @nateraw
It should be possible to tweak the ... | [
-1.1209107637405396,
-0.8946842551231384,
-0.8350309729576111,
1.5535743236541748,
-0.20843717455863953,
-1.2693958282470703,
0.1250711977481842,
-1.1027363538742065,
1.681578516960144,
-0.9570409655570984,
0.4714111387729645,
-1.5528804063796997,
0.0343179851770401,
-0.6533536314964294,
... |
https://github.com/huggingface/datasets/issues/5488 | Error loading MP3 files from CommonVoice | Hi @kradonneoh, thanks for reporting.
Please note that to work with audio datasets (and specifically with MP3 files) we have detailed installation instructions in our docs: https://huggingface.co/docs/datasets/installation#audio
- one of the requirements is torchaudio<0.12.0
Let us know if the problem persists a... | ### Describe the bug
When loading a CommonVoice dataset with `datasets==2.9.0` and `torchaudio>=0.12.0`, I get an error reading the audio arrays:
```python
---------------------------------------------------------------------------
LibsndfileError Traceback (most recent call last)
~/.l... | 316 | 45 | Error loading MP3 files from CommonVoice
### Describe the bug
When loading a CommonVoice dataset with `datasets==2.9.0` and `torchaudio>=0.12.0`, I get an error reading the audio arrays:
```python
---------------------------------------------------------------------------
LibsndfileError ... | [
-1.2167824506759644,
-0.7806340456008911,
-0.4928833842277527,
1.4399757385253906,
-0.05507618933916092,
-1.3829612731933594,
0.13091428577899933,
-0.8968741297721863,
1.5224993228912354,
-0.8785139322280884,
0.40658533573150635,
-1.6223372220993042,
0.05925789102911949,
-0.610502600669860... |
https://github.com/huggingface/datasets/issues/5488 | Error loading MP3 files from CommonVoice | I saw that and have followed it (hence the Expected Behavior section of the bug report).
Is there no intention of updating to the latest version? It does limit the version of `torch` I can use, which isn’t ideal. | ### Describe the bug
When loading a CommonVoice dataset with `datasets==2.9.0` and `torchaudio>=0.12.0`, I get an error reading the audio arrays:
```python
---------------------------------------------------------------------------
LibsndfileError Traceback (most recent call last)
~/.l... | 316 | 39 | Error loading MP3 files from CommonVoice
### Describe the bug
When loading a CommonVoice dataset with `datasets==2.9.0` and `torchaudio>=0.12.0`, I get an error reading the audio arrays:
```python
---------------------------------------------------------------------------
LibsndfileError ... | [
-1.2167824506759644,
-0.7806340456008911,
-0.4928833842277527,
1.4399757385253906,
-0.05507618933916092,
-1.3829612731933594,
0.13091428577899933,
-0.8968741297721863,
1.5224993228912354,
-0.8785139322280884,
0.40658533573150635,
-1.6223372220993042,
0.05925789102911949,
-0.610502600669860... |
https://github.com/huggingface/datasets/issues/5488 | Error loading MP3 files from CommonVoice | @kradonneoh hey! actually with `ffmpeg4` loading of mp3 files should work, so this is a not expected behavior and we need to investigate it. It works on my side with `torchaudio==0.13` and `ffmpeg==4.2.7`. Which `torchaudio` version do you use?
`datasets` should support decoding of mp3 files with `torchaudio` when i... | ### Describe the bug
When loading a CommonVoice dataset with `datasets==2.9.0` and `torchaudio>=0.12.0`, I get an error reading the audio arrays:
```python
---------------------------------------------------------------------------
LibsndfileError Traceback (most recent call last)
~/.l... | 316 | 137 | Error loading MP3 files from CommonVoice
### Describe the bug
When loading a CommonVoice dataset with `datasets==2.9.0` and `torchaudio>=0.12.0`, I get an error reading the audio arrays:
```python
---------------------------------------------------------------------------
LibsndfileError ... | [
-1.2167824506759644,
-0.7806340456008911,
-0.4928833842277527,
1.4399757385253906,
-0.05507618933916092,
-1.3829612731933594,
0.13091428577899933,
-0.8968741297721863,
1.5224993228912354,
-0.8785139322280884,
0.40658533573150635,
-1.6223372220993042,
0.05925789102911949,
-0.610502600669860... |
https://github.com/huggingface/datasets/issues/5487 | Incorrect filepath for dill module | Hi! The correct path is still `dill._dill.XXXX` in the latest release. What do you get when you run `python -c "import dill; print(dill.__version__)"` in your environment? | ### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import datasets
File "/Users/avivbrokman/... | 317 | 26 | Incorrect filepath for dill module
### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import da... | [
-1.1599488258361816,
-0.8837870359420776,
-0.7099396586418152,
1.3716564178466797,
-0.08292015641927719,
-1.3277747631072998,
0.0947943702340126,
-1.1010955572128296,
1.5574061870574951,
-0.6909691095352173,
0.2057599425315857,
-1.688206672668457,
-0.19245323538780212,
-0.4060768485069275,... |
https://github.com/huggingface/datasets/issues/5487 | Incorrect filepath for dill module | `0.3.6` I feel like that's bad news, because it's probably not the issue.
My mistake, about the wrong path guess. I think I didn't notice that the first `dill` in the path isn't supposed to be included in the path specification in python.
<img width="146" alt="Screen Shot 2023-01-31 at 12 58 32 PM" src="https://us... | ### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import datasets
File "/Users/avivbrokman/... | 317 | 54 | Incorrect filepath for dill module
### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import da... | [
-1.1599488258361816,
-0.8837870359420776,
-0.7099396586418152,
1.3716564178466797,
-0.08292015641927719,
-1.3277747631072998,
0.0947943702340126,
-1.1010955572128296,
1.5574061870574951,
-0.6909691095352173,
0.2057599425315857,
-1.688206672668457,
-0.19245323538780212,
-0.4060768485069275,... |
https://github.com/huggingface/datasets/issues/5487 | Incorrect filepath for dill module | Hi, @avivbrokman, this issue you report appeared only with old versions of dill. See:
- #288
Are you sure you are in the right Python environment?
- Please note that Jupyter (where I guess you get the error) may have multiple execution backends (IPython kernels) that might be different from the Python environment ... | ### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import datasets
File "/Users/avivbrokman/... | 317 | 85 | Incorrect filepath for dill module
### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import da... | [
-1.1599488258361816,
-0.8837870359420776,
-0.7099396586418152,
1.3716564178466797,
-0.08292015641927719,
-1.3277747631072998,
0.0947943702340126,
-1.1010955572128296,
1.5574061870574951,
-0.6909691095352173,
0.2057599425315857,
-1.688206672668457,
-0.19245323538780212,
-0.4060768485069275,... |
https://github.com/huggingface/datasets/issues/5487 | Incorrect filepath for dill module | I'm using spyder, and I am still getting `0.3.6` for `dill`, so unfortunately #288 isn't applicable, I think. However, I found something odd that I believe is a clue:
```
import inspect
import dill
inspect.getfile(dill)
>>> '/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/dill/__init__.py'
```
... | ### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import datasets
File "/Users/avivbrokman/... | 317 | 152 | Incorrect filepath for dill module
### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import da... | [
-1.1599488258361816,
-0.8837870359420776,
-0.7099396586418152,
1.3716564178466797,
-0.08292015641927719,
-1.3277747631072998,
0.0947943702340126,
-1.1010955572128296,
1.5574061870574951,
-0.6909691095352173,
0.2057599425315857,
-1.688206672668457,
-0.19245323538780212,
-0.4060768485069275,... |
https://github.com/huggingface/datasets/issues/5487 | Incorrect filepath for dill module | As I said above, I guess this is not a problem with `datasets`. I think you have different Python environments: one with the new dill version (the one you get while using pip) and other with the old dill version (the one where you get the AttributeError).
You should update `dill` in the Python environment you are us... | ### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import datasets
File "/Users/avivbrokman/... | 317 | 76 | Incorrect filepath for dill module
### Describe the bug
I installed the `datasets` package and when I try to `import` it, I get the following error:
```
Traceback (most recent call last):
File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module>
import da... | [
-1.1599488258361816,
-0.8837870359420776,
-0.7099396586418152,
1.3716564178466797,
-0.08292015641927719,
-1.3277747631072998,
0.0947943702340126,
-1.1010955572128296,
1.5574061870574951,
-0.6909691095352173,
0.2057599425315857,
-1.688206672668457,
-0.19245323538780212,
-0.4060768485069275,... |
https://github.com/huggingface/datasets/issues/5486 | Adding `sep` to TextConfig | Hi @omar-araboghli, thanks for your proposal.
Have you tried to use "csv" loader instead of "text"? That already has a `sep` argument. | I have a local a `.txt` file that follows the `CONLL2003` format which I need to load using `load_script`. However, by using `sample_by='line'`, one can only split the dataset into lines without splitting each line into columns. Would it be reasonable to add a `sep` argument in combination with `sample_by='paragraph'` ... | 318 | 22 | Adding `sep` to TextConfig
I have a local a `.txt` file that follows the `CONLL2003` format which I need to load using `load_script`. However, by using `sample_by='line'`, one can only split the dataset into lines without splitting each line into columns. Would it be reasonable to add a `sep` argument in combination... | [
-1.1810396909713745,
-0.8709943294525146,
-0.7613259553909302,
1.5773694515228271,
-0.22678525745868683,
-1.1137773990631104,
0.13493561744689941,
-1.102150559425354,
1.7744081020355225,
-0.763896107673645,
0.3840358853340149,
-1.6877012252807617,
0.06789879500865936,
-0.7430889010429382,
... |
https://github.com/huggingface/datasets/issues/5486 | Adding `sep` to TextConfig | Hi @albertvillanova, thanks for the quick response!
Indeed, I have been trying to use `csv` instead of `text`. However I am still not able to define range of rows as one sequence, that is achievable with passing `sample_by='paragraph'` to the `TextConfig`
For instance, the below code
```python
import datasets... | I have a local a `.txt` file that follows the `CONLL2003` format which I need to load using `load_script`. However, by using `sample_by='line'`, one can only split the dataset into lines without splitting each line into columns. Would it be reasonable to add a `sep` argument in combination with `sample_by='paragraph'` ... | 318 | 116 | Adding `sep` to TextConfig
I have a local a `.txt` file that follows the `CONLL2003` format which I need to load using `load_script`. However, by using `sample_by='line'`, one can only split the dataset into lines without splitting each line into columns. Would it be reasonable to add a `sep` argument in combination... | [
-1.1810396909713745,
-0.8709943294525146,
-0.7613259553909302,
1.5773694515228271,
-0.22678525745868683,
-1.1137773990631104,
0.13493561744689941,
-1.102150559425354,
1.7744081020355225,
-0.763896107673645,
0.3840358853340149,
-1.6877012252807617,
0.06789879500865936,
-0.7430889010429382,
... |
https://github.com/huggingface/datasets/issues/5482 | Reload features from Parquet metadata | I'd be happy to have a look, if nobody else has started working on this yet @lhoestq.
It seems to me that for the `arrow` format features are currently attached as metadata [in `datasets.arrow_writer`](https://github.com/huggingface/datasets/blob/5f810b7011a8a4ab077a1847c024d2d9e267b065/src/datasets/arrow_writer.py... | The idea would be to allow this :
```python
ds.to_parquet("my_dataset/ds.parquet")
reloaded = load_dataset("my_dataset")
assert ds.features == reloaded.features
```
And it should also work with Image and Audio types (right now they're reloaded as a dict type)
This can be implemented by storing and reading th... | 320 | 66 | Reload features from Parquet metadata
The idea would be to allow this :
```python
ds.to_parquet("my_dataset/ds.parquet")
reloaded = load_dataset("my_dataset")
assert ds.features == reloaded.features
```
And it should also work with Image and Audio types (right now they're reloaded as a dict type)
This can ... | [
-1.166128158569336,
-0.8522657752037048,
-0.7657282948493958,
1.5049726963043213,
-0.08720782399177551,
-1.2787712812423706,
0.06049301102757454,
-1.0508406162261963,
1.6568968296051025,
-0.6973925232887268,
0.29063931107521057,
-1.6855441331863403,
0.01580018177628517,
-0.5152422189712524... |
https://github.com/huggingface/datasets/issues/5482 | Reload features from Parquet metadata | Thanks @MFreidank ! That's correct :)
Reading the metadata to infer the features can be ideally done in the `parquet.py` file in `packaged_builder` when a parquet file is read. You can cast the arrow table to the schema you get from the features.arrow_schema | The idea would be to allow this :
```python
ds.to_parquet("my_dataset/ds.parquet")
reloaded = load_dataset("my_dataset")
assert ds.features == reloaded.features
```
And it should also work with Image and Audio types (right now they're reloaded as a dict type)
This can be implemented by storing and reading th... | 320 | 43 | Reload features from Parquet metadata
The idea would be to allow this :
```python
ds.to_parquet("my_dataset/ds.parquet")
reloaded = load_dataset("my_dataset")
assert ds.features == reloaded.features
```
And it should also work with Image and Audio types (right now they're reloaded as a dict type)
This can ... | [
-1.197356939315796,
-0.8986155390739441,
-0.824112057685852,
1.652525544166565,
-0.22901135683059692,
-1.4266657829284668,
0.11700111627578735,
-1.1063463687896729,
1.7173707485198975,
-0.9330558776855469,
0.4493255317211151,
-1.6518818140029907,
0.18065370619297028,
-0.5315032005310059,
... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | Hi ! Sure :) you can comment `#self-assign` to assign yourself to this issue.
I can give you some pointers to get started:
`load_dataset` works roughly this way:
1. it instantiate a dataset builder using `load_dataset_builder()`
2. the builder download and prepare the dataset as Arrow files in the cache using `... | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 146 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.1753803491592407,
-0.8470456004142761,
-0.704535961151123,
1.5490326881408691,
-0.16414037346839905,
-1.296331524848938,
0.1660868376493454,
-1.0722166299819946,
1.7976151704788208,
-0.8254375457763672,
0.452565997838974,
-1.6749439239501953,
0.1041148230433464,
-0.5741820335388184,
-... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | This use-case is a bit specific, and `load_dataset` already has enough parameters (plus, `streaming=True` also returns an iterable dataset, so we would have to explain the difference), so I think it would be better to add `IterableDataset.from_file` to the API (more flexible and aligned with the goal from https://githu... | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 50 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.2134650945663452,
-0.9165612459182739,
-0.8333216905593872,
1.5663814544677734,
-0.14164571464061737,
-1.298644781112671,
0.10598117113113403,
-0.971045732498169,
1.614971399307251,
-0.9010605216026306,
0.4977528154850006,
-1.7379769086837769,
0.028258066624403,
-0.5117005705833435,
-... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | > This use-case is a bit specific
This allows to use `datasets` for large scale training where map-style datasets are too slow and use too much memory in PyTorch. So I would still consider adding it.
Alternatively we could add this feature one level bellow:
```python
builder = load_dataset_builder(...)
builder... | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 54 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.2192336320877075,
-0.8842707276344299,
-0.701069176197052,
1.5191925764083862,
-0.20752127468585968,
-1.333116054534912,
0.18966422975063324,
-1.0095540285110474,
1.7087229490280151,
-0.928950309753418,
0.45042845606803894,
-1.6462763547897339,
0.07562471926212311,
-0.5040363669395447,
... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | Yes, I see how this can be useful. Still, I think `Dataset.to_iterable` + `IterableDataset.from_file` would be much cleaner in terms of the API design (and more flexible since `load_dataset` can only access the "initial" (unprocessed) version of a dataset).
And since it can be tricky to manually find the "initial" v... | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 75 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.1892433166503906,
-0.9370976090431213,
-0.7828351259231567,
1.5670956373214722,
-0.22921042144298553,
-1.299664855003357,
0.1782594472169876,
-0.9872884750366211,
1.7259507179260254,
-0.927904486656189,
0.44232386350631714,
-1.7147488594055176,
0.05562882125377655,
-0.508151650428772,
... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | > This allows to use datasets for large scale training where map-style datasets are too slow and use too much memory in PyTorch.
I second that. e.g. In my last experiment Oscar-en uses 16GB RSS RAM per process and when using multiple processes the host quickly runs out cpu memory. | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 50 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.277882695198059,
-0.9373113512992859,
-0.7740324139595032,
1.5673296451568604,
-0.2550923824310303,
-1.2900761365890503,
0.0703420341014862,
-1.0109537839889526,
1.6413134336471558,
-0.8726053833961487,
0.3990180492401123,
-1.6514431238174438,
0.05712546035647392,
-0.41546377539634705,
... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | >And since it can be tricky to manually find the "initial" version of a dataset in the cache, maybe load_dataset could return an iterable dataset streamed from the cache if streaming=True and the cache is up-to-date.
This is exactly the need on JeanZay (HPC) - I have the dataset cache ready, but the compute node is ... | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 81 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.2290716171264648,
-0.9090748429298401,
-0.806638777256012,
1.565148115158081,
-0.22343139350414276,
-1.2788716554641724,
0.0906263217329979,
-1.0198851823806763,
1.6406433582305908,
-0.9181777834892273,
0.4443301260471344,
-1.7047456502914429,
0.11137185990810394,
-0.4876655042171478,
... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | > Yes, I see how this can be useful. Still, I think Dataset.to_iterable + IterableDataset.from_file would be much cleaner in terms of the API design (and more flexible since load_dataset can only access the "initial" (unprocessed) version of a dataset).
I like `IterableDataset.from_file` as well. On the other hand `... | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 172 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.2615983486175537,
-0.9546722173690796,
-0.732764482498169,
1.4907504320144653,
-0.2046111524105072,
-1.2499337196350098,
0.11575119197368622,
-1.0419470071792603,
1.6616331338882446,
-0.8744317293167114,
0.38723519444465637,
-1.6974396705627441,
0.015965620055794716,
-0.4623053073883056... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | > I was suggesting a different parameter just to make explicit the difference between
MosaicML's `streaming` library does the same (tries to stream from the local cache if possible), so logging a message should be explicit enough :). | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 38 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.2099270820617676,
-0.8309966325759888,
-0.8755233883857727,
1.6023204326629639,
-0.25128743052482605,
-1.2368806600570679,
0.1659827083349228,
-1.0083129405975342,
1.6647257804870605,
-0.7686957120895386,
0.46065738797187805,
-1.6777592897415161,
0.03902047872543335,
-0.4802227318286896... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | Hi Both! It has been a while since my first issue so I am gonna go for this one ! #self-assign | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 21 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.2416939735412598,
-0.8539114594459534,
-0.8113563060760498,
1.5327818393707275,
-0.20977532863616943,
-1.3059695959091187,
0.1383461207151413,
-0.956865668296814,
1.6455330848693848,
-0.8499376773834229,
0.5143492817878723,
-1.693522572517395,
0.027579037472605705,
-0.46484267711639404,... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | https://github.com/huggingface/datasets/pull/5821 should be helpful to implement `IterableDataset.from_file`, since it defines a new ArrowExamplesIterable that takes an Arrow tables generator function (e.g. from a file) and can be used in an IterableDataset | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 31 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.205568790435791,
-0.9849755764007568,
-0.8147204518318176,
1.6263556480407715,
-0.17988377809524536,
-1.2975077629089355,
0.07710582762956619,
-0.9588920474052429,
1.5921838283538818,
-0.8183634877204895,
0.530085027217865,
-1.6791902780532837,
0.0354824997484684,
-0.4295259714126587,
... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | So what's recommanded usage of `IterableDataset.from_file` and `load_dataset`? How about I have multiple arrow files and `load_dataset` is often convenient to handle that. | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 23 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.1906013488769531,
-0.9140706658363342,
-0.8065270185470581,
1.6290780305862427,
-0.19189764559268951,
-1.3308740854263306,
0.25316500663757324,
-0.9579043984413147,
1.6622309684753418,
-0.9153589010238647,
0.492199182510376,
-1.7103694677352905,
-0.005908402614295483,
-0.503328204154968... |
https://github.com/huggingface/datasets/issues/5481 | Load a cached dataset as iterable | If you have multiple Arrow files you can load them using
```python
from datasets import load_dataset
data_files = {"train": ["path/to/0.arrow", "path/to/1.arrow", ..., "path/to/n.arrow"]}
ds = load_dataset("arrow", data_files=data_files, streaming=True)
```
This is equivalent to calling `IterableDataset.f... | The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when streaming=True | 321 | 37 | Load a cached dataset as iterable
The idea would be to allow something like
```python
ds = load_dataset("c4", "en", as_iterable=True)
```
To be used to train models. It would load an IterableDataset from the cached Arrow files.
Cc @stas00
Edit : from the discussions we may load from cache when stream... | [
-1.1421035528182983,
-0.9280318021774292,
-0.7541816234588623,
1.5897581577301025,
-0.18992190062999725,
-1.205849528312683,
0.18088871240615845,
-1.0219354629516602,
1.6403167247772217,
-0.8541223406791687,
0.4649083614349365,
-1.70931077003479,
0.02902182564139366,
-0.5327165722846985,
... |
https://github.com/huggingface/datasets/issues/5477 | Unpin sqlalchemy once issue is fixed | @albertvillanova It looks like that issue has been fixed so I made a PR to unpin sqlalchemy! | Once the source issue is fixed:
- pandas-dev/pandas#51015
we should revert the pin introduced in:
- #5476 | 322 | 17 | Unpin sqlalchemy once issue is fixed
Once the source issue is fixed:
- pandas-dev/pandas#51015
we should revert the pin introduced in:
- #5476
@albertvillanova It looks like that issue has been fixed so I made a PR to unpin sqlalchemy! | [
-1.0940613746643066,
-0.8565547466278076,
-0.8060352206230164,
1.3778626918792725,
-0.22861014306545258,
-1.200943946838379,
-0.0018995129503309727,
-0.9439231157302856,
1.7429156303405762,
-0.7648454904556274,
0.4368641674518585,
-1.7347999811172485,
-0.07652505487203598,
-0.6610858440399... |
https://github.com/huggingface/datasets/issues/5477 | Unpin sqlalchemy once issue is fixed | The source issue:
- https://github.com/pandas-dev/pandas/issues/40686
has been fixed:
- https://github.com/pandas-dev/pandas/pull/48576
The fix was released yesterday (2023-04-03) only in `pandas-2.0.0`:
- https://github.com/pandas-dev/pandas/releases/tag/v2.0.0
but it will not be back-ported to `pandas-1`:... | Once the source issue is fixed:
- pandas-dev/pandas#51015
we should revert the pin introduced in:
- #5476 | 322 | 62 | Unpin sqlalchemy once issue is fixed
Once the source issue is fixed:
- pandas-dev/pandas#51015
we should revert the pin introduced in:
- #5476
The source issue:
- https://github.com/pandas-dev/pandas/issues/40686
has been fixed:
- https://github.com/pandas-dev/pandas/pull/48576
The fix was released yes... | [
-1.0830845832824707,
-0.7927930355072021,
-0.7707298994064331,
1.4383342266082764,
0.002129746600985527,
-1.3204591274261475,
0.002570660784840584,
-0.8994882702827454,
1.5556992292404175,
-0.6449288129806519,
0.34706997871398926,
-1.695032000541687,
-0.0441645011305809,
-0.686315953731536... |
https://github.com/huggingface/datasets/issues/5475 | Dataset scan time is much slower than using native arrow | Hi ! In your code you only iterate on the Arrow buffers - you don't actually load the data as python objects. For a fair comparison, you can modify your code using:
```diff
- for _ in range(0, len(table), bsz):
- _ = {k:table[k][_ : _ + bsz] for k in cols}
+ for _ in range(0, len(table), bsz):
+ ... | ### Describe the bug
I'm basically running the same scanning experiment from the tutorials https://huggingface.co/course/chapter5/4?fw=pt except now I'm comparing to a native pyarrow version.
I'm finding that the native pyarrow approach is much faster (2 orders of magnitude). Is there something I'm missing that exp... | 323 | 85 | Dataset scan time is much slower than using native arrow
### Describe the bug
I'm basically running the same scanning experiment from the tutorials https://huggingface.co/course/chapter5/4?fw=pt except now I'm comparing to a native pyarrow version.
I'm finding that the native pyarrow approach is much faster (2 or... | [
-1.2639212608337402,
-0.9404871463775635,
-0.7692417502403259,
1.4909127950668335,
-0.08855834603309631,
-1.222729206085205,
0.08676759898662567,
-1.0652048587799072,
1.5927857160568237,
-0.7779014110565186,
0.34533968567848206,
-1.6053789854049683,
0.10426580905914307,
-0.5708794593811035... |
https://github.com/huggingface/datasets/issues/5475 | Dataset scan time is much slower than using native arrow | Ah I see, datasets is implicitly making this conversion. Thanks for pointing that out!
If it's not too much, I would also suggest updating some of your docs with the same `.to_pylist()` conversion in the code snippet that follows [here](https://huggingface.co/course/chapter5/4?fw=pt#:~:text=let%E2%80%99s%20run%20a%2... | ### Describe the bug
I'm basically running the same scanning experiment from the tutorials https://huggingface.co/course/chapter5/4?fw=pt except now I'm comparing to a native pyarrow version.
I'm finding that the native pyarrow approach is much faster (2 orders of magnitude). Is there something I'm missing that exp... | 323 | 40 | Dataset scan time is much slower than using native arrow
### Describe the bug
I'm basically running the same scanning experiment from the tutorials https://huggingface.co/course/chapter5/4?fw=pt except now I'm comparing to a native pyarrow version.
I'm finding that the native pyarrow approach is much faster (2 or... | [
-1.24544095993042,
-0.9433806538581848,
-0.7787543535232544,
1.4426261186599731,
-0.01119820773601532,
-1.1739522218704224,
0.03638889268040657,
-1.0511670112609863,
1.514532446861267,
-0.8187382221221924,
0.3589247167110443,
-1.6049509048461914,
0.019263528287410736,
-0.4325990080833435,
... |
https://github.com/huggingface/datasets/issues/5475 | Dataset scan time is much slower than using native arrow | This code snippet shows `datasets` code that reads the Arrow data as python objects already, there is no need to add to_pylist. Or were you thinking about something else ? | ### Describe the bug
I'm basically running the same scanning experiment from the tutorials https://huggingface.co/course/chapter5/4?fw=pt except now I'm comparing to a native pyarrow version.
I'm finding that the native pyarrow approach is much faster (2 orders of magnitude). Is there something I'm missing that exp... | 323 | 30 | Dataset scan time is much slower than using native arrow
### Describe the bug
I'm basically running the same scanning experiment from the tutorials https://huggingface.co/course/chapter5/4?fw=pt except now I'm comparing to a native pyarrow version.
I'm finding that the native pyarrow approach is much faster (2 or... | [
-1.2313355207443237,
-0.9755416512489319,
-0.8328843712806702,
1.4891432523727417,
-0.08112998306751251,
-1.2118287086486816,
0.10704300552606583,
-1.0513474941253662,
1.6313530206680298,
-0.8384964466094971,
0.3347957730293274,
-1.6132891178131104,
0.05242260918021202,
-0.5046954154968262... |
https://github.com/huggingface/datasets/issues/5474 | Column project operation on `datasets.Dataset` | Hi ! This would be a nice addition indeed :) This sounds like a duplicate of https://github.com/huggingface/datasets/issues/5468
> Not sure. Some of my PRs are still open and some do not have any discussions.
Sorry to hear that, feel free to ping me on those PRs | ### Feature request
There is no operation to select a subset of columns of original dataset. Expected API follows.
```python
a = Dataset.from_dict({
'int': [0, 1, 2]
'char': ['a', 'b', 'c'],
'none': [None] * 3,
})
b = a.project('int', 'char') # usually, .select()
print(a.column_names) # std... | 324 | 46 | Column project operation on `datasets.Dataset`
### Feature request
There is no operation to select a subset of columns of original dataset. Expected API follows.
```python
a = Dataset.from_dict({
'int': [0, 1, 2]
'char': ['a', 'b', 'c'],
'none': [None] * 3,
})
b = a.project('int', 'char') # u... | [
-1.145315408706665,
-0.9957084059715271,
-0.7484076023101807,
1.6048171520233154,
-0.1513364017009735,
-1.22935152053833,
0.025476552546024323,
-1.0114284753799438,
1.7112200260162354,
-0.7468509674072266,
0.2529119849205017,
-1.71471107006073,
-0.05852080509066582,
-0.7319914698600769,
... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | Hi! I agree it would be nice to have a method like that. Instead of `keep_columns`, we can name it `select_columns` to be more aligned with PyArrow's naming convention (`pa.Table.select`). | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 30 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.1291453838348389,
-0.8903515338897705,
-0.7809796929359436,
1.5449533462524414,
-0.03898512199521065,
-1.2388007640838623,
0.24319131672382355,
-1.049644112586975,
1.6578035354614258,
-0.8081666231155396,
0.3863639533519745,
-1.6128418445587158,
0.09286542981863022,
-0.7118729948997498,... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | Hi, I am a newbie to open source and would like to contribute. @mariosasko can I take up this issue ? | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 21 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.154079556465149,
-0.9099525809288025,
-0.8128042817115784,
1.554242491722107,
-0.06166444718837738,
-1.2450250387191772,
0.19262157380580902,
-1.0791751146316528,
1.5956411361694336,
-0.8156757950782776,
0.38709521293640137,
-1.5716898441314697,
0.083161361515522,
-0.7508717775344849,
... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | Hey, I also want to work on this issue I am a newbie to open source. | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 16 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.1514225006103516,
-0.914588987827301,
-0.8264305591583252,
1.5600435733795166,
-0.06846963614225388,
-1.2471632957458496,
0.19848120212554932,
-1.0974475145339966,
1.606711506843567,
-0.8056549429893494,
0.3895787000656128,
-1.5868126153945923,
0.08885469287633896,
-0.7562308311462402,
... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | This sounds related to https://github.com/huggingface/datasets/issues/5474
I'm fine with `select_columns`, or we could also override `select` to also accept a list of columns maybe ? | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 24 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.1184251308441162,
-0.908778727054596,
-0.8135619759559631,
1.5733529329299927,
-0.02999010682106018,
-1.2494052648544312,
0.2329113781452179,
-1.049363613128662,
1.6241936683654785,
-0.7745755910873413,
0.370052695274353,
-1.6105045080184937,
0.04719815403223038,
-0.7533395886421204,
... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | @lhoestq, I am planning to add a member function to the dataset class to perform the selection operation. Do you think its the right way to proceed? or there is a better option ? | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 34 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.1540052890777588,
-0.9093753695487976,
-0.8194781541824341,
1.5442306995391846,
-0.05323059856891632,
-1.2564054727554321,
0.1866016536951065,
-1.096221685409546,
1.6075315475463867,
-0.7959123253822327,
0.37741926312446594,
-1.6004903316497803,
0.09928056597709656,
-0.7280531525611877,... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | Unless @mariosasko thinks otherwise, I think it can go in `Dataset.select()` :)
Though some parameters like keep_in_memory, indices_cache_file_name or writer_batch_size wouldn't when selecting columns, so we would need to update the docstring as well | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 34 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.1602294445037842,
-0.8891732096672058,
-0.7769017219543457,
1.5607486963272095,
-0.04427981376647949,
-1.2542165517807007,
0.21812741458415985,
-1.0332354307174683,
1.5975195169448853,
-0.8014072179794312,
0.37465348839759827,
-1.5837072134017944,
0.07537930458784103,
-0.724290728569030... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | If someone wants to give it a shot, feel free to comment `#self-assign` and it will assign the issue to you.
Feel free to ping us here if you have questions or if we can help :) | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 37 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.160651683807373,
-0.895520031452179,
-0.8101415038108826,
1.591101050376892,
-0.07814791053533554,
-1.243514895439148,
0.18331453204154968,
-1.0582562685012817,
1.6232577562332153,
-0.8004816770553589,
0.3806670308113098,
-1.590840458869934,
0.13563409447669983,
-0.7386615872383118,
-... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | I would rather have this functionality as a separate method. IMO it's always better to be explicit than to have an API where a single method can do different/uncorrelated things (somewhat reminds me of Pandas, and there is probably a good reason why PyArrow is more rigid in this aspect). | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 50 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.1706547737121582,
-0.9165715575218201,
-0.7993765473365784,
1.5321871042251587,
-0.061782486736774445,
-1.2707600593566895,
0.18224802613258362,
-1.0692819356918335,
1.6468626260757446,
-0.820591151714325,
0.3929552733898163,
-1.61613929271698,
0.0563240572810173,
-0.7124407887458801,
... |
https://github.com/huggingface/datasets/issues/5468 | Allow opposite of remove_columns on Dataset and DatasetDict | In the end I also think it would be nice to have it as a separate method, this way we can also have it for `IterableDataset` (which can't have `select` for indices) | ### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP)
gigaspeech = gigaspeech.remove_columns(column... | 325 | 32 | Allow opposite of remove_columns on Dataset and DatasetDict
### Feature request
In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code:
```python
COLUMNS_TO_KEEP = ["text", "audio"]
all_columns = gigaspeech["train"].column_names
columns_to_remove = set(all_columns) - set(COL... | [
-1.1840059757232666,
-0.908944845199585,
-0.8097745180130005,
1.5823670625686646,
-0.048173509538173676,
-1.2653944492340088,
0.17485898733139038,
-1.0706654787063599,
1.6354033946990967,
-0.8036683797836304,
0.397324800491333,
-1.597825288772583,
0.09538523852825165,
-0.7207145690917969,
... |
https://github.com/huggingface/datasets/issues/5464 | NonMatchingChecksumError for hendrycks_test | Thanks for reporting, @sarahwie.
Please note this issue was already fixed in `datasets` 2.6.0 version:
- #5040
If you update your `datasets` version, you will be able to load the dataset:
```
pip install -U datasets
``` | ### Describe the bug
The checksum of the file has likely changed on the remote host.
### Steps to reproduce the bug
`dataset = nlp.load_dataset("hendrycks_test", "anatomy")`
### Expected behavior
no error thrown
### Environment info
- `datasets` version: 2.2.1
- Platform: macOS-13.1-arm64-arm-64bit
- Pyt... | 326 | 37 | NonMatchingChecksumError for hendrycks_test
### Describe the bug
The checksum of the file has likely changed on the remote host.
### Steps to reproduce the bug
`dataset = nlp.load_dataset("hendrycks_test", "anatomy")`
### Expected behavior
no error thrown
### Environment info
- `datasets` version: 2.2.1
... | [
-1.1270794868469238,
-0.8291107416152954,
-0.736801028251648,
1.4307700395584106,
-0.1305014193058014,
-1.2828882932662964,
0.15632571280002594,
-1.0467504262924194,
1.7284445762634277,
-0.8059728145599365,
0.35828888416290283,
-1.7001981735229492,
0.025911666452884674,
-0.604555070400238,... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Ccing @dwofk (the author of `fast-depth`).
Thanks, @awsaf49 for reporting this. I believe this is because the NYU Depth V2 shipped from `fast-depth` is already preprocessed.
If you think it might be better to have the NYU Depth V2 dataset from BTS [here](https://huggingface.co/datasets/sayakpaul/nyu_depth_v2) f... | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 56 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Good catch ! Ideally it would be nice to have the datasets in the raw form, this way users can choose whatever processing they want to apply | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 27 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | > Ccing @dwofk (the author of `fast-depth`).
>
> Thanks, @awsaf49 for reporting this. I believe this is because the NYU Depth V2 shipped from `fast-depth` is already preprocessed.
>
> If you think it might be better to have the NYU Depth V2 dataset from BTS [here](https://huggingface.co/datasets/sayakpaul/nyu_dep... | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 155 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | I think BTS. Repositories like https://github.com/vinvino02/GLPDepth usually use BTS. Also, just for clarity, the PR will be to https://huggingface.co/datasets/sayakpaul/nyu_depth_v2. Once we have worked it out, we can update the following things:
* https://github.com/huggingface/blog/pull/718
* https://huggingface... | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 55 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | @sayakpaul If I get this right I have to,
1. Create a PR on https://huggingface.co/datasets/sayakpaul/nyu_depth_v2
2. Create a PR on https://github.com/huggingface/blog
3. Create a PR on https://github.com/huggingface/datasets to update https://github.com/huggingface/datasets/blob/main/docs/source/depth_estimation.m... | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 30 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Yup opening a PR to use BTS on https://huggingface.co/datasets/sayakpaul/nyu_depth_v2 sounds good :) Thanks for the help ! | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 17 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Finally, I have found the origin of the **discretized depth map**. When I first loaded the datasets from HF I noticed it was 30GB but in DenseDepth data is only 4GB with dtype=uint8. This means data from fast-depth (before loading to HF) must have high precision. So when I tried to dig deeper by directly loading depth_... | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 270 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Thanks so much for digging into this.
Since the second solution entails changes to core datatypes in `datasets`, I think it's better to go with the first solution.
@lhoestq WDYT? | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 30 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | @sayakpaul Yes, Solution 1 requires minimal change and provides no precision loss. But I think support for `uint16` image would be a great addition as many datasets come with `uint16` image. For example [UW-Madison GI Tract Image Segmentation](https://www.kaggle.com/competitions/uw-madison-gi-tract-image-segmentation) ... | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 77 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Using Array2D makes it not practical to use to train a model - in `transformers` we expect an image type.
There is a pull request to support more precision than uint8 in Image() here: https://github.com/huggingface/datasets/pull/5365/files
we can probably merge it today and do a release right away | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 47 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Fantastic, @lhoestq!
@awsaf49 then let's wait for the PR to get merged and then take the next steps? | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 18 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | The PR adds support for uint16 which is ok for BTS if I understand correctly, would it be ok for you ? | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 22 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | If the main issue with the current version of NYU we have on the Hub is related to the precision loss stemming from `Image()`, I'd prefer if `Image()` supported float32 as well. | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 32 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | I also prefer `float32` as it offers more precision. But I'm not sure if we'll be able to visualize image with `float32` precision. | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 23 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | We could have a separate loading for the float32 one using Array2D, but I feel like it's less convenient to use due to the amount of disk space and because it's not an Image() type. That's why I think uint16 is a better solution for users | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 46 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Yep. I just tested from the PR and it works:
```python
>>> train_dataset = load_dataset("sayakpaul/nyu_depth_v2", split="train", streaming=True)
Downloading readme: 100%|██████████████████| 8.71k/8.71k [00:00<00:00, 3.60MB/s]
>>> next(iter(train_dataset))
{'... | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 120 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Great! the case is closed! This issue has been solved and I have to say, it was quite the thrill ride. I felt like Sherlock Holmes, solving a mystery and finding the bug🕵️♂️. But in all seriousness, it was a pleasure working on this issue and I'm glad we could get to the bottom of it.
On another note, should I con... | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 77 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Haha thanks Mr Holmes :p
maybe let's close this issue when we're done updating the blog post and the documentation | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 20 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | @awsaf49 thank you for your hard work!
I am a little unsure why the other links need to be updated, though. They all rely on datasets internally. | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 27 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | I think depth_map still shows discretized version. It would be nice to have corrected one.
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets/depth_est_target_viz.png" width = 300> | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 20 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Also, I think we need to make some changes in the code to visualize depth_map as it is `float32` . `plot.imshow()` supports either [0, 1] + float32 or [0. 255] + uint8 | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 32 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5461 | Discrepancy in `nyu_depth_v2` dataset | Oh yes! Do you want to start with the fixes? Please feel free to say no but I wanted to make sure your contributions are reflected properly in our doc and the blog :) | ### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-sid... | 327 | 34 | Discrepancy in `nyu_depth_v2` dataset
### Describe the bug
I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different... | [
-1.2323219776153564,
-0.9139350652694702,
-0.7461334466934204,
1.4001511335372925,
-0.17456850409507751,
-1.2431707382202148,
0.08405475318431854,
-1.0615496635437012,
1.6305798292160034,
-0.7169316411018372,
0.2860521078109741,
-1.66029691696167,
-0.011840286664664745,
-0.5992560386657715... |
https://github.com/huggingface/datasets/issues/5458 | slice split while streaming | Hi! Yes, that's correct. When `streaming` is `True`, only split names can be specified as `split`, and for slicing, you have to use `.skip`/`.take` instead.
E.g.
`load_dataset("lhoestq/demo1",revision=None, streaming=True, split="train[:3]")`
rewritten with `.skip`/`.take`:
`load_dataset("lhoestq/demo1",revisi... | ### Describe the bug
When using the `load_dataset` function with streaming set to True, slicing splits is apparently not supported.
Did I miss this in the documentation?
### Steps to reproduce the bug
`load_dataset("lhoestq/demo1",revision=None, streaming=True, split="train[:3]")`
causes ValueError: Bad split:... | 328 | 35 | slice split while streaming
### Describe the bug
When using the `load_dataset` function with streaming set to True, slicing splits is apparently not supported.
Did I miss this in the documentation?
### Steps to reproduce the bug
`load_dataset("lhoestq/demo1",revision=None, streaming=True, split="train[:3]")`
... | [
-1.167305827140808,
-0.9285175800323486,
-0.6649534702301025,
1.4969956874847412,
-0.1945435255765915,
-1.1529728174209595,
0.1905038207769394,
-1.076942801475525,
1.6675045490264893,
-0.7395195960998535,
0.3413146436214447,
-1.6742693185806274,
-0.03626634180545807,
-0.6227717399597168,
... |
https://github.com/huggingface/datasets/issues/5457 | prebuilt dataset relies on `downloads/extracted` | Hi!
This issue is due to our audio/image datasets not being self-contained. This allows us to save disk space (files are written only once) but also leads to the issues like this one. We plan to make all our datasets self-contained in Datasets 3.0.
In the meantime, you can run the following map to ensure your da... | ### Describe the bug
I pre-built the dataset:
```
python -c 'import sys; from datasets import load_dataset; ds=load_dataset(sys.argv[1])' HuggingFaceM4/general-pmd-synthetic-testing
```
and it can be used just fine.
now I wipe out `downloads/extracted` and it no longer works.
```
rm -r ~/.cache/huggingface... | 329 | 76 | prebuilt dataset relies on `downloads/extracted`
### Describe the bug
I pre-built the dataset:
```
python -c 'import sys; from datasets import load_dataset; ds=load_dataset(sys.argv[1])' HuggingFaceM4/general-pmd-synthetic-testing
```
and it can be used just fine.
now I wipe out `downloads/extracted` and it... | [
-1.157984972000122,
-0.9018886685371399,
-0.7691069841384888,
1.5294994115829468,
-0.02931012213230133,
-1.293380618095398,
0.09620700776576996,
-0.9256891012191772,
1.568138599395752,
-0.7634966373443604,
0.3926277160644531,
-1.7433313131332397,
-0.05582356080412865,
-0.556559681892395,
... |
https://github.com/huggingface/datasets/issues/5457 | prebuilt dataset relies on `downloads/extracted` | Understood. Thank you, Mario.
Perhaps the solution could be very simple - move the extracted files into the directory of the cached dataset? Which would make it self-contained already and won't require waiting for a new major release. Unless I'm missing some back-compat nuance.
But regardless if X relies on Y - i... | ### Describe the bug
I pre-built the dataset:
```
python -c 'import sys; from datasets import load_dataset; ds=load_dataset(sys.argv[1])' HuggingFaceM4/general-pmd-synthetic-testing
```
and it can be used just fine.
now I wipe out `downloads/extracted` and it no longer works.
```
rm -r ~/.cache/huggingface... | 329 | 76 | prebuilt dataset relies on `downloads/extracted`
### Describe the bug
I pre-built the dataset:
```
python -c 'import sys; from datasets import load_dataset; ds=load_dataset(sys.argv[1])' HuggingFaceM4/general-pmd-synthetic-testing
```
and it can be used just fine.
now I wipe out `downloads/extracted` and it... | [
-1.157984972000122,
-0.9018886685371399,
-0.7691069841384888,
1.5294994115829468,
-0.02931012213230133,
-1.293380618095398,
0.09620700776576996,
-0.9256891012191772,
1.568138599395752,
-0.7634966373443604,
0.3926277160644531,
-1.7433313131332397,
-0.05582356080412865,
-0.556559681892395,
... |
https://github.com/huggingface/datasets/issues/5454 | Save and resume the state of a DataLoader | Something that'd be nice to have is "manual update of state". One of the learning from training LLMs is the ability to skip some batches whenever we notice huge spike might be handy. | It would be nice when using `datasets` with a PyTorch DataLoader to be able to resume a training from a DataLoader state (e.g. to resume a training that crashed)
What I have in mind (but lmk if you have other ideas or comments):
For map-style datasets, this requires to have a PyTorch Sampler state that can be sav... | 330 | 33 | Save and resume the state of a DataLoader
It would be nice when using `datasets` with a PyTorch DataLoader to be able to resume a training from a DataLoader state (e.g. to resume a training that crashed)
What I have in mind (but lmk if you have other ideas or comments):
For map-style datasets, this requires to ... | [
-1.210391640663147,
-0.9462154507637024,
-0.8508016467094421,
1.3852672576904297,
-0.17072966694831848,
-1.3206204175949097,
0.04050850868225098,
-1.0508242845535278,
1.6814910173416138,
-0.917292058467865,
0.3468030095100403,
-1.7072945833206177,
0.19839245080947876,
-0.5438280701637268,
... |
https://github.com/huggingface/datasets/issues/5454 | Save and resume the state of a DataLoader | Your outline spec is very sound and clear, @lhoestq - thank you!
@thomasw21, indeed that would be a wonderful extra feature. In Megatron-Deepspeed we manually drained the dataloader for the range we wanted. I wasn't very satisfied with the way we did it, since its behavior would change if you were to do multiple ran... | It would be nice when using `datasets` with a PyTorch DataLoader to be able to resume a training from a DataLoader state (e.g. to resume a training that crashed)
What I have in mind (but lmk if you have other ideas or comments):
For map-style datasets, this requires to have a PyTorch Sampler state that can be sav... | 330 | 97 | Save and resume the state of a DataLoader
It would be nice when using `datasets` with a PyTorch DataLoader to be able to resume a training from a DataLoader state (e.g. to resume a training that crashed)
What I have in mind (but lmk if you have other ideas or comments):
For map-style datasets, this requires to ... | [
-1.2777979373931885,
-0.9837959408760071,
-0.755339503288269,
1.356204867362976,
-0.17536111176013947,
-1.2910062074661255,
0.03848791867494583,
-1.072117805480957,
1.6808425188064575,
-0.8867015242576599,
0.33993273973464966,
-1.6807563304901123,
0.16393020749092102,
-0.521294355392456,
... |
https://github.com/huggingface/datasets/issues/5451 | ImageFolder BadZipFile: Bad offset for central directory | Hi ! Could you share the full stack trace ? Which dataset did you try to load ?
it may be related to https://github.com/huggingface/datasets/pull/5640 | ### Describe the bug
I'm getting the following exception:
```
lib/python3.10/zipfile.py:1353 in _RealGetContents │
│ │
│ 1350 │ │ # self.start_dir: Position of start of central directory ... | 331 | 24 | ImageFolder BadZipFile: Bad offset for central directory
### Describe the bug
I'm getting the following exception:
```
lib/python3.10/zipfile.py:1353 in _RealGetContents │
│ │
│ 1350 │ │ # self.start_dir... | [
-1.1801573038101196,
-0.8495473861694336,
-0.6627269387245178,
1.5185000896453857,
-0.19597958028316498,
-1.228810429573059,
0.24361631274223328,
-1.032873272895813,
1.5680439472198486,
-0.7590497732162476,
0.3073175847530365,
-1.6794732809066772,
0.13305480778217316,
-0.5226603746414185,
... |
https://github.com/huggingface/datasets/issues/5451 | ImageFolder BadZipFile: Bad offset for central directory | The `BadZipFile` error means the ZIP file is corrupted, so I'm closing this issue as it's not directly related to `datasets`. | ### Describe the bug
I'm getting the following exception:
```
lib/python3.10/zipfile.py:1353 in _RealGetContents │
│ │
│ 1350 │ │ # self.start_dir: Position of start of central directory ... | 331 | 21 | ImageFolder BadZipFile: Bad offset for central directory
### Describe the bug
I'm getting the following exception:
```
lib/python3.10/zipfile.py:1353 in _RealGetContents │
│ │
│ 1350 │ │ # self.start_dir... | [
-1.1939665079116821,
-0.8490809202194214,
-0.6932085752487183,
1.5191311836242676,
-0.2111867517232895,
-1.2388697862625122,
0.26149576902389526,
-1.0506815910339355,
1.5851229429244995,
-0.7903262376785278,
0.3210504353046417,
-1.6974228620529175,
0.1194610595703125,
-0.546713650226593,
... |
https://github.com/huggingface/datasets/issues/5451 | ImageFolder BadZipFile: Bad offset for central directory | For others that find this issue following a `BadZipFile` error, I had the same problem because I had a file in a folder dataset `my-image.target` and the datasets library was incorrectly determining that the (PNG) file was a zip archive. When it tried to extract the file, this error occurred.
Updating to `datasets=... | ### Describe the bug
I'm getting the following exception:
```
lib/python3.10/zipfile.py:1353 in _RealGetContents │
│ │
│ 1350 │ │ # self.start_dir: Position of start of central directory ... | 331 | 58 | ImageFolder BadZipFile: Bad offset for central directory
### Describe the bug
I'm getting the following exception:
```
lib/python3.10/zipfile.py:1353 in _RealGetContents │
│ │
│ 1350 │ │ # self.start_dir... | [
-1.2359187602996826,
-0.8569865822792053,
-0.6782492399215698,
1.4818766117095947,
-0.21162763237953186,
-1.261195182800293,
0.23142798244953156,
-1.0930302143096924,
1.6386640071868896,
-0.8028919100761414,
0.26819145679473877,
-1.6773107051849365,
0.13818736374378204,
-0.5428324937820435... |
https://github.com/huggingface/datasets/issues/5450 | to_tf_dataset with a TF collator causes bizarrely persistent slowdown | A possible hint: The function it seems to be spending a lot of time in (when iterating over the original dataset) is `_get_mp` in the PIL JPEG decoder:

| ### Describe the bug
This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing)
Briefly, there are several datasets that, when you iterate over them with `to_tf_dataset` **and** a data colla... | 332 | 29 | to_tf_dataset with a TF collator causes bizarrely persistent slowdown
### Describe the bug
This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing)
Briefly, there are several datasets th... | [
-1.1414843797683716,
-0.9540003538131714,
-0.714849054813385,
1.382982850074768,
-0.2000240981578827,
-1.2782437801361084,
0.1743442714214325,
-1.1151078939437866,
1.7982470989227295,
-0.8330002427101135,
0.24018649756908417,
-1.7073076963424683,
0.051127463579177856,
-0.6064121723175049,
... |
https://github.com/huggingface/datasets/issues/5450 | to_tf_dataset with a TF collator causes bizarrely persistent slowdown | If "mp" is multiprocessing, this might suggest some kind of negative interaction between the JPEG decoder and TF's handling of processes/threads. Note that we haven't merged the parallel `to_tf_dataset` PR yet, so it's not caused by that PR! | ### Describe the bug
This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing)
Briefly, there are several datasets that, when you iterate over them with `to_tf_dataset` **and** a data colla... | 332 | 38 | to_tf_dataset with a TF collator causes bizarrely persistent slowdown
### Describe the bug
This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing)
Briefly, there are several datasets th... | [
-1.1403965950012207,
-0.9472631216049194,
-0.7174069881439209,
1.387075424194336,
-0.2061592936515808,
-1.2699196338653564,
0.17962321639060974,
-1.1113897562026978,
1.7824658155441284,
-0.8306373357772827,
0.23737767338752747,
-1.7016518115997314,
0.050736263394355774,
-0.6071605086326599... |
https://github.com/huggingface/datasets/issues/5450 | to_tf_dataset with a TF collator causes bizarrely persistent slowdown | Update: MP isn't multiprocessing at all, it's an internal PIL method for loading metadata from JPEG files. No idea why that would be a bottleneck, but I'll see if a Python profiler can't figure out where the time is actually being spent. | ### Describe the bug
This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing)
Briefly, there are several datasets that, when you iterate over them with `to_tf_dataset` **and** a data colla... | 332 | 42 | to_tf_dataset with a TF collator causes bizarrely persistent slowdown
### Describe the bug
This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing)
Briefly, there are several datasets th... | [
-1.141693353652954,
-0.9571100473403931,
-0.7233511209487915,
1.385690450668335,
-0.20012611150741577,
-1.2680176496505737,
0.1750892698764801,
-1.1131263971328735,
1.7796988487243652,
-0.8293800950050354,
0.22731220722198486,
-1.697354793548584,
0.05005267634987831,
-0.6047587394714355,
... |
https://github.com/huggingface/datasets/issues/5450 | to_tf_dataset with a TF collator causes bizarrely persistent slowdown | After further profiling, the slowdown is in the C methods for JPEG decoding that are included as part of PIL. Because Python profilers can't inspect inside that, I don't have any further information on which lines exactly are responsible for the slowdown or why.
In the meantime, I'm going to suggest switching from `... | ### Describe the bug
This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing)
Briefly, there are several datasets that, when you iterate over them with `to_tf_dataset` **and** a data colla... | 332 | 76 | to_tf_dataset with a TF collator causes bizarrely persistent slowdown
### Describe the bug
This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing)
Briefly, there are several datasets th... | [
-1.1419216394424438,
-0.9476639628410339,
-0.7116085886955261,
1.3707692623138428,
-0.20300760865211487,
-1.2872434854507446,
0.18772634863853455,
-1.1250361204147339,
1.7960673570632935,
-0.837628960609436,
0.22950974106788635,
-1.699198603630066,
0.06453695893287659,
-0.5998414754867554,... |
https://github.com/huggingface/datasets/issues/5444 | info messages logged as warnings | Looks like a duplicate of https://github.com/huggingface/datasets/issues/1948.
I also think these should be logged as INFO messages, but let's see what @lhoestq thinks. | ### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from the Python docs for reference:
* I... | 333 | 22 | info messages logged as warnings
### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from ... | [
-1.2012889385223389,
-0.9273018836975098,
-0.7546572685241699,
1.4767590761184692,
-0.167137011885643,
-1.2878497838974,
0.1407080888748169,
-1.0096759796142578,
1.6865419149398804,
-0.8163502812385559,
0.3636291027069092,
-1.626015067100525,
-0.005010148044675589,
-0.5913500189781189,
-... |
https://github.com/huggingface/datasets/issues/5444 | info messages logged as warnings | It can be considered unexpected to see a `map` function return instantaneously. The warning is here to explain this case by mentioning that the cache was used. I don't expect first time users (only seeing warnings) to guess that the cache works this way | ### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from the Python docs for reference:
* I... | 333 | 44 | info messages logged as warnings
### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from ... | [
-1.236562967300415,
-0.916245698928833,
-0.7499101161956787,
1.4740647077560425,
-0.1550338864326477,
-1.2958605289459229,
0.12919111549854279,
-1.0116636753082275,
1.7336289882659912,
-0.8567910194396973,
0.3752639889717102,
-1.6075327396392822,
0.01064293272793293,
-0.584657609462738,
... |
https://github.com/huggingface/datasets/issues/5444 | info messages logged as warnings | Oh, so it's intentional? Do all Hugging Face packages use `warning` when using cache?
I guess feel free to close this issue then. | ### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from the Python docs for reference:
* I... | 333 | 23 | info messages logged as warnings
### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from ... | [
-1.2241021394729614,
-0.9066925644874573,
-0.7336789965629578,
1.4768046140670776,
-0.1610291451215744,
-1.2774314880371094,
0.1545158475637436,
-1.0086294412612915,
1.7311090230941772,
-0.8399935364723206,
0.3464798033237457,
-1.6371490955352783,
0.03219063580036163,
-0.602618932723999,
... |
https://github.com/huggingface/datasets/issues/5444 | info messages logged as warnings | Yes it's intentional for `map`. For `load_dataset` it's also intentional but for a different reason: it shows where in the cache the dataset is located, in case the user wants to clear the cache. | ### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from the Python docs for reference:
* I... | 333 | 34 | info messages logged as warnings
### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from ... | [
-1.2135945558547974,
-0.9044673442840576,
-0.756445050239563,
1.4845976829528809,
-0.15900501608848572,
-1.2960442304611206,
0.15157468616962433,
-1.0147960186004639,
1.7531806230545044,
-0.8709537982940674,
0.3316876292228699,
-1.6485707759857178,
0.00995037704706192,
-0.5834614038467407,... |
https://github.com/huggingface/datasets/issues/5444 | info messages logged as warnings | OK I see. It's surprising to me that these are considered "something unexpected happened", the concept of cache is pretty common.
Has a user every actually complained that they ran their code once, and it took a minute while the data downloaded, then ran their code again and it ran really fast (and completed success... | ### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from the Python docs for reference:
* I... | 333 | 183 | info messages logged as warnings
### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from ... | [
-1.255374789237976,
-0.9473716616630554,
-0.7523675560951233,
1.4561004638671875,
-0.1649414300918579,
-1.2520487308502197,
0.0686570405960083,
-1.0906296968460083,
1.6898003816604614,
-0.8039632439613342,
0.305169939994812,
-1.6514534950256348,
0.04116402938961983,
-0.6184441447257996,
... |
https://github.com/huggingface/datasets/issues/5444 | info messages logged as warnings | I see your point and I'm not closed to switching to INFO, but I think those logs are important to make the library less opaque. I also just checked `transformers` scripts and they default to INFO which is nice. However for colab users the default is still WARNING iirc, and it counts as one of the main env where `datase... | ### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from the Python docs for reference:
* I... | 333 | 94 | info messages logged as warnings
### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from ... | [
-1.202043056488037,
-0.9279480576515198,
-0.7376196980476379,
1.464721441268921,
-0.15710747241973877,
-1.2811208963394165,
0.1379225254058838,
-1.106611967086792,
1.726864218711853,
-0.8411441445350647,
0.34654518961906433,
-1.6311951875686646,
0.00867342110723257,
-0.5844743251800537,
... |
https://github.com/huggingface/datasets/issues/5444 | info messages logged as warnings | These kinds of messages are logged as INFO in Transformers, so we should probably be consistent with them | ### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from the Python docs for reference:
* I... | 333 | 18 | info messages logged as warnings
### Describe the bug
Code in `datasets` is using `logger.warning` when it should be using `logger.info`.
Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category.
Definitions from ... | [
-1.2303543090820312,
-0.8941532373428345,
-0.7463688254356384,
1.490765929222107,
-0.19838981330394745,
-1.3130637407302856,
0.13688313961029053,
-1.0077579021453857,
1.7194478511810303,
-0.8827599287033081,
0.3706198036670685,
-1.6408991813659668,
0.030980480834841728,
-0.6354483962059021... |
https://github.com/huggingface/datasets/issues/5442 | OneDrive Integrations with HF Datasets | Hi!
We use [`fsspec`](https://github.com/fsspec/filesystem_spec) to integrate with storage providers. You can find more info (and the usage examples) in [our docs](https://huggingface.co/docs/datasets/v2.8.0/filesystems#download-and-prepare-a-dataset-into-a-cloud-storage).
[`gdrivefs`](https://github.com/fsspec/... | ### Feature request
First of all , I would like to thank all community who are developed DataSet storage and make it free available
How to integrate our Onedrive account or any other possible storage clouds (like google drive,...) with the **HF** datasets section.
For example, if I have **50GB** on my **Onedrive*... | 334 | 67 | OneDrive Integrations with HF Datasets
### Feature request
First of all , I would like to thank all community who are developed DataSet storage and make it free available
How to integrate our Onedrive account or any other possible storage clouds (like google drive,...) with the **HF** datasets section.
For exa... | [
-1.24271821975708,
-0.959071159362793,
-0.7232421040534973,
1.3471896648406982,
-0.11604730039834976,
-1.284589171409607,
0.12122423946857452,
-1.082366943359375,
1.6358044147491455,
-0.7559685111045837,
0.28111669421195984,
-1.7193418741226196,
-0.004981605801731348,
-0.5995687246322632,
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.