html_url stringlengths 51 51 | title stringlengths 6 280 | comments stringlengths 67 24.7k | body stringlengths 51 36.2k | __index_level_0__ int64 1 1.17k | comment_length int64 16 1.45k | text stringlengths 190 38.3k | embeddings list |
|---|---|---|---|---|---|---|---|
https://github.com/huggingface/datasets/issues/6109 | Problems in downloading Amazon reviews from HF | Thanks for reporting, @610v4nn1.
Indeed, the source data files are no longer available. We have contacted the authors of the dataset and they report that Amazon has decided to stop distributing the multilingual reviews dataset.
We are adding a notification about this issue to the dataset card.
See: https://hug... | ### Describe the bug
I have a script downloading `amazon_reviews_multi`.
When the download starts, I get
```
Downloading data files: 0%| | 0/1 [00:00<?, ?it/s]
Downloading data: 243B [00:00, 1.43MB/s]
Downloading data files: 100%|ββββββββββ| 1/1 [00:01<00:00, 1.54s/it]
Extracting data files: 100%... | 1 | 49 | Problems in downloading Amazon reviews from HF
### Describe the bug
I have a script downloading `amazon_reviews_multi`.
When the download starts, I get
```
Downloading data files: 0%| | 0/1 [00:00<?, ?it/s]
Downloading data: 243B [00:00, 1.43MB/s]
Downloading data files: 100%|ββββββββββ| 1/1 [00:... | [
-1.1749063730239868,
-0.8698778748512268,
-0.8900013566017151,
1.4243459701538086,
-0.13573047518730164,
-1.2226994037628174,
0.15154093503952026,
-1.2511365413665771,
1.432878017425537,
-0.6011939644813538,
0.4040875732898712,
-1.727347731590271,
0.04589912295341492,
-0.4552827477455139,
... |
https://github.com/huggingface/datasets/issues/6108 | Loading local datasets got strangely stuck | Yesterday I waited for more than 12 hours to make sure it was really **stuck** instead of proceeding too slow. | ### Describe the bug
I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as:
```python
ds = load_dataset("json", data_files=LIST_OF_FILE_PATHS, num_proc=... | 2 | 20 | Loading local datasets got strangely stuck
### Describe the bug
I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as:
```python
ds = load_dataset("js... | [
-1.1857736110687256,
-0.9407454133033752,
-0.7441576719284058,
1.4746325016021729,
-0.11657063663005829,
-1.2399708032608032,
0.11755771934986115,
-1.0479322671890259,
1.7225706577301025,
-0.7684071063995361,
0.38269880414009094,
-1.6014268398284912,
0.03821048140525818,
-0.620981097221374... |
https://github.com/huggingface/datasets/issues/6108 | Loading local datasets got strangely stuck | I've had similar weird issues with `load_dataset` as well. Not multiple files, but dataset is quite big, about 50G. | ### Describe the bug
I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as:
```python
ds = load_dataset("json", data_files=LIST_OF_FILE_PATHS, num_proc=... | 2 | 19 | Loading local datasets got strangely stuck
### Describe the bug
I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as:
```python
ds = load_dataset("js... | [
-1.1857736110687256,
-0.9407454133033752,
-0.7441576719284058,
1.4746325016021729,
-0.11657063663005829,
-1.2399708032608032,
0.11755771934986115,
-1.0479322671890259,
1.7225706577301025,
-0.7684071063995361,
0.38269880414009094,
-1.6014268398284912,
0.03821048140525818,
-0.620981097221374... |
https://github.com/huggingface/datasets/issues/6099 | How do i get "amazon_us_reviews | Seems like the problem isn't with the library, but the dataset itself hosted on AWS S3.
Its [homepage](https://s3.amazonaws.com/amazon-reviews-pds/readme.html) returns an `AccessDenied` XML response, which is the same thing you get if you try to log the `record` that triggers the exception
```python
try:
ex... | ### Feature request
I have been trying to load 'amazon_us_dataset" but unable to do so.
`amazon_us_reviews = load_dataset('amazon_us_reviews')`
`print(amazon_us_reviews)`
> [ValueError: Config name is missing.
Please pick one among the available configs: ['Wireless_v1_00', 'Watches_v1_00', 'Video_Games_v1... | 5 | 67 | How do i get "amazon_us_reviews
### Feature request
I have been trying to load 'amazon_us_dataset" but unable to do so.
`amazon_us_reviews = load_dataset('amazon_us_reviews')`
`print(amazon_us_reviews)`
> [ValueError: Config name is missing.
Please pick one among the available configs: ['Wireless_v1_00'... | [
-1.210225224494934,
-1.2076855897903442,
-0.5899178385734558,
1.6656270027160645,
0.004276080057024956,
-1.407726764678955,
0.13049446046352386,
-1.0258605480194092,
1.3568669557571411,
-0.8192351460456848,
0.10499061644077301,
-1.3370153903961182,
-0.003387860022485256,
-0.735708951950073... |
https://github.com/huggingface/datasets/issues/6089 | AssertionError: daemonic processes are not allowed to have children | We could add a "threads" parallel backend to `datasets.parallel.parallel_backend` to support downloading with threads but note that `download_and_extract` also decompresses archives, and this is a CPU-intensive task, which is not ideal for (Python) threads (good for IO-intensive tasks). | ### Describe the bug
When I load_dataset with num_proc > 0 in a deamon process, I got an error:
```python
File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract
return self.extract(self.download(url_or_urls))
^^^^^^^^^^^^^^^^^
File "/Users... | 8 | 38 | AssertionError: daemonic processes are not allowed to have children
### Describe the bug
When I load_dataset with num_proc > 0 in a deamon process, I got an error:
```python
File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract
return self.extr... | [
-1.3304325342178345,
-0.843843400478363,
-0.6240853667259216,
1.2293169498443604,
0.023741507902741432,
-0.9174615740776062,
0.06877093762159348,
-1.0991508960723877,
1.8150602579116821,
-0.6466710567474365,
0.21736067533493042,
-1.551055669784546,
-0.1861899197101593,
-0.48568251729011536... |
https://github.com/huggingface/datasets/issues/6089 | AssertionError: daemonic processes are not allowed to have children | > We could add a "threads" parallel backend to `datasets.parallel.parallel_backend` to support downloading with threads but note that `download_and_extract` also decompresses archives, and this is a CPU-intensive task, which is not ideal for (Python) threads (good for IO-intensive tasks).
Great! Download takes more ... | ### Describe the bug
When I load_dataset with num_proc > 0 in a deamon process, I got an error:
```python
File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract
return self.extract(self.download(url_or_urls))
^^^^^^^^^^^^^^^^^
File "/Users... | 8 | 58 | AssertionError: daemonic processes are not allowed to have children
### Describe the bug
When I load_dataset with num_proc > 0 in a deamon process, I got an error:
```python
File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract
return self.extr... | [
-1.3304325342178345,
-0.843843400478363,
-0.6240853667259216,
1.2293169498443604,
0.023741507902741432,
-0.9174615740776062,
0.06877093762159348,
-1.0991508960723877,
1.8150602579116821,
-0.6466710567474365,
0.21736067533493042,
-1.551055669784546,
-0.1861899197101593,
-0.48568251729011536... |
https://github.com/huggingface/datasets/issues/6086 | Support `fsspec` in `Dataset.to_<format>` methods | I'm assuming this should just cover `to_csv`, `to_parquet`, and `to_json`, right? As `to_list` and `to_dict` just return Python objects, `to_pandas` returns a `pandas.DataFrame` and `to_sql` just inserts into a SQL DB, is that right? | Supporting this should be fairly easy.
Requested on the forum [here](https://discuss.huggingface.co/t/how-can-i-convert-a-loaded-dataset-in-to-a-parquet-file-and-save-it-to-the-s3/48353). | 10 | 34 | Support `fsspec` in `Dataset.to_<format>` methods
Supporting this should be fairly easy.
Requested on the forum [here](https://discuss.huggingface.co/t/how-can-i-convert-a-loaded-dataset-in-to-a-parquet-file-and-save-it-to-the-s3/48353).
I'm assuming this should just cover `to_csv`, `to_parquet`, and `to_json`, ... | [
-1.0311846733093262,
-0.7468673586845398,
-0.7434991002082825,
1.6008455753326416,
-0.025689486414194107,
-1.4986146688461304,
0.25253990292549133,
-1.046440601348877,
1.7896806001663208,
-0.9305419921875,
0.49904119968414307,
-1.5891073942184448,
0.06446303427219391,
-0.7518453001976013,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | When the process starts to hang, can you interrupt it with CTRL + C and paste the error stack trace here? | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 21 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2703862190246582,
-1.0200567245483398,
-0.7282028198242188,
1.5863037109375,
-0.22804024815559387,
-1.1206165552139282,
0.14635612070560455,
-0.9283672571182251,
1.6327399015426636,
-0.8532999157905579,
0.25511327385902405,
-1.5935802459716797,
0.06626027077436447,
-0.5541383624076843,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Thanks @mariosasko for your prompt response, here's the stack trace:
```
KeyboardInterrupt Traceback (most recent call last)
Cell In[12], line 4
2 t = time.time()
3 iter_ = 0
----> 4 for batch in train_dataloader:
5 #batch_proc = streaming_obj.collect_streaming_dat... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 308 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2974315881729126,
-1.0190706253051758,
-0.664889931678772,
1.5647151470184326,
-0.20822866261005402,
-1.1625877618789673,
0.14594830572605133,
-0.9586323499679565,
1.6330128908157349,
-0.8328601121902466,
0.23161981999874115,
-1.5734355449676514,
0.07717040926218033,
-0.5745692253112793... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Update: If i let it run, it eventually fails with:
```
RuntimeError Traceback (most recent call last)
Cell In[16], line 4
2 t = time.time()
3 iter_ = 0
----> 4 for batch in train_dataloader:
5 #batch_proc = streaming_obj.collect_streaming_data_batch(batch)
... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 416 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.270894169807434,
-1.0062732696533203,
-0.6609833240509033,
1.563029408454895,
-0.22833463549613953,
-1.170157790184021,
0.1303490400314331,
-0.9474954009056091,
1.6263947486877441,
-0.8306624889373779,
0.2395327091217041,
-1.5603952407836914,
0.07044895738363266,
-0.571514368057251,
-... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | PyTorch tensors cannot store `Decimal` objects. Casting the column with decimals to `float` should fix the issue. | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 17 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2665154933929443,
-1.0163578987121582,
-0.7126186490058899,
1.5815691947937012,
-0.22405150532722473,
-1.126129150390625,
0.14313547313213348,
-0.9426345229148865,
1.658698558807373,
-0.8402796387672424,
0.2593388259410858,
-1.6109161376953125,
0.07901172339916229,
-0.5566502213478088,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | I already have cast in collate_fn, in which I perform .astype(float) for each numerical field.
On the same instance, I installed a conda env with python 3.6, and this works well.
Sample:
```
def streaming_data_collate_fn(batch):
df = pd.DataFrame.from_dict(batch)
feat_vals = torch.FloatTensor(np.nan_t... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 42 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.3096867799758911,
-1.0151026248931885,
-0.654405415058136,
1.5715867280960083,
-0.2311745584011078,
-1.1411608457565308,
0.1739429086446762,
-0.9530844688415527,
1.66796875,
-0.8388645648956299,
0.2485904097557068,
-1.5883809328079224,
0.09396015852689743,
-0.5771110653877258,
-0.6925... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | `collate_fn` is applied after the `torch` formatting step, so I think the only option when working with an `IterableDataset` is to remove the `with_format` call and perform the conversion from Python values to PyTorch tensors in `collate_fn`. The standard `Dataset` supports `with_format("numpy")`, which should make thi... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 48 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2474640607833862,
-0.9869871735572815,
-0.7190800905227661,
1.585418462753296,
-0.23098866641521454,
-1.1668672561645508,
0.17919516563415527,
-0.9628450274467468,
1.6689121723175049,
-0.8653399348258972,
0.28742289543151855,
-1.6140029430389404,
0.06157983839511871,
-0.5838105082511902... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Thanks!
Python 3.10 conda-env: After replacing with_format("torch") with with_format("numpy"), the error went away. However, it was still taking over 2 minutes to load a very small batch of 64 samples with num_workers set to 32. Once I removed with_format call altogether, it is finishing in 11 seconds.
Python 3.6 ... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 83 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2701588869094849,
-0.997541069984436,
-0.6988049149513245,
1.51397705078125,
-0.25449058413505554,
-1.1525267362594604,
0.17723368108272552,
-0.9602234959602356,
1.6439114809036255,
-0.8577772378921509,
0.24207332730293274,
-1.6130080223083496,
0.08785462379455566,
-0.5229309797286987,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Can you share the `datasets` and `torch` versions installed in these conda envs?
> Once I removed with_format call altogether, it is finishing in 11 seconds.
Hmm, that's surprising. What are your dataset's `.features`? | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 34 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2404839992523193,
-1.0219470262527466,
-0.7118972539901733,
1.579450011253357,
-0.21030646562576294,
-1.133574366569519,
0.14982429146766663,
-0.9466209411621094,
1.682141661643982,
-0.8303287625312805,
0.23833730816841125,
-1.617666482925415,
0.05280683934688568,
-0.5681951642036438,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Python 3.6:
datasets.__version__ 2.4.0
torch.__version__ 1.10.1+cu102
Python 3.10:
datasets.__version__ 2.14.0
torch.__version__ 2.0.0
Anonymized features are of the form (subset shown here):
{
'string_feature_i': Value(dtype='string', id=None),
'numerical_feature_i': Value(dtype='decimal128(38, 0)', id=N... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 46 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.3298096656799316,
-1.0528417825698853,
-0.6605807542800903,
1.6113710403442383,
-0.2553796172142029,
-1.1346224546432495,
0.14628013968467712,
-0.9305062890052795,
1.6034176349639893,
-0.8199122548103333,
0.21513554453849792,
-1.5423533916473389,
0.0535685159265995,
-0.5379592776298523,... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | One more thing, in python 3.10 based kernel, interestingly increasing num_workers seem to be increasing the runtime of iterating I was trying out. In python 3.10 kernel execution, I do not even see multiple CPU cores spiking unlike in 3.6.
512 batch size on 32 workers executes in 2.4 seconds on python 3.6 kernel, wh... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 61 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2805341482162476,
-1.0410552024841309,
-0.6732608079910278,
1.5353178977966309,
-0.254350870847702,
-1.1552599668502808,
0.1737479567527771,
-0.9447788596153259,
1.6417279243469238,
-0.8380560278892517,
0.23958228528499603,
-1.5837104320526123,
0.048319969326257706,
-0.5743476748466492,... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | **Update**: It seems the latency part is more of a multiprocessing issue with torch and some host specific issue, and I had to scourge through relevant pytorch issues, when I stumbled across these threads:
1. https://github.com/pytorch/pytorch/issues/102494
2. https://github.com/pytorch/pytorch/issues/102269
3. http... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 11 | 87 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2661398649215698,
-1.0214083194732666,
-0.7204177379608154,
1.5396506786346436,
-0.22598449885845184,
-1.1263196468353271,
0.09709709882736206,
-0.9557428359985352,
1.616188406944275,
-0.767447292804718,
0.2418806105852127,
-1.6014881134033203,
0.014191004447638988,
-0.5700944066047668,... |
https://github.com/huggingface/datasets/issues/6078 | resume_download with streaming=True | Currently, it's not possible to efficiently resume streaming after an error. Eventually, we plan to support this for Parquet (see https://github.com/huggingface/datasets/issues/5380). | ### Describe the bug
I used:
```
dataset = load_dataset(
"oscar-corpus/OSCAR-2201",
token=True,
language="fr",
streaming=True,
split="train"
)
```
Unfortunately, the server had a problem during the training process. I saved the step my training stopped at.
But how can I resume download f... | 12 | 21 | resume_download with streaming=True
### Describe the bug
I used:
```
dataset = load_dataset(
"oscar-corpus/OSCAR-2201",
token=True,
language="fr",
streaming=True,
split="train"
)
```
Unfortunately, the server had a problem during the training process. I saved the step my training stoppe... | [
-1.2152302265167236,
-0.9335026144981384,
-0.6795473694801331,
1.4461760520935059,
-0.1448366641998291,
-1.1738520860671997,
0.18506090342998505,
-1.0231943130493164,
1.5911850929260254,
-0.8055459260940552,
0.3376910090446472,
-1.6882258653640747,
-0.015008553862571716,
-0.554406762123107... |
https://github.com/huggingface/datasets/issues/6077 | Mapping gets stuck at 99% | The `MAX_MAP_BATCH_SIZE = 1_000_000_000` hack is bad as it loads the entire dataset into RAM when performing `.map`. Instead, it's best to use `.iter(batch_size)` to iterate over the data batches and compute `mean` for each column. (`stddev` can be computed in another pass).
Also, these arrays are big, so it makes s... | ### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize the features of the dataset, ... | 13 | 62 | Mapping gets stuck at 99%
### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize t... | [
-1.2467619180679321,
-0.9692441821098328,
-0.7773353457450867,
1.3650438785552979,
-0.10279692709445953,
-1.2577497959136963,
0.0820453092455864,
-1.0771145820617676,
1.6699507236480713,
-0.7752376198768616,
0.2535783052444458,
-1.6741188764572144,
0.0001330627128481865,
-0.519218504428863... |
https://github.com/huggingface/datasets/issues/6077 | Mapping gets stuck at 99% | Hi @mariosasko !
I agree, it's an ugly hack, but it was convenient since the resulting `mean_std` could be cached by the library. For my large dataset (which doesn't fit in RAM), I'm actually using something similar to what you suggested. I got rid of the first mapping in the above scripts and replaced it with an it... | ### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize the features of the dataset, ... | 13 | 67 | Mapping gets stuck at 99%
### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize t... | [
-1.2467619180679321,
-0.9692441821098328,
-0.7773353457450867,
1.3650438785552979,
-0.10279692709445953,
-1.2577497959136963,
0.0820453092455864,
-1.0771145820617676,
1.6699507236480713,
-0.7752376198768616,
0.2535783052444458,
-1.6741188764572144,
0.0001330627128481865,
-0.519218504428863... |
https://github.com/huggingface/datasets/issues/6077 | Mapping gets stuck at 99% | Have you tried to reduce `batch_size`/`writer_batch_size` in the 2nd `.map`? Also, can you interrupt the process when it gets stuck and share the error stack trace? | ### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize the features of the dataset, ... | 13 | 26 | Mapping gets stuck at 99%
### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize t... | [
-1.2467619180679321,
-0.9692441821098328,
-0.7773353457450867,
1.3650438785552979,
-0.10279692709445953,
-1.2577497959136963,
0.0820453092455864,
-1.0771145820617676,
1.6699507236480713,
-0.7752376198768616,
0.2535783052444458,
-1.6741188764572144,
0.0001330627128481865,
-0.519218504428863... |
https://github.com/huggingface/datasets/issues/6077 | Mapping gets stuck at 99% | I think `batch_size/writer_batch_size` is already at its lowest in the 2nd `.map` since `batched=False` implies `batch_size=1` and `len(ds) = 1000 = writer_batch_size`.
Here is also a bunch of stack traces when I interrupted the process:
<details>
<summary>stack trace 1</summary>
```python
(pyg)[d623204@ro... | ### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize the features of the dataset, ... | 13 | 1,454 | Mapping gets stuck at 99%
### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize t... | [
-1.2467619180679321,
-0.9692441821098328,
-0.7773353457450867,
1.3650438785552979,
-0.10279692709445953,
-1.2577497959136963,
0.0820453092455864,
-1.0771145820617676,
1.6699507236480713,
-0.7752376198768616,
0.2535783052444458,
-1.6741188764572144,
0.0001330627128481865,
-0.519218504428863... |
https://github.com/huggingface/datasets/issues/6075 | Error loading music files using `load_dataset` | This code behaves as expected on my local machine or in Colab. Which version of `soundfile` do you have installed? MP3 requires `soundfile>=0.12.1`. | ### Describe the bug
I tried to load a music file using `datasets.load_dataset()` from the repository - https://huggingface.co/datasets/susnato/pop2piano_real_music_test
I got the following error -
```
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/susnato/anaconda3/en... | 14 | 23 | Error loading music files using `load_dataset`
### Describe the bug
I tried to load a music file using `datasets.load_dataset()` from the repository - https://huggingface.co/datasets/susnato/pop2piano_real_music_test
I got the following error -
```
Traceback (most recent call last):
File "<stdin>", line 1... | [
-1.2145147323608398,
-0.921959638595581,
-0.7346850633621216,
1.3431165218353271,
-0.05712354928255081,
-1.3600084781646729,
0.1440393477678299,
-1.0172046422958374,
1.5832189321517944,
-0.7139288187026978,
0.3244938552379608,
-1.6303914785385132,
-0.20887859165668488,
-0.43720120191574097... |
https://github.com/huggingface/datasets/issues/6073 | version2.3.2 load_dataset()data_files can't include .xxxx in path | Version 2.3.2 is over one year old, so please use the latest release (2.14.0) to get the expected behavior. Version 2.3.2 does not contain some fixes we made to fix resolving hidden files/directories (starting with a dot). | ### Describe the bug
First, I cd workdir.
Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"})
that couldn't work and
<FileNotFoundError: Unable to find
'/a/b/c/.d/train/train.jsonl' at
/a/b/c/.d/>
And I debug, it is fine in version2.1.2... | 15 | 37 | version2.3.2 load_dataset()data_files can't include .xxxx in path
### Describe the bug
First, I cd workdir.
Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"})
that couldn't work and
<FileNotFoundError: Unable to find
'/a/b/c/.d/train/tr... | [
-1.1433175802230835,
-0.7783908247947693,
-0.4615917205810547,
1.541563630104065,
-0.12804095447063446,
-1.145755648612976,
0.23625068366527557,
-1.0691317319869995,
1.5106480121612549,
-0.679083526134491,
0.15528185665607452,
-1.651146411895752,
0.1933295875787735,
-0.5603920221328735,
... |
https://github.com/huggingface/datasets/issues/6071 | storage_options provided to load_dataset not fully piping through since datasets 2.14.0 | Hi ! Thanks for reporting, I opened a PR to fix this
What filesystem are you using ? | ### Describe the bug
Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that need those options to be set.
I think this is because of the new `_prepare_path_and_sto... | 16 | 18 | storage_options provided to load_dataset not fully piping through since datasets 2.14.0
### Describe the bug
Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that ... | [
-1.1659444570541382,
-0.9065359830856323,
-0.6957008242607117,
1.4630454778671265,
-0.13110680878162384,
-1.281360149383545,
0.20910535752773285,
-1.0998634099960327,
1.6670114994049072,
-0.7348120808601379,
0.32562580704689026,
-1.6995145082473755,
0.009933054447174072,
-0.575554907321929... |
https://github.com/huggingface/datasets/issues/6071 | storage_options provided to load_dataset not fully piping through since datasets 2.14.0 | Hi @lhoestq ! Thank you so much π
It's a bit of a custom setup, but in practice I am using a [pyarrow.fs.S3FileSystem](https://arrow.apache.org/docs/python/generated/pyarrow.fs.S3FileSystem.html) (wrapped in a `fsspec.implementations.arrow.ArrowFSWrapper` [to make it](https://arrow.apache.org/docs/python/filesyste... | ### Describe the bug
Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that need those options to be set.
I think this is because of the new `_prepare_path_and_sto... | 16 | 86 | storage_options provided to load_dataset not fully piping through since datasets 2.14.0
### Describe the bug
Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that ... | [
-1.1659444570541382,
-0.9065359830856323,
-0.6957008242607117,
1.4630454778671265,
-0.13110680878162384,
-1.281360149383545,
0.20910535752773285,
-1.0998634099960327,
1.6670114994049072,
-0.7348120808601379,
0.32562580704689026,
-1.6995145082473755,
0.009933054447174072,
-0.575554907321929... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | You can list the dataset's columns with `ds.column_names` before `.map` to check whether the dataset has an `image` column. If it doesn't, then this is a bug. Otherwise, please paste the line with the `.map` call.
| ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 17 | 36 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.17133367061615,
-0.9671884775161743,
-0.9597904086112976,
1.3718503713607788,
-0.1919693946838379,
-1.2990189790725708,
0.16572007536888123,
-1.1901322603225708,
1.7549738883972168,
-0.969964325428009,
0.27610474824905396,
-1.708359956741333,
0.08246801793575287,
-0.5398345589637756,
... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | This is the piece of code I am running:
```
data_transforms = utils.get_data_augmentation(args)
image_dataset = utils.load_image_dataset(args.dataset)
def resize(examples):
examples["pixel_values"] = [image.convert("RGB").resize((300, 300)) for image in examples["image"]]
return examples
def preproce... | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 17 | 1,035 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.32200288772583,
-0.9935063719749451,
-0.7597574591636658,
1.4973571300506592,
-0.1484074741601944,
-1.2504184246063232,
0.17417672276496887,
-1.1684865951538086,
1.676836371421814,
-0.8942831754684448,
0.34366580843925476,
-1.6277977228164673,
0.02760835736989975,
-0.595337450504303,
... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | The `remove_columns=["image"]` argument in the `.map` call removes the `image` column from the output, so drop this argument to preserve it. | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 17 | 21 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.1705082654953003,
-0.9761772155761719,
-0.9978625178337097,
1.4372361898422241,
-0.19130155444145203,
-1.3484487533569336,
0.14734625816345215,
-1.2010046243667603,
1.7176117897033691,
-0.9727316498756409,
0.27451565861701965,
-1.7099891901016235,
0.08859211206436157,
-0.573822081089019... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | The problem is not with the removal of the image key. The bug is why only the labels are sent to be process, instead of all the featues or dictionary keys.
P.S. I just dropped the removal argument as you've suggested, but that didn't solve the problem, because only the labels are being sent to be processed | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 17 | 57 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.1786153316497803,
-0.9939639568328857,
-0.9947094917297363,
1.3766529560089111,
-0.24062664806842804,
-1.2966949939727783,
0.09172168374061584,
-1.222991704940796,
1.6651229858398438,
-0.9775608777999878,
0.26388615369796753,
-1.6799390316009521,
0.0680551677942276,
-0.5430256724357605,... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | All the `image_dataset.column_names` after the `map` call should also be present in `preprocess_train `/`preprocess_val` unless (input) `columns` in `set_transform` are specified.
If that's not the case, we need a full reproducer (not snippets) with the environment info. | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 17 | 37 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.1579222679138184,
-0.939723789691925,
-0.9462810754776001,
1.4220147132873535,
-0.1891966313123703,
-1.2969825267791748,
0.16110125184059143,
-1.1809947490692139,
1.7300562858581543,
-1.0112223625183105,
0.29581353068351746,
-1.6760975122451782,
0.04168058559298515,
-0.615085780620575,
... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | I have resolved the error after including a collate function as indicated in the Quick Start session of the Datasets docs.:
Here is what I did:
```
data_transforms = utils.get_data_augmentation(args)
image_dataset = utils.load_image_dataset(args.dataset)
def preprocess_train(example_batch):
example_batch[... | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 17 | 139 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.339882731437683,
-0.9851165413856506,
-0.8010908961296082,
1.4915612936019897,
-0.19218501448631287,
-1.2128950357437134,
0.16310033202171326,
-1.1704480648040771,
1.6644611358642578,
-0.8980790972709656,
0.3183603584766388,
-1.6021389961242676,
0.006425078026950359,
-0.5804527997970581... |
https://github.com/huggingface/datasets/issues/6066 | AttributeError: '_tqdm_cls' object has no attribute '_lock' | Hi ! I opened https://github.com/huggingface/datasets/pull/6067 to add the missing `_lock`
We'll do a patch release soon, but feel free to install `datasets` from source in the meantime | ### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p... | 18 | 27 | AttributeError: '_tqdm_cls' object has no attribute '_lock'
### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "... | [
-1.3521512746810913,
-0.8465626835823059,
-0.5868315100669861,
1.283677339553833,
0.03677382320165634,
-0.8991362452507019,
0.05178180709481239,
-1.024348497390747,
1.9279330968856812,
-0.5640751123428345,
0.19435115158557892,
-1.4659757614135742,
-0.12366081774234772,
-0.3550596237182617,... |
https://github.com/huggingface/datasets/issues/6066 | AttributeError: '_tqdm_cls' object has no attribute '_lock' | I have tested the latest main, it does not work.
I add more logs to reproduce this issue, it looks like a multi threading bug:
```python
@contextmanager
def ensure_lock(tqdm_class, lock_name=""):
"""get (create if necessary) and then restore `tqdm_class`'s lock"""
import os
import threading
... | ### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p... | 18 | 184 | AttributeError: '_tqdm_cls' object has no attribute '_lock'
### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "... | [
-1.3521512746810913,
-0.8465626835823059,
-0.5868315100669861,
1.283677339553833,
0.03677382320165634,
-0.8991362452507019,
0.05178180709481239,
-1.024348497390747,
1.9279330968856812,
-0.5640751123428345,
0.19435115158557892,
-1.4659757614135742,
-0.12366081774234772,
-0.3550596237182617,... |
https://github.com/huggingface/datasets/issues/6066 | AttributeError: '_tqdm_cls' object has no attribute '_lock' | Maybe it is a bug of tqdm? I think simply use `try ... except AttributeError ...` wraps `del tqdm_class._lock` should work. | ### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p... | 18 | 21 | AttributeError: '_tqdm_cls' object has no attribute '_lock'
### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "... | [
-1.3521512746810913,
-0.8465626835823059,
-0.5868315100669861,
1.283677339553833,
0.03677382320165634,
-0.8991362452507019,
0.05178180709481239,
-1.024348497390747,
1.9279330968856812,
-0.5640751123428345,
0.19435115158557892,
-1.4659757614135742,
-0.12366081774234772,
-0.3550596237182617,... |
https://github.com/huggingface/datasets/issues/6066 | AttributeError: '_tqdm_cls' object has no attribute '_lock' | Yes it looks like a bug on their end indeed, do you want to open a PR on tqdm ?
Let me see if I can find a workaround in the meantime | ### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p... | 18 | 32 | AttributeError: '_tqdm_cls' object has no attribute '_lock'
### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "... | [
-1.3521512746810913,
-0.8465626835823059,
-0.5868315100669861,
1.283677339553833,
0.03677382320165634,
-0.8991362452507019,
0.05178180709481239,
-1.024348497390747,
1.9279330968856812,
-0.5640751123428345,
0.19435115158557892,
-1.4659757614135742,
-0.12366081774234772,
-0.3550596237182617,... |
https://github.com/huggingface/datasets/issues/6060 | Dataset.map() execute twice when in PyTorch DDP mode | Sorry for asking a duplicate question about `num_proc`, I searched the forum and find the solution.
But I still can't make the trick with `torch.distributed.barrier()` to only map at the main process work. The [post on forum]( https://discuss.huggingface.co/t/slow-processing-with-map-when-using-deepspeed-or-fairscal... | ### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W... | 19 | 40 | Dataset.map() execute twice when in PyTorch DDP mode
### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only... | [
-1.3185489177703857,
-0.9868546724319458,
-0.5761328935623169,
1.4660086631774902,
-0.17576950788497925,
-1.1413180828094482,
0.1571463942527771,
-1.0458317995071411,
1.5937193632125854,
-0.8820594549179077,
0.28995469212532043,
-1.597267985343933,
0.055676743388175964,
-0.522131085395813,... |
https://github.com/huggingface/datasets/issues/6060 | Dataset.map() execute twice when in PyTorch DDP mode | If it does the `map` twice then it means the hash of your map function is not some same between your two processes.
Can you make sure your map functions have the same hash in different processes ?
```python
from datasets.fingerprint import Hasher
print(Hasher.hash(lambda x: cut_reorder_keys(x, num_stations_li... | ### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W... | 19 | 95 | Dataset.map() execute twice when in PyTorch DDP mode
### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only... | [
-1.3185489177703857,
-0.9868546724319458,
-0.5761328935623169,
1.4660086631774902,
-0.17576950788497925,
-1.1413180828094482,
0.1571463942527771,
-1.0458317995071411,
1.5937193632125854,
-0.8820594549179077,
0.28995469212532043,
-1.597267985343933,
0.055676743388175964,
-0.522131085395813,... |
https://github.com/huggingface/datasets/issues/6060 | Dataset.map() execute twice when in PyTorch DDP mode | Thanks for help! I find the fingerprint between processes don't have same hash:
```
Rank 0: Gpu 0 cut_reorder_keys fingerprint c7f47f40e9a67657
Rank 0: Gpu 0 random_shift fingerprint 240a0ce79831e7d4
Rank 1: Gpu 1 cut_reorder_keys fingerprint 20edd3d9cf284001
Rank 1: Gpu 1 random_shift fingerprint 819f7c1c18e773... | ### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W... | 19 | 90 | Dataset.map() execute twice when in PyTorch DDP mode
### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only... | [
-1.3185489177703857,
-0.9868546724319458,
-0.5761328935623169,
1.4660086631774902,
-0.17576950788497925,
-1.1413180828094482,
0.1571463942527771,
-1.0458317995071411,
1.5937193632125854,
-0.8820594549179077,
0.28995469212532043,
-1.597267985343933,
0.055676743388175964,
-0.522131085395813,... |
https://github.com/huggingface/datasets/issues/6058 | laion-coco download error | This can also mean one of the files was not downloaded correctly.
We log an erroneous file's name before raising the reader's error, so this is how you can find the problematic file. Then, you should delete it and call `load_dataset` again.
(I checked all the uploaded files, and they seem to be valid Parquet file... | ### Describe the bug
The full trace:
```
/home/bian/anaconda3/envs/sd/lib/python3.10/site-packages/datasets/load.py:1744: FutureWarning: 'ignore_verifications' was de
precated in favor of 'verification_mode' in version 2.9.1 and will be removed in 3.0.0.
You can remove this warning by passing 'verification_mode=no... | 20 | 67 | laion-coco download error
### Describe the bug
The full trace:
```
/home/bian/anaconda3/envs/sd/lib/python3.10/site-packages/datasets/load.py:1744: FutureWarning: 'ignore_verifications' was de
precated in favor of 'verification_mode' in version 2.9.1 and will be removed in 3.0.0.
You can remove this warning by p... | [
-1.228498935699463,
-0.9487924575805664,
-0.7148016691207886,
1.4198180437088013,
-0.07668789476156235,
-1.150172472000122,
0.09865113347768784,
-1.0615023374557495,
1.4952726364135742,
-0.6555337905883789,
0.2628440856933594,
-1.7097371816635132,
-0.08578065782785416,
-0.5050936937332153,... |
https://github.com/huggingface/datasets/issues/6057 | Why is the speed difference of gen example so big? | Hi!
It's hard to explain this behavior without more information. Can you profile the slower version with the following code
```python
import cProfile, pstats
from datasets import load_dataset
with cProfile.Profile() as profiler:
ds = load_dataset(...)
stats = pstats.Stats(profiler).sort_stats("cumtime"... | ```python
def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir):
with open(metadata_path, 'r') as file:
metadata = json.load(file)
for idx, item in enumerate(metadata):
image_path = item.get('image_path')
text_content = item.get('tex... | 21 | 44 | Why is the speed difference of gen example so big?
```python
def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir):
with open(metadata_path, 'r') as file:
metadata = json.load(file)
for idx, item in enumerate(metadata):
image_path = item.get('i... | [
-1.216934323310852,
-0.9637471437454224,
-0.751177191734314,
1.5022461414337158,
-0.20957061648368835,
-1.2565745115280151,
0.1558200716972351,
-1.0771048069000244,
1.6094040870666504,
-0.8362681269645691,
0.3122471272945404,
-1.6983448266983032,
-0.053490594029426575,
-0.6072121262550354,... |
https://github.com/huggingface/datasets/issues/6051 | Skipping shard in the remote repo and resume upload | Hi! `_select_contiguous` fetches a (zero-copy) slice of the dataset's Arrow table to build a shard, so I don't think this part is the problem. To me, the issue seems to be the step where we embed external image files' bytes (a lot of file reads). You can use `.map` with multiprocessing to perform this step before `push... | ### Describe the bug
For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading.
From my understanding, the problem is in this part of the code:
arrow_dataset.py
```python
for index, shard in logging.tqdm(
enume... | 23 | 111 | Skipping shard in the remote repo and resume upload
### Describe the bug
For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading.
From my understanding, the problem is in this part of the code:
arrow_dataset.py
```python... | [
-1.2337567806243896,
-0.90196293592453,
-0.678459882736206,
1.487384557723999,
-0.22792723774909973,
-1.3088724613189697,
0.11573110520839691,
-1.0518202781677246,
1.6316242218017578,
-0.8330173492431641,
0.36152344942092896,
-1.6441075801849365,
0.030849676579236984,
-0.6168583035469055,
... |
https://github.com/huggingface/datasets/issues/6051 | Skipping shard in the remote repo and resume upload | Hi, thanks, this solution saves some time.
But can't we avoid embedding all external image files bytes with each push, skipping the images that have already been pushed into the repo?
Edit: Ok I missed the part of cache it manually on the disk the first time, this solves the problem. Thank you | ### Describe the bug
For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading.
From my understanding, the problem is in this part of the code:
arrow_dataset.py
```python
for index, shard in logging.tqdm(
enume... | 23 | 53 | Skipping shard in the remote repo and resume upload
### Describe the bug
For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading.
From my understanding, the problem is in this part of the code:
arrow_dataset.py
```python... | [
-1.228704571723938,
-0.9043618440628052,
-0.6839578747749329,
1.5109764337539673,
-0.23591430485248566,
-1.3173547983169556,
0.1201028823852539,
-1.0454732179641724,
1.6178052425384521,
-0.835773766040802,
0.34961482882499695,
-1.637833833694458,
0.06114544719457626,
-0.597659170627594,
... |
https://github.com/huggingface/datasets/issues/6048 | when i use datasets.load_dataset, i encounter the http connect error! | The `audiofolder` loader is not available in version `2.3.2`, hence the error. Please run the `pip install -U datasets` command to update the `datasets` installation to make `load_dataset("audiofolder", ...)` work. | ### Describe the bug
`common_voice_test = load_dataset("audiofolder", data_dir="./dataset/",cache_dir="./cache",split=datasets.Split.TEST)`
when i run the code above, i got the error as below:
--------------------------------------------
ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/... | 24 | 30 | when i use datasets.load_dataset, i encounter the http connect error!
### Describe the bug
`common_voice_test = load_dataset("audiofolder", data_dir="./dataset/",cache_dir="./cache",split=datasets.Split.TEST)`
when i run the code above, i got the error as below:
--------------------------------------------
Connec... | [
-1.1496117115020752,
-0.7972938418388367,
-0.5717766880989075,
1.3664467334747314,
-0.026249855756759644,
-1.354591727256775,
0.0974016860127449,
-0.8682535290718079,
1.5897915363311768,
-0.8269336223602295,
0.35338395833969116,
-1.7552576065063477,
0.027949972078204155,
-0.568594634532928... |
https://github.com/huggingface/datasets/issues/6043 | Compression kwargs have no effect when saving datasets as csv | Hello @exs-avianello, I have reproduced the bug successfully and have understood the problem. But I am confused regarding this part of the statement, "`pandas.DataFrame.to_csv` is always called with a buf-like `path_or_buf`".
Can you please elaborate on it?
Thanks! | ### Describe the bug
Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed.
A warning is raised if explicitly providing a `compression` kwarg, ... | 25 | 38 | Compression kwargs have no effect when saving datasets as csv
### Describe the bug
Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed.
A w... | [
-1.1168360710144043,
-0.900812566280365,
-0.7744753956794739,
1.420461654663086,
-0.10985736548900604,
-1.2823952436447144,
0.10812487453222275,
-1.1060352325439453,
1.7183713912963867,
-0.8180946111679077,
0.30754318833351135,
-1.7111681699752808,
-0.037234142422676086,
-0.628463268280029... |
https://github.com/huggingface/datasets/issues/6043 | Compression kwargs have no effect when saving datasets as csv | Hi @aryanxk02 ! Sure, what I actually meant is that when passing a path-like `path_or_buf` here
https://github.com/huggingface/datasets/blob/14f6edd9222e577dccb962ed5338b79b73502fa5/src/datasets/arrow_dataset.py#L4708-L4714
it gets converted to a file object behind the scenes here
https://github.com/huggingfa... | ### Describe the bug
Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed.
A warning is raised if explicitly providing a `compression` kwarg, ... | 25 | 52 | Compression kwargs have no effect when saving datasets as csv
### Describe the bug
Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed.
A w... | [
-1.1168360710144043,
-0.900812566280365,
-0.7744753956794739,
1.420461654663086,
-0.10985736548900604,
-1.2823952436447144,
0.10812487453222275,
-1.1060352325439453,
1.7183713912963867,
-0.8180946111679077,
0.30754318833351135,
-1.7111681699752808,
-0.037234142422676086,
-0.628463268280029... |
https://github.com/huggingface/datasets/issues/6043 | Compression kwargs have no effect when saving datasets as csv | @exs-avianello When `path_or_buf` is set to None, the `to_csv()` method will return the CSV data as a string instead of saving it to a file. Hence the compression doesn't take place. I think setting `path_or_buf=self.path_or_buf` should work. What you say? | ### Describe the bug
Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed.
A warning is raised if explicitly providing a `compression` kwarg, ... | 25 | 40 | Compression kwargs have no effect when saving datasets as csv
### Describe the bug
Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed.
A w... | [
-1.1168360710144043,
-0.900812566280365,
-0.7744753956794739,
1.420461654663086,
-0.10985736548900604,
-1.2823952436447144,
0.10812487453222275,
-1.1060352325439453,
1.7183713912963867,
-0.8180946111679077,
0.30754318833351135,
-1.7111681699752808,
-0.037234142422676086,
-0.628463268280029... |
https://github.com/huggingface/datasets/issues/6038 | File "/home/zhizhou/anaconda3/envs/pytorch/lib/python3.10/site-packages/datasets/builder.py", line 992, in _download_and_prepare if str(split_generator.split_info.name).lower() == "all": AttributeError: 'str' object has no attribute 'split_info'. Did you mean: 'splitlines'? | Instead of writing the loading script, you can use the built-in loader to [load JSON files](https://huggingface.co/docs/datasets/loading#json):
```python
from datasets import load_dataset
ds = load_dataset("json", data_files={"train": os.path.join(data_dir["train"]), "dev": os.path.join(data_dir["dev"])})
``` | Hi, I use the code below to load local file
```
def _split_generators(self, dl_manager):
# TODO: This method is tasked with downloading/extracting the data and defining the splits depending on the configuration
# If several configurations are possible (listed in BUILDER_CONFIGS), the configurati... | 26 | 29 | File "/home/zhizhou/anaconda3/envs/pytorch/lib/python3.10/site-packages/datasets/builder.py", line 992, in _download_and_prepare if str(split_generator.split_info.name).lower() == "all": AttributeError: 'str' object has no attribute 'split_info'. Did you mean: 'splitlines'?
Hi, I use the code below to load loca... | [
-1.2361313104629517,
-0.9707532525062561,
-0.7087602615356445,
1.4899070262908936,
-0.20801934599876404,
-1.2016712427139282,
0.14858272671699524,
-1.0610185861587524,
1.639115333557129,
-0.8069785833358765,
0.30962151288986206,
-1.6301939487457275,
0.004292638972401619,
-0.565640330314636... |
https://github.com/huggingface/datasets/issues/6037 | Documentation links to examples are broken | These docs are outdated (version 1.2.1 is over two years old). Please refer to [this](https://huggingface.co/docs/datasets/dataset_script) version instead.
Initially, we hosted datasets in this repo, but now you can find them [on the HF Hub](https://huggingface.co/datasets) (e.g. the [`ag_news`](https://huggingface.... | ### Describe the bug
The links at the bottom of [add_dataset](https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html) to examples of specific datasets are all broken, for example
- text classification: [ag_news](https://github.com/huggingface/datasets/blob/master/datasets/ag_news/ag_news.py) (original data ... | 27 | 38 | Documentation links to examples are broken
### Describe the bug
The links at the bottom of [add_dataset](https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html) to examples of specific datasets are all broken, for example
- text classification: [ag_news](https://github.com/huggingface/datasets/blob/master... | [
-1.201838731765747,
-0.8922097682952881,
-0.6922095417976379,
1.4508154392242432,
-0.05030794069170952,
-1.3228124380111694,
0.06119988486170769,
-0.9711858630180359,
1.5745917558670044,
-0.6364167332649231,
0.32652443647384644,
-1.6934266090393066,
-0.08025117218494415,
-0.499392360448837... |
https://github.com/huggingface/datasets/issues/6034 | load_dataset hangs on WSL | Even if a dataset is cached, we still make requests to check whether the cache is up-to-date. [This](https://huggingface.co/docs/datasets/v2.13.1/en/loading#offline) section in the docs explains how to avoid them and directly load the cached version. | ### Describe the bug
load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local cache is available so I am not... | 28 | 33 | load_dataset hangs on WSL
### Describe the bug
load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local ca... | [
-1.2228678464889526,
-0.9821550846099854,
-0.7424820065498352,
1.4652351140975952,
-0.06262974441051483,
-1.1958858966827393,
0.06592405587434769,
-1.0115690231323242,
1.7043321132659912,
-0.717893660068512,
0.24771787226200104,
-1.6565674543380737,
-0.04175431653857231,
-0.518666923046112... |
https://github.com/huggingface/datasets/issues/6034 | load_dataset hangs on WSL | Thanks - that works! However it doesn't resolve the original issue (but I am not sure if it is a WSL problem) | ### Describe the bug
load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local cache is available so I am not... | 28 | 22 | load_dataset hangs on WSL
### Describe the bug
load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local ca... | [
-1.2606147527694702,
-1.025841474533081,
-0.7423518300056458,
1.4611889123916626,
-0.08388272672891617,
-1.1834825277328491,
0.060363709926605225,
-1.0104409456253052,
1.7091079950332642,
-0.726340115070343,
0.27476975321769714,
-1.6232661008834839,
0.004051377065479755,
-0.540695071220398... |
https://github.com/huggingface/datasets/issues/6034 | load_dataset hangs on WSL | We use `requests` to make HTTP requests (and `aiohttp` in the streaming mode), so I don't think we can provide much help regarding the socket issue (it probably has something to do with WSL). | ### Describe the bug
load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local cache is available so I am not... | 28 | 34 | load_dataset hangs on WSL
### Describe the bug
load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local ca... | [
-1.2554733753204346,
-1.0141922235488892,
-0.7413086295127869,
1.4800934791564941,
-0.10498398542404175,
-1.1705819368362427,
0.08941688388586044,
-1.0425829887390137,
1.704733967781067,
-0.7188597321510315,
0.2837793231010437,
-1.6243011951446533,
-0.002514083869755268,
-0.527570307254791... |
https://github.com/huggingface/datasets/issues/6032 | DownloadConfig.proxies not work when load_dataset_builder calling HfApi.dataset_info | `HfApi` comes from the `huggingface_hub` package. You can use [this](https://huggingface.co/docs/huggingface_hub/v0.16.3/en/package_reference/utilities#huggingface_hub.configure_http_backend) utility to change the `huggingface_hub`'s `Session` proxies (see the example).
We plan to implement https://github.com/huggin... | ### Describe the bug
```python
download_config = DownloadConfig(proxies={'https': '<my proxy>'})
builder = load_dataset_builder(..., download_config=download_config)
```
But, when getting the dataset_info from HfApi, the http requests not using the proxies.
### Steps to reproduce the bug
1. Setup proxies i... | 29 | 32 | DownloadConfig.proxies not work when load_dataset_builder calling HfApi.dataset_info
### Describe the bug
```python
download_config = DownloadConfig(proxies={'https': '<my proxy>'})
builder = load_dataset_builder(..., download_config=download_config)
```
But, when getting the dataset_info from HfApi, the htt... | [
-1.1104506254196167,
-0.900575578212738,
-0.676376461982727,
1.509926438331604,
-0.10628755390644073,
-1.2705026865005493,
0.15166087448596954,
-1.0402348041534424,
1.57485032081604,
-0.6813511252403259,
0.24466240406036377,
-1.692107081413269,
-0.0123829310759902,
-0.5783460140228271,
-... |
https://github.com/huggingface/datasets/issues/6025 | Using a dataset for a use other than it was intended for. | I've opened a PR with a fix. In the meantime, you can avoid the error by deleting `task_templates` with `dataset.info.task_templates = None` before the `interleave_datasets` call.
` | ### Describe the bug
Hi, I want to use the rotten tomatoes dataset but for a task other than classification, but when I interleave the dataset, it throws ```'ValueError: Column label is not present in features.'```. It seems that the label_col must be there in the dataset for some reason?
Here is the full stacktra... | 31 | 27 | Using a dataset for a use other than it was intended for.
### Describe the bug
Hi, I want to use the rotten tomatoes dataset but for a task other than classification, but when I interleave the dataset, it throws ```'ValueError: Column label is not present in features.'```. It seems that the label_col must be there i... | [
-1.1893870830535889,
-0.8744972348213196,
-0.6626822352409363,
1.4390053749084473,
-0.15856659412384033,
-1.3036210536956787,
0.06782419234514236,
-1.0595897436141968,
1.5827667713165283,
-0.763502836227417,
0.24396421015262604,
-1.686629056930542,
-0.07698502391576767,
-0.513750433921814,... |
https://github.com/huggingface/datasets/issues/6020 | Inconsistent "The features can't be aligned" error when combining map, multiprocessing, and variable length outputs | This scenario currently requires explicitly passing the target features (to avoid the error):
```python
import datasets
...
features = dataset.features
features["output"] = = [{"test": datasets.Value("int64")}]
test2 = dataset.map(lambda row, idx: test_func(row, idx), with_indices=True, num_proc=32, features... | ### Describe the bug
I'm using a dataset with map and multiprocessing to run a function that returned a variable length list of outputs. This output list may be empty. Normally this is handled fine, but there is an edge case that crops up when using multiprocessing. In some cases, an empty list result ends up in a dat... | 33 | 36 | Inconsistent "The features can't be aligned" error when combining map, multiprocessing, and variable length outputs
### Describe the bug
I'm using a dataset with map and multiprocessing to run a function that returned a variable length list of outputs. This output list may be empty. Normally this is handled fine, bu... | [
-1.2154974937438965,
-1.0155515670776367,
-0.7719004154205322,
1.5813425779342651,
-0.2697145640850067,
-1.105621576309204,
0.07095621526241302,
-1.0311253070831299,
1.575103759765625,
-0.7578650116920471,
0.27627527713775635,
-1.6430742740631104,
0.032659441232681274,
-0.7017241716384888,... |
https://github.com/huggingface/datasets/issues/6109 | Problems in downloading Amazon reviews from HF | Thanks for reporting, @610v4nn1.
Indeed, the source data files are no longer available. We have contacted the authors of the dataset and they report that Amazon has decided to stop distributing the multilingual reviews dataset.
We are adding a notification about this issue to the dataset card.
See: https://hug... | ### Describe the bug
I have a script downloading `amazon_reviews_multi`.
When the download starts, I get
```
Downloading data files: 0%| | 0/1 [00:00<?, ?it/s]
Downloading data: 243B [00:00, 1.43MB/s]
Downloading data files: 100%|ββββββββββ| 1/1 [00:01<00:00, 1.54s/it]
Extracting data files: 100%... | 35 | 49 | Problems in downloading Amazon reviews from HF
### Describe the bug
I have a script downloading `amazon_reviews_multi`.
When the download starts, I get
```
Downloading data files: 0%| | 0/1 [00:00<?, ?it/s]
Downloading data: 243B [00:00, 1.43MB/s]
Downloading data files: 100%|ββββββββββ| 1/1 [00:... | [
-1.1749063730239868,
-0.8698778748512268,
-0.8900013566017151,
1.4243459701538086,
-0.13573047518730164,
-1.2226994037628174,
0.15154093503952026,
-1.2511365413665771,
1.432878017425537,
-0.6011939644813538,
0.4040875732898712,
-1.727347731590271,
0.04589912295341492,
-0.4552827477455139,
... |
https://github.com/huggingface/datasets/issues/6108 | Loading local datasets got strangely stuck | Yesterday I waited for more than 12 hours to make sure it was really **stuck** instead of proceeding too slow. | ### Describe the bug
I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as:
```python
ds = load_dataset("json", data_files=LIST_OF_FILE_PATHS, num_proc=... | 36 | 20 | Loading local datasets got strangely stuck
### Describe the bug
I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as:
```python
ds = load_dataset("js... | [
-1.1857736110687256,
-0.9407454133033752,
-0.7441576719284058,
1.4746325016021729,
-0.11657063663005829,
-1.2399708032608032,
0.11755771934986115,
-1.0479322671890259,
1.7225706577301025,
-0.7684071063995361,
0.38269880414009094,
-1.6014268398284912,
0.03821048140525818,
-0.620981097221374... |
https://github.com/huggingface/datasets/issues/6108 | Loading local datasets got strangely stuck | I've had similar weird issues with `load_dataset` as well. Not multiple files, but dataset is quite big, about 50G. | ### Describe the bug
I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as:
```python
ds = load_dataset("json", data_files=LIST_OF_FILE_PATHS, num_proc=... | 36 | 19 | Loading local datasets got strangely stuck
### Describe the bug
I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as:
```python
ds = load_dataset("js... | [
-1.1857736110687256,
-0.9407454133033752,
-0.7441576719284058,
1.4746325016021729,
-0.11657063663005829,
-1.2399708032608032,
0.11755771934986115,
-1.0479322671890259,
1.7225706577301025,
-0.7684071063995361,
0.38269880414009094,
-1.6014268398284912,
0.03821048140525818,
-0.620981097221374... |
https://github.com/huggingface/datasets/issues/6099 | How do i get "amazon_us_reviews | Seems like the problem isn't with the library, but the dataset itself hosted on AWS S3.
Its [homepage](https://s3.amazonaws.com/amazon-reviews-pds/readme.html) returns an `AccessDenied` XML response, which is the same thing you get if you try to log the `record` that triggers the exception
```python
try:
ex... | ### Feature request
I have been trying to load 'amazon_us_dataset" but unable to do so.
`amazon_us_reviews = load_dataset('amazon_us_reviews')`
`print(amazon_us_reviews)`
> [ValueError: Config name is missing.
Please pick one among the available configs: ['Wireless_v1_00', 'Watches_v1_00', 'Video_Games_v1... | 39 | 67 | How do i get "amazon_us_reviews
### Feature request
I have been trying to load 'amazon_us_dataset" but unable to do so.
`amazon_us_reviews = load_dataset('amazon_us_reviews')`
`print(amazon_us_reviews)`
> [ValueError: Config name is missing.
Please pick one among the available configs: ['Wireless_v1_00'... | [
-1.210225224494934,
-1.2076855897903442,
-0.5899178385734558,
1.6656270027160645,
0.004276080057024956,
-1.407726764678955,
0.13049446046352386,
-1.0258605480194092,
1.3568669557571411,
-0.8192351460456848,
0.10499061644077301,
-1.3370153903961182,
-0.003387860022485256,
-0.735708951950073... |
https://github.com/huggingface/datasets/issues/6089 | AssertionError: daemonic processes are not allowed to have children | We could add a "threads" parallel backend to `datasets.parallel.parallel_backend` to support downloading with threads but note that `download_and_extract` also decompresses archives, and this is a CPU-intensive task, which is not ideal for (Python) threads (good for IO-intensive tasks). | ### Describe the bug
When I load_dataset with num_proc > 0 in a deamon process, I got an error:
```python
File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract
return self.extract(self.download(url_or_urls))
^^^^^^^^^^^^^^^^^
File "/Users... | 42 | 38 | AssertionError: daemonic processes are not allowed to have children
### Describe the bug
When I load_dataset with num_proc > 0 in a deamon process, I got an error:
```python
File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract
return self.extr... | [
-1.3304325342178345,
-0.843843400478363,
-0.6240853667259216,
1.2293169498443604,
0.023741507902741432,
-0.9174615740776062,
0.06877093762159348,
-1.0991508960723877,
1.8150602579116821,
-0.6466710567474365,
0.21736067533493042,
-1.551055669784546,
-0.1861899197101593,
-0.48568251729011536... |
https://github.com/huggingface/datasets/issues/6089 | AssertionError: daemonic processes are not allowed to have children | > We could add a "threads" parallel backend to `datasets.parallel.parallel_backend` to support downloading with threads but note that `download_and_extract` also decompresses archives, and this is a CPU-intensive task, which is not ideal for (Python) threads (good for IO-intensive tasks).
Great! Download takes more ... | ### Describe the bug
When I load_dataset with num_proc > 0 in a deamon process, I got an error:
```python
File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract
return self.extract(self.download(url_or_urls))
^^^^^^^^^^^^^^^^^
File "/Users... | 42 | 58 | AssertionError: daemonic processes are not allowed to have children
### Describe the bug
When I load_dataset with num_proc > 0 in a deamon process, I got an error:
```python
File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract
return self.extr... | [
-1.3304325342178345,
-0.843843400478363,
-0.6240853667259216,
1.2293169498443604,
0.023741507902741432,
-0.9174615740776062,
0.06877093762159348,
-1.0991508960723877,
1.8150602579116821,
-0.6466710567474365,
0.21736067533493042,
-1.551055669784546,
-0.1861899197101593,
-0.48568251729011536... |
https://github.com/huggingface/datasets/issues/6086 | Support `fsspec` in `Dataset.to_<format>` methods | I'm assuming this should just cover `to_csv`, `to_parquet`, and `to_json`, right? As `to_list` and `to_dict` just return Python objects, `to_pandas` returns a `pandas.DataFrame` and `to_sql` just inserts into a SQL DB, is that right? | Supporting this should be fairly easy.
Requested on the forum [here](https://discuss.huggingface.co/t/how-can-i-convert-a-loaded-dataset-in-to-a-parquet-file-and-save-it-to-the-s3/48353). | 44 | 34 | Support `fsspec` in `Dataset.to_<format>` methods
Supporting this should be fairly easy.
Requested on the forum [here](https://discuss.huggingface.co/t/how-can-i-convert-a-loaded-dataset-in-to-a-parquet-file-and-save-it-to-the-s3/48353).
I'm assuming this should just cover `to_csv`, `to_parquet`, and `to_json`, ... | [
-1.0311846733093262,
-0.7468673586845398,
-0.7434991002082825,
1.6008455753326416,
-0.025689486414194107,
-1.4986146688461304,
0.25253990292549133,
-1.046440601348877,
1.7896806001663208,
-0.9305419921875,
0.49904119968414307,
-1.5891073942184448,
0.06446303427219391,
-0.7518453001976013,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | When the process starts to hang, can you interrupt it with CTRL + C and paste the error stack trace here? | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 21 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2703862190246582,
-1.0200567245483398,
-0.7282028198242188,
1.5863037109375,
-0.22804024815559387,
-1.1206165552139282,
0.14635612070560455,
-0.9283672571182251,
1.6327399015426636,
-0.8532999157905579,
0.25511327385902405,
-1.5935802459716797,
0.06626027077436447,
-0.5541383624076843,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Thanks @mariosasko for your prompt response, here's the stack trace:
```
KeyboardInterrupt Traceback (most recent call last)
Cell In[12], line 4
2 t = time.time()
3 iter_ = 0
----> 4 for batch in train_dataloader:
5 #batch_proc = streaming_obj.collect_streaming_dat... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 308 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2974315881729126,
-1.0190706253051758,
-0.664889931678772,
1.5647151470184326,
-0.20822866261005402,
-1.1625877618789673,
0.14594830572605133,
-0.9586323499679565,
1.6330128908157349,
-0.8328601121902466,
0.23161981999874115,
-1.5734355449676514,
0.07717040926218033,
-0.5745692253112793... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Update: If i let it run, it eventually fails with:
```
RuntimeError Traceback (most recent call last)
Cell In[16], line 4
2 t = time.time()
3 iter_ = 0
----> 4 for batch in train_dataloader:
5 #batch_proc = streaming_obj.collect_streaming_data_batch(batch)
... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 416 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.270894169807434,
-1.0062732696533203,
-0.6609833240509033,
1.563029408454895,
-0.22833463549613953,
-1.170157790184021,
0.1303490400314331,
-0.9474954009056091,
1.6263947486877441,
-0.8306624889373779,
0.2395327091217041,
-1.5603952407836914,
0.07044895738363266,
-0.571514368057251,
-... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | PyTorch tensors cannot store `Decimal` objects. Casting the column with decimals to `float` should fix the issue. | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 17 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2665154933929443,
-1.0163578987121582,
-0.7126186490058899,
1.5815691947937012,
-0.22405150532722473,
-1.126129150390625,
0.14313547313213348,
-0.9426345229148865,
1.658698558807373,
-0.8402796387672424,
0.2593388259410858,
-1.6109161376953125,
0.07901172339916229,
-0.5566502213478088,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | I already have cast in collate_fn, in which I perform .astype(float) for each numerical field.
On the same instance, I installed a conda env with python 3.6, and this works well.
Sample:
```
def streaming_data_collate_fn(batch):
df = pd.DataFrame.from_dict(batch)
feat_vals = torch.FloatTensor(np.nan_t... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 42 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.3096867799758911,
-1.0151026248931885,
-0.654405415058136,
1.5715867280960083,
-0.2311745584011078,
-1.1411608457565308,
0.1739429086446762,
-0.9530844688415527,
1.66796875,
-0.8388645648956299,
0.2485904097557068,
-1.5883809328079224,
0.09396015852689743,
-0.5771110653877258,
-0.6925... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | `collate_fn` is applied after the `torch` formatting step, so I think the only option when working with an `IterableDataset` is to remove the `with_format` call and perform the conversion from Python values to PyTorch tensors in `collate_fn`. The standard `Dataset` supports `with_format("numpy")`, which should make thi... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 48 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2474640607833862,
-0.9869871735572815,
-0.7190800905227661,
1.585418462753296,
-0.23098866641521454,
-1.1668672561645508,
0.17919516563415527,
-0.9628450274467468,
1.6689121723175049,
-0.8653399348258972,
0.28742289543151855,
-1.6140029430389404,
0.06157983839511871,
-0.5838105082511902... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Thanks!
Python 3.10 conda-env: After replacing with_format("torch") with with_format("numpy"), the error went away. However, it was still taking over 2 minutes to load a very small batch of 64 samples with num_workers set to 32. Once I removed with_format call altogether, it is finishing in 11 seconds.
Python 3.6 ... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 83 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2701588869094849,
-0.997541069984436,
-0.6988049149513245,
1.51397705078125,
-0.25449058413505554,
-1.1525267362594604,
0.17723368108272552,
-0.9602234959602356,
1.6439114809036255,
-0.8577772378921509,
0.24207332730293274,
-1.6130080223083496,
0.08785462379455566,
-0.5229309797286987,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Can you share the `datasets` and `torch` versions installed in these conda envs?
> Once I removed with_format call altogether, it is finishing in 11 seconds.
Hmm, that's surprising. What are your dataset's `.features`? | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 34 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2404839992523193,
-1.0219470262527466,
-0.7118972539901733,
1.579450011253357,
-0.21030646562576294,
-1.133574366569519,
0.14982429146766663,
-0.9466209411621094,
1.682141661643982,
-0.8303287625312805,
0.23833730816841125,
-1.617666482925415,
0.05280683934688568,
-0.5681951642036438,
... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | Python 3.6:
datasets.__version__ 2.4.0
torch.__version__ 1.10.1+cu102
Python 3.10:
datasets.__version__ 2.14.0
torch.__version__ 2.0.0
Anonymized features are of the form (subset shown here):
{
'string_feature_i': Value(dtype='string', id=None),
'numerical_feature_i': Value(dtype='decimal128(38, 0)', id=N... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 46 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.3298096656799316,
-1.0528417825698853,
-0.6605807542800903,
1.6113710403442383,
-0.2553796172142029,
-1.1346224546432495,
0.14628013968467712,
-0.9305062890052795,
1.6034176349639893,
-0.8199122548103333,
0.21513554453849792,
-1.5423533916473389,
0.0535685159265995,
-0.5379592776298523,... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | One more thing, in python 3.10 based kernel, interestingly increasing num_workers seem to be increasing the runtime of iterating I was trying out. In python 3.10 kernel execution, I do not even see multiple CPU cores spiking unlike in 3.6.
512 batch size on 32 workers executes in 2.4 seconds on python 3.6 kernel, wh... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 61 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2805341482162476,
-1.0410552024841309,
-0.6732608079910278,
1.5353178977966309,
-0.254350870847702,
-1.1552599668502808,
0.1737479567527771,
-0.9447788596153259,
1.6417279243469238,
-0.8380560278892517,
0.23958228528499603,
-1.5837104320526123,
0.048319969326257706,
-0.5743476748466492,... |
https://github.com/huggingface/datasets/issues/6079 | Iterating over DataLoader based on HF datasets is stuck forever | **Update**: It seems the latency part is more of a multiprocessing issue with torch and some host specific issue, and I had to scourge through relevant pytorch issues, when I stumbled across these threads:
1. https://github.com/pytorch/pytorch/issues/102494
2. https://github.com/pytorch/pytorch/issues/102269
3. http... | ### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou... | 45 | 87 | Iterating over DataLoader based on HF datasets is stuck forever
### Describe the bug
I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment.
I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo... | [
-1.2661398649215698,
-1.0214083194732666,
-0.7204177379608154,
1.5396506786346436,
-0.22598449885845184,
-1.1263196468353271,
0.09709709882736206,
-0.9557428359985352,
1.616188406944275,
-0.767447292804718,
0.2418806105852127,
-1.6014881134033203,
0.014191004447638988,
-0.5700944066047668,... |
https://github.com/huggingface/datasets/issues/6078 | resume_download with streaming=True | Currently, it's not possible to efficiently resume streaming after an error. Eventually, we plan to support this for Parquet (see https://github.com/huggingface/datasets/issues/5380). | ### Describe the bug
I used:
```
dataset = load_dataset(
"oscar-corpus/OSCAR-2201",
token=True,
language="fr",
streaming=True,
split="train"
)
```
Unfortunately, the server had a problem during the training process. I saved the step my training stopped at.
But how can I resume download f... | 46 | 21 | resume_download with streaming=True
### Describe the bug
I used:
```
dataset = load_dataset(
"oscar-corpus/OSCAR-2201",
token=True,
language="fr",
streaming=True,
split="train"
)
```
Unfortunately, the server had a problem during the training process. I saved the step my training stoppe... | [
-1.2152302265167236,
-0.9335026144981384,
-0.6795473694801331,
1.4461760520935059,
-0.1448366641998291,
-1.1738520860671997,
0.18506090342998505,
-1.0231943130493164,
1.5911850929260254,
-0.8055459260940552,
0.3376910090446472,
-1.6882258653640747,
-0.015008553862571716,
-0.554406762123107... |
https://github.com/huggingface/datasets/issues/6077 | Mapping gets stuck at 99% | The `MAX_MAP_BATCH_SIZE = 1_000_000_000` hack is bad as it loads the entire dataset into RAM when performing `.map`. Instead, it's best to use `.iter(batch_size)` to iterate over the data batches and compute `mean` for each column. (`stddev` can be computed in another pass).
Also, these arrays are big, so it makes s... | ### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize the features of the dataset, ... | 47 | 62 | Mapping gets stuck at 99%
### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize t... | [
-1.2467619180679321,
-0.9692441821098328,
-0.7773353457450867,
1.3650438785552979,
-0.10279692709445953,
-1.2577497959136963,
0.0820453092455864,
-1.0771145820617676,
1.6699507236480713,
-0.7752376198768616,
0.2535783052444458,
-1.6741188764572144,
0.0001330627128481865,
-0.519218504428863... |
https://github.com/huggingface/datasets/issues/6077 | Mapping gets stuck at 99% | Hi @mariosasko !
I agree, it's an ugly hack, but it was convenient since the resulting `mean_std` could be cached by the library. For my large dataset (which doesn't fit in RAM), I'm actually using something similar to what you suggested. I got rid of the first mapping in the above scripts and replaced it with an it... | ### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize the features of the dataset, ... | 47 | 67 | Mapping gets stuck at 99%
### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize t... | [
-1.2467619180679321,
-0.9692441821098328,
-0.7773353457450867,
1.3650438785552979,
-0.10279692709445953,
-1.2577497959136963,
0.0820453092455864,
-1.0771145820617676,
1.6699507236480713,
-0.7752376198768616,
0.2535783052444458,
-1.6741188764572144,
0.0001330627128481865,
-0.519218504428863... |
https://github.com/huggingface/datasets/issues/6077 | Mapping gets stuck at 99% | Have you tried to reduce `batch_size`/`writer_batch_size` in the 2nd `.map`? Also, can you interrupt the process when it gets stuck and share the error stack trace? | ### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize the features of the dataset, ... | 47 | 26 | Mapping gets stuck at 99%
### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize t... | [
-1.2467619180679321,
-0.9692441821098328,
-0.7773353457450867,
1.3650438785552979,
-0.10279692709445953,
-1.2577497959136963,
0.0820453092455864,
-1.0771145820617676,
1.6699507236480713,
-0.7752376198768616,
0.2535783052444458,
-1.6741188764572144,
0.0001330627128481865,
-0.519218504428863... |
https://github.com/huggingface/datasets/issues/6077 | Mapping gets stuck at 99% | I think `batch_size/writer_batch_size` is already at its lowest in the 2nd `.map` since `batched=False` implies `batch_size=1` and `len(ds) = 1000 = writer_batch_size`.
Here is also a bunch of stack traces when I interrupted the process:
<details>
<summary>stack trace 1</summary>
```python
(pyg)[d623204@ro... | ### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize the features of the dataset, ... | 47 | 1,454 | Mapping gets stuck at 99%
### Describe the bug
Hi !
I'm currently working with a large (~150GB) unnormalized dataset at work.
The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it.
I want to normalize t... | [
-1.2467619180679321,
-0.9692441821098328,
-0.7773353457450867,
1.3650438785552979,
-0.10279692709445953,
-1.2577497959136963,
0.0820453092455864,
-1.0771145820617676,
1.6699507236480713,
-0.7752376198768616,
0.2535783052444458,
-1.6741188764572144,
0.0001330627128481865,
-0.519218504428863... |
https://github.com/huggingface/datasets/issues/6075 | Error loading music files using `load_dataset` | This code behaves as expected on my local machine or in Colab. Which version of `soundfile` do you have installed? MP3 requires `soundfile>=0.12.1`. | ### Describe the bug
I tried to load a music file using `datasets.load_dataset()` from the repository - https://huggingface.co/datasets/susnato/pop2piano_real_music_test
I got the following error -
```
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/susnato/anaconda3/en... | 48 | 23 | Error loading music files using `load_dataset`
### Describe the bug
I tried to load a music file using `datasets.load_dataset()` from the repository - https://huggingface.co/datasets/susnato/pop2piano_real_music_test
I got the following error -
```
Traceback (most recent call last):
File "<stdin>", line 1... | [
-1.2145147323608398,
-0.921959638595581,
-0.7346850633621216,
1.3431165218353271,
-0.05712354928255081,
-1.3600084781646729,
0.1440393477678299,
-1.0172046422958374,
1.5832189321517944,
-0.7139288187026978,
0.3244938552379608,
-1.6303914785385132,
-0.20887859165668488,
-0.43720120191574097... |
https://github.com/huggingface/datasets/issues/6073 | version2.3.2 load_dataset()data_files can't include .xxxx in path | Version 2.3.2 is over one year old, so please use the latest release (2.14.0) to get the expected behavior. Version 2.3.2 does not contain some fixes we made to fix resolving hidden files/directories (starting with a dot). | ### Describe the bug
First, I cd workdir.
Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"})
that couldn't work and
<FileNotFoundError: Unable to find
'/a/b/c/.d/train/train.jsonl' at
/a/b/c/.d/>
And I debug, it is fine in version2.1.2... | 49 | 37 | version2.3.2 load_dataset()data_files can't include .xxxx in path
### Describe the bug
First, I cd workdir.
Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"})
that couldn't work and
<FileNotFoundError: Unable to find
'/a/b/c/.d/train/tr... | [
-1.1433175802230835,
-0.7783908247947693,
-0.4615917205810547,
1.541563630104065,
-0.12804095447063446,
-1.145755648612976,
0.23625068366527557,
-1.0691317319869995,
1.5106480121612549,
-0.679083526134491,
0.15528185665607452,
-1.651146411895752,
0.1933295875787735,
-0.5603920221328735,
... |
https://github.com/huggingface/datasets/issues/6071 | storage_options provided to load_dataset not fully piping through since datasets 2.14.0 | Hi ! Thanks for reporting, I opened a PR to fix this
What filesystem are you using ? | ### Describe the bug
Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that need those options to be set.
I think this is because of the new `_prepare_path_and_sto... | 50 | 18 | storage_options provided to load_dataset not fully piping through since datasets 2.14.0
### Describe the bug
Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that ... | [
-1.1659444570541382,
-0.9065359830856323,
-0.6957008242607117,
1.4630454778671265,
-0.13110680878162384,
-1.281360149383545,
0.20910535752773285,
-1.0998634099960327,
1.6670114994049072,
-0.7348120808601379,
0.32562580704689026,
-1.6995145082473755,
0.009933054447174072,
-0.575554907321929... |
https://github.com/huggingface/datasets/issues/6071 | storage_options provided to load_dataset not fully piping through since datasets 2.14.0 | Hi @lhoestq ! Thank you so much π
It's a bit of a custom setup, but in practice I am using a [pyarrow.fs.S3FileSystem](https://arrow.apache.org/docs/python/generated/pyarrow.fs.S3FileSystem.html) (wrapped in a `fsspec.implementations.arrow.ArrowFSWrapper` [to make it](https://arrow.apache.org/docs/python/filesyste... | ### Describe the bug
Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that need those options to be set.
I think this is because of the new `_prepare_path_and_sto... | 50 | 86 | storage_options provided to load_dataset not fully piping through since datasets 2.14.0
### Describe the bug
Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that ... | [
-1.1659444570541382,
-0.9065359830856323,
-0.6957008242607117,
1.4630454778671265,
-0.13110680878162384,
-1.281360149383545,
0.20910535752773285,
-1.0998634099960327,
1.6670114994049072,
-0.7348120808601379,
0.32562580704689026,
-1.6995145082473755,
0.009933054447174072,
-0.575554907321929... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | You can list the dataset's columns with `ds.column_names` before `.map` to check whether the dataset has an `image` column. If it doesn't, then this is a bug. Otherwise, please paste the line with the `.map` call.
| ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 51 | 36 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.17133367061615,
-0.9671884775161743,
-0.9597904086112976,
1.3718503713607788,
-0.1919693946838379,
-1.2990189790725708,
0.16572007536888123,
-1.1901322603225708,
1.7549738883972168,
-0.969964325428009,
0.27610474824905396,
-1.708359956741333,
0.08246801793575287,
-0.5398345589637756,
... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | This is the piece of code I am running:
```
data_transforms = utils.get_data_augmentation(args)
image_dataset = utils.load_image_dataset(args.dataset)
def resize(examples):
examples["pixel_values"] = [image.convert("RGB").resize((300, 300)) for image in examples["image"]]
return examples
def preproce... | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 51 | 1,035 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.32200288772583,
-0.9935063719749451,
-0.7597574591636658,
1.4973571300506592,
-0.1484074741601944,
-1.2504184246063232,
0.17417672276496887,
-1.1684865951538086,
1.676836371421814,
-0.8942831754684448,
0.34366580843925476,
-1.6277977228164673,
0.02760835736989975,
-0.595337450504303,
... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | The `remove_columns=["image"]` argument in the `.map` call removes the `image` column from the output, so drop this argument to preserve it. | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 51 | 21 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.1705082654953003,
-0.9761772155761719,
-0.9978625178337097,
1.4372361898422241,
-0.19130155444145203,
-1.3484487533569336,
0.14734625816345215,
-1.2010046243667603,
1.7176117897033691,
-0.9727316498756409,
0.27451565861701965,
-1.7099891901016235,
0.08859211206436157,
-0.573822081089019... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | The problem is not with the removal of the image key. The bug is why only the labels are sent to be process, instead of all the featues or dictionary keys.
P.S. I just dropped the removal argument as you've suggested, but that didn't solve the problem, because only the labels are being sent to be processed | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 51 | 57 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.1786153316497803,
-0.9939639568328857,
-0.9947094917297363,
1.3766529560089111,
-0.24062664806842804,
-1.2966949939727783,
0.09172168374061584,
-1.222991704940796,
1.6651229858398438,
-0.9775608777999878,
0.26388615369796753,
-1.6799390316009521,
0.0680551677942276,
-0.5430256724357605,... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | All the `image_dataset.column_names` after the `map` call should also be present in `preprocess_train `/`preprocess_val` unless (input) `columns` in `set_transform` are specified.
If that's not the case, we need a full reproducer (not snippets) with the environment info. | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 51 | 37 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.1579222679138184,
-0.939723789691925,
-0.9462810754776001,
1.4220147132873535,
-0.1891966313123703,
-1.2969825267791748,
0.16110125184059143,
-1.1809947490692139,
1.7300562858581543,
-1.0112223625183105,
0.29581353068351746,
-1.6760975122451782,
0.04168058559298515,
-0.615085780620575,
... |
https://github.com/huggingface/datasets/issues/6069 | KeyError: dataset has no key "image" | I have resolved the error after including a collate function as indicated in the Quick Start session of the Datasets docs.:
Here is what I did:
```
data_transforms = utils.get_data_augmentation(args)
image_dataset = utils.load_image_dataset(args.dataset)
def preprocess_train(example_batch):
example_batch[... | ### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ... | 51 | 139 | KeyError: dataset has no key "image"
### Describe the bug
I've loaded a local image dataset with:
`ds = laod_dataset("imagefolder", data_dir=path-to-data)`
And defined a transform to process the data, following the Datasets docs.
However, I get a keyError error, indicating there's no "image" key in my dataset... | [
-1.339882731437683,
-0.9851165413856506,
-0.8010908961296082,
1.4915612936019897,
-0.19218501448631287,
-1.2128950357437134,
0.16310033202171326,
-1.1704480648040771,
1.6644611358642578,
-0.8980790972709656,
0.3183603584766388,
-1.6021389961242676,
0.006425078026950359,
-0.5804527997970581... |
https://github.com/huggingface/datasets/issues/6066 | AttributeError: '_tqdm_cls' object has no attribute '_lock' | Hi ! I opened https://github.com/huggingface/datasets/pull/6067 to add the missing `_lock`
We'll do a patch release soon, but feel free to install `datasets` from source in the meantime | ### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p... | 52 | 27 | AttributeError: '_tqdm_cls' object has no attribute '_lock'
### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "... | [
-1.3521512746810913,
-0.8465626835823059,
-0.5868315100669861,
1.283677339553833,
0.03677382320165634,
-0.8991362452507019,
0.05178180709481239,
-1.024348497390747,
1.9279330968856812,
-0.5640751123428345,
0.19435115158557892,
-1.4659757614135742,
-0.12366081774234772,
-0.3550596237182617,... |
https://github.com/huggingface/datasets/issues/6066 | AttributeError: '_tqdm_cls' object has no attribute '_lock' | I have tested the latest main, it does not work.
I add more logs to reproduce this issue, it looks like a multi threading bug:
```python
@contextmanager
def ensure_lock(tqdm_class, lock_name=""):
"""get (create if necessary) and then restore `tqdm_class`'s lock"""
import os
import threading
... | ### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p... | 52 | 184 | AttributeError: '_tqdm_cls' object has no attribute '_lock'
### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "... | [
-1.3521512746810913,
-0.8465626835823059,
-0.5868315100669861,
1.283677339553833,
0.03677382320165634,
-0.8991362452507019,
0.05178180709481239,
-1.024348497390747,
1.9279330968856812,
-0.5640751123428345,
0.19435115158557892,
-1.4659757614135742,
-0.12366081774234772,
-0.3550596237182617,... |
https://github.com/huggingface/datasets/issues/6066 | AttributeError: '_tqdm_cls' object has no attribute '_lock' | Maybe it is a bug of tqdm? I think simply use `try ... except AttributeError ...` wraps `del tqdm_class._lock` should work. | ### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p... | 52 | 21 | AttributeError: '_tqdm_cls' object has no attribute '_lock'
### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "... | [
-1.3521512746810913,
-0.8465626835823059,
-0.5868315100669861,
1.283677339553833,
0.03677382320165634,
-0.8991362452507019,
0.05178180709481239,
-1.024348497390747,
1.9279330968856812,
-0.5640751123428345,
0.19435115158557892,
-1.4659757614135742,
-0.12366081774234772,
-0.3550596237182617,... |
https://github.com/huggingface/datasets/issues/6066 | AttributeError: '_tqdm_cls' object has no attribute '_lock' | Yes it looks like a bug on their end indeed, do you want to open a PR on tqdm ?
Let me see if I can find a workaround in the meantime | ### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p... | 52 | 32 | AttributeError: '_tqdm_cls' object has no attribute '_lock'
### Describe the bug
```python
File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module
data_files = DataFilesDict.from_patterns(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "... | [
-1.3521512746810913,
-0.8465626835823059,
-0.5868315100669861,
1.283677339553833,
0.03677382320165634,
-0.8991362452507019,
0.05178180709481239,
-1.024348497390747,
1.9279330968856812,
-0.5640751123428345,
0.19435115158557892,
-1.4659757614135742,
-0.12366081774234772,
-0.3550596237182617,... |
https://github.com/huggingface/datasets/issues/6060 | Dataset.map() execute twice when in PyTorch DDP mode | Sorry for asking a duplicate question about `num_proc`, I searched the forum and find the solution.
But I still can't make the trick with `torch.distributed.barrier()` to only map at the main process work. The [post on forum]( https://discuss.huggingface.co/t/slow-processing-with-map-when-using-deepspeed-or-fairscal... | ### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W... | 53 | 40 | Dataset.map() execute twice when in PyTorch DDP mode
### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only... | [
-1.3185489177703857,
-0.9868546724319458,
-0.5761328935623169,
1.4660086631774902,
-0.17576950788497925,
-1.1413180828094482,
0.1571463942527771,
-1.0458317995071411,
1.5937193632125854,
-0.8820594549179077,
0.28995469212532043,
-1.597267985343933,
0.055676743388175964,
-0.522131085395813,... |
https://github.com/huggingface/datasets/issues/6060 | Dataset.map() execute twice when in PyTorch DDP mode | If it does the `map` twice then it means the hash of your map function is not some same between your two processes.
Can you make sure your map functions have the same hash in different processes ?
```python
from datasets.fingerprint import Hasher
print(Hasher.hash(lambda x: cut_reorder_keys(x, num_stations_li... | ### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W... | 53 | 95 | Dataset.map() execute twice when in PyTorch DDP mode
### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only... | [
-1.3185489177703857,
-0.9868546724319458,
-0.5761328935623169,
1.4660086631774902,
-0.17576950788497925,
-1.1413180828094482,
0.1571463942527771,
-1.0458317995071411,
1.5937193632125854,
-0.8820594549179077,
0.28995469212532043,
-1.597267985343933,
0.055676743388175964,
-0.522131085395813,... |
https://github.com/huggingface/datasets/issues/6060 | Dataset.map() execute twice when in PyTorch DDP mode | Thanks for help! I find the fingerprint between processes don't have same hash:
```
Rank 0: Gpu 0 cut_reorder_keys fingerprint c7f47f40e9a67657
Rank 0: Gpu 0 random_shift fingerprint 240a0ce79831e7d4
Rank 1: Gpu 1 cut_reorder_keys fingerprint 20edd3d9cf284001
Rank 1: Gpu 1 random_shift fingerprint 819f7c1c18e773... | ### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W... | 53 | 90 | Dataset.map() execute twice when in PyTorch DDP mode
### Describe the bug
I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only... | [
-1.3185489177703857,
-0.9868546724319458,
-0.5761328935623169,
1.4660086631774902,
-0.17576950788497925,
-1.1413180828094482,
0.1571463942527771,
-1.0458317995071411,
1.5937193632125854,
-0.8820594549179077,
0.28995469212532043,
-1.597267985343933,
0.055676743388175964,
-0.522131085395813,... |
https://github.com/huggingface/datasets/issues/6058 | laion-coco download error | This can also mean one of the files was not downloaded correctly.
We log an erroneous file's name before raising the reader's error, so this is how you can find the problematic file. Then, you should delete it and call `load_dataset` again.
(I checked all the uploaded files, and they seem to be valid Parquet file... | ### Describe the bug
The full trace:
```
/home/bian/anaconda3/envs/sd/lib/python3.10/site-packages/datasets/load.py:1744: FutureWarning: 'ignore_verifications' was de
precated in favor of 'verification_mode' in version 2.9.1 and will be removed in 3.0.0.
You can remove this warning by passing 'verification_mode=no... | 54 | 67 | laion-coco download error
### Describe the bug
The full trace:
```
/home/bian/anaconda3/envs/sd/lib/python3.10/site-packages/datasets/load.py:1744: FutureWarning: 'ignore_verifications' was de
precated in favor of 'verification_mode' in version 2.9.1 and will be removed in 3.0.0.
You can remove this warning by p... | [
-1.228498935699463,
-0.9487924575805664,
-0.7148016691207886,
1.4198180437088013,
-0.07668789476156235,
-1.150172472000122,
0.09865113347768784,
-1.0615023374557495,
1.4952726364135742,
-0.6555337905883789,
0.2628440856933594,
-1.7097371816635132,
-0.08578065782785416,
-0.5050936937332153,... |
https://github.com/huggingface/datasets/issues/6057 | Why is the speed difference of gen example so big? | Hi!
It's hard to explain this behavior without more information. Can you profile the slower version with the following code
```python
import cProfile, pstats
from datasets import load_dataset
with cProfile.Profile() as profiler:
ds = load_dataset(...)
stats = pstats.Stats(profiler).sort_stats("cumtime"... | ```python
def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir):
with open(metadata_path, 'r') as file:
metadata = json.load(file)
for idx, item in enumerate(metadata):
image_path = item.get('image_path')
text_content = item.get('tex... | 55 | 44 | Why is the speed difference of gen example so big?
```python
def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir):
with open(metadata_path, 'r') as file:
metadata = json.load(file)
for idx, item in enumerate(metadata):
image_path = item.get('i... | [
-1.216934323310852,
-0.9637471437454224,
-0.751177191734314,
1.5022461414337158,
-0.20957061648368835,
-1.2565745115280151,
0.1558200716972351,
-1.0771048069000244,
1.6094040870666504,
-0.8362681269645691,
0.3122471272945404,
-1.6983448266983032,
-0.053490594029426575,
-0.6072121262550354,... |
https://github.com/huggingface/datasets/issues/6051 | Skipping shard in the remote repo and resume upload | Hi! `_select_contiguous` fetches a (zero-copy) slice of the dataset's Arrow table to build a shard, so I don't think this part is the problem. To me, the issue seems to be the step where we embed external image files' bytes (a lot of file reads). You can use `.map` with multiprocessing to perform this step before `push... | ### Describe the bug
For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading.
From my understanding, the problem is in this part of the code:
arrow_dataset.py
```python
for index, shard in logging.tqdm(
enume... | 57 | 111 | Skipping shard in the remote repo and resume upload
### Describe the bug
For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading.
From my understanding, the problem is in this part of the code:
arrow_dataset.py
```python... | [
-1.2337567806243896,
-0.90196293592453,
-0.678459882736206,
1.487384557723999,
-0.22792723774909973,
-1.3088724613189697,
0.11573110520839691,
-1.0518202781677246,
1.6316242218017578,
-0.8330173492431641,
0.36152344942092896,
-1.6441075801849365,
0.030849676579236984,
-0.6168583035469055,
... |
https://github.com/huggingface/datasets/issues/6051 | Skipping shard in the remote repo and resume upload | Hi, thanks, this solution saves some time.
But can't we avoid embedding all external image files bytes with each push, skipping the images that have already been pushed into the repo?
Edit: Ok I missed the part of cache it manually on the disk the first time, this solves the problem. Thank you | ### Describe the bug
For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading.
From my understanding, the problem is in this part of the code:
arrow_dataset.py
```python
for index, shard in logging.tqdm(
enume... | 57 | 53 | Skipping shard in the remote repo and resume upload
### Describe the bug
For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading.
From my understanding, the problem is in this part of the code:
arrow_dataset.py
```python... | [
-1.228704571723938,
-0.9043618440628052,
-0.6839578747749329,
1.5109764337539673,
-0.23591430485248566,
-1.3173547983169556,
0.1201028823852539,
-1.0454732179641724,
1.6178052425384521,
-0.835773766040802,
0.34961482882499695,
-1.637833833694458,
0.06114544719457626,
-0.597659170627594,
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.