html_url
stringlengths
51
51
title
stringlengths
6
280
comments
stringlengths
67
24.7k
body
stringlengths
51
36.2k
__index_level_0__
int64
1
1.17k
comment_length
int64
16
1.45k
text
stringlengths
190
38.3k
embeddings
list
https://github.com/huggingface/datasets/issues/6109
Problems in downloading Amazon reviews from HF
Thanks for reporting, @610v4nn1. Indeed, the source data files are no longer available. We have contacted the authors of the dataset and they report that Amazon has decided to stop distributing the multilingual reviews dataset. We are adding a notification about this issue to the dataset card. See: https://hug...
### Describe the bug I have a script downloading `amazon_reviews_multi`. When the download starts, I get ``` Downloading data files: 0%| | 0/1 [00:00<?, ?it/s] Downloading data: 243B [00:00, 1.43MB/s] Downloading data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:01<00:00, 1.54s/it] Extracting data files: 100%...
1
49
Problems in downloading Amazon reviews from HF ### Describe the bug I have a script downloading `amazon_reviews_multi`. When the download starts, I get ``` Downloading data files: 0%| | 0/1 [00:00<?, ?it/s] Downloading data: 243B [00:00, 1.43MB/s] Downloading data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:...
[ -1.1749063730239868, -0.8698778748512268, -0.8900013566017151, 1.4243459701538086, -0.13573047518730164, -1.2226994037628174, 0.15154093503952026, -1.2511365413665771, 1.432878017425537, -0.6011939644813538, 0.4040875732898712, -1.727347731590271, 0.04589912295341492, -0.4552827477455139, ...
https://github.com/huggingface/datasets/issues/6108
Loading local datasets got strangely stuck
Yesterday I waited for more than 12 hours to make sure it was really **stuck** instead of proceeding too slow.
### Describe the bug I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as: ```python ds = load_dataset("json", data_files=LIST_OF_FILE_PATHS, num_proc=...
2
20
Loading local datasets got strangely stuck ### Describe the bug I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as: ```python ds = load_dataset("js...
[ -1.1857736110687256, -0.9407454133033752, -0.7441576719284058, 1.4746325016021729, -0.11657063663005829, -1.2399708032608032, 0.11755771934986115, -1.0479322671890259, 1.7225706577301025, -0.7684071063995361, 0.38269880414009094, -1.6014268398284912, 0.03821048140525818, -0.620981097221374...
https://github.com/huggingface/datasets/issues/6108
Loading local datasets got strangely stuck
I've had similar weird issues with `load_dataset` as well. Not multiple files, but dataset is quite big, about 50G.
### Describe the bug I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as: ```python ds = load_dataset("json", data_files=LIST_OF_FILE_PATHS, num_proc=...
2
19
Loading local datasets got strangely stuck ### Describe the bug I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as: ```python ds = load_dataset("js...
[ -1.1857736110687256, -0.9407454133033752, -0.7441576719284058, 1.4746325016021729, -0.11657063663005829, -1.2399708032608032, 0.11755771934986115, -1.0479322671890259, 1.7225706577301025, -0.7684071063995361, 0.38269880414009094, -1.6014268398284912, 0.03821048140525818, -0.620981097221374...
https://github.com/huggingface/datasets/issues/6099
How do i get "amazon_us_reviews
Seems like the problem isn't with the library, but the dataset itself hosted on AWS S3. Its [homepage](https://s3.amazonaws.com/amazon-reviews-pds/readme.html) returns an `AccessDenied` XML response, which is the same thing you get if you try to log the `record` that triggers the exception ```python try: ex...
### Feature request I have been trying to load 'amazon_us_dataset" but unable to do so. `amazon_us_reviews = load_dataset('amazon_us_reviews')` `print(amazon_us_reviews)` > [ValueError: Config name is missing. Please pick one among the available configs: ['Wireless_v1_00', 'Watches_v1_00', 'Video_Games_v1...
5
67
How do i get "amazon_us_reviews ### Feature request I have been trying to load 'amazon_us_dataset" but unable to do so. `amazon_us_reviews = load_dataset('amazon_us_reviews')` `print(amazon_us_reviews)` > [ValueError: Config name is missing. Please pick one among the available configs: ['Wireless_v1_00'...
[ -1.210225224494934, -1.2076855897903442, -0.5899178385734558, 1.6656270027160645, 0.004276080057024956, -1.407726764678955, 0.13049446046352386, -1.0258605480194092, 1.3568669557571411, -0.8192351460456848, 0.10499061644077301, -1.3370153903961182, -0.003387860022485256, -0.735708951950073...
https://github.com/huggingface/datasets/issues/6089
AssertionError: daemonic processes are not allowed to have children
We could add a "threads" parallel backend to `datasets.parallel.parallel_backend` to support downloading with threads but note that `download_and_extract` also decompresses archives, and this is a CPU-intensive task, which is not ideal for (Python) threads (good for IO-intensive tasks).
### Describe the bug When I load_dataset with num_proc > 0 in a deamon process, I got an error: ```python File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract return self.extract(self.download(url_or_urls)) ^^^^^^^^^^^^^^^^^ File "/Users...
8
38
AssertionError: daemonic processes are not allowed to have children ### Describe the bug When I load_dataset with num_proc > 0 in a deamon process, I got an error: ```python File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract return self.extr...
[ -1.3304325342178345, -0.843843400478363, -0.6240853667259216, 1.2293169498443604, 0.023741507902741432, -0.9174615740776062, 0.06877093762159348, -1.0991508960723877, 1.8150602579116821, -0.6466710567474365, 0.21736067533493042, -1.551055669784546, -0.1861899197101593, -0.48568251729011536...
https://github.com/huggingface/datasets/issues/6089
AssertionError: daemonic processes are not allowed to have children
> We could add a "threads" parallel backend to `datasets.parallel.parallel_backend` to support downloading with threads but note that `download_and_extract` also decompresses archives, and this is a CPU-intensive task, which is not ideal for (Python) threads (good for IO-intensive tasks). Great! Download takes more ...
### Describe the bug When I load_dataset with num_proc > 0 in a deamon process, I got an error: ```python File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract return self.extract(self.download(url_or_urls)) ^^^^^^^^^^^^^^^^^ File "/Users...
8
58
AssertionError: daemonic processes are not allowed to have children ### Describe the bug When I load_dataset with num_proc > 0 in a deamon process, I got an error: ```python File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract return self.extr...
[ -1.3304325342178345, -0.843843400478363, -0.6240853667259216, 1.2293169498443604, 0.023741507902741432, -0.9174615740776062, 0.06877093762159348, -1.0991508960723877, 1.8150602579116821, -0.6466710567474365, 0.21736067533493042, -1.551055669784546, -0.1861899197101593, -0.48568251729011536...
https://github.com/huggingface/datasets/issues/6086
Support `fsspec` in `Dataset.to_<format>` methods
I'm assuming this should just cover `to_csv`, `to_parquet`, and `to_json`, right? As `to_list` and `to_dict` just return Python objects, `to_pandas` returns a `pandas.DataFrame` and `to_sql` just inserts into a SQL DB, is that right?
Supporting this should be fairly easy. Requested on the forum [here](https://discuss.huggingface.co/t/how-can-i-convert-a-loaded-dataset-in-to-a-parquet-file-and-save-it-to-the-s3/48353).
10
34
Support `fsspec` in `Dataset.to_<format>` methods Supporting this should be fairly easy. Requested on the forum [here](https://discuss.huggingface.co/t/how-can-i-convert-a-loaded-dataset-in-to-a-parquet-file-and-save-it-to-the-s3/48353). I'm assuming this should just cover `to_csv`, `to_parquet`, and `to_json`, ...
[ -1.0311846733093262, -0.7468673586845398, -0.7434991002082825, 1.6008455753326416, -0.025689486414194107, -1.4986146688461304, 0.25253990292549133, -1.046440601348877, 1.7896806001663208, -0.9305419921875, 0.49904119968414307, -1.5891073942184448, 0.06446303427219391, -0.7518453001976013, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
When the process starts to hang, can you interrupt it with CTRL + C and paste the error stack trace here?
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
21
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2703862190246582, -1.0200567245483398, -0.7282028198242188, 1.5863037109375, -0.22804024815559387, -1.1206165552139282, 0.14635612070560455, -0.9283672571182251, 1.6327399015426636, -0.8532999157905579, 0.25511327385902405, -1.5935802459716797, 0.06626027077436447, -0.5541383624076843, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Thanks @mariosasko for your prompt response, here's the stack trace: ``` KeyboardInterrupt Traceback (most recent call last) Cell In[12], line 4 2 t = time.time() 3 iter_ = 0 ----> 4 for batch in train_dataloader: 5 #batch_proc = streaming_obj.collect_streaming_dat...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
308
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2974315881729126, -1.0190706253051758, -0.664889931678772, 1.5647151470184326, -0.20822866261005402, -1.1625877618789673, 0.14594830572605133, -0.9586323499679565, 1.6330128908157349, -0.8328601121902466, 0.23161981999874115, -1.5734355449676514, 0.07717040926218033, -0.5745692253112793...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Update: If i let it run, it eventually fails with: ``` RuntimeError Traceback (most recent call last) Cell In[16], line 4 2 t = time.time() 3 iter_ = 0 ----> 4 for batch in train_dataloader: 5 #batch_proc = streaming_obj.collect_streaming_data_batch(batch) ...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
416
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.270894169807434, -1.0062732696533203, -0.6609833240509033, 1.563029408454895, -0.22833463549613953, -1.170157790184021, 0.1303490400314331, -0.9474954009056091, 1.6263947486877441, -0.8306624889373779, 0.2395327091217041, -1.5603952407836914, 0.07044895738363266, -0.571514368057251, -...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
PyTorch tensors cannot store `Decimal` objects. Casting the column with decimals to `float` should fix the issue.
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
17
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2665154933929443, -1.0163578987121582, -0.7126186490058899, 1.5815691947937012, -0.22405150532722473, -1.126129150390625, 0.14313547313213348, -0.9426345229148865, 1.658698558807373, -0.8402796387672424, 0.2593388259410858, -1.6109161376953125, 0.07901172339916229, -0.5566502213478088, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
I already have cast in collate_fn, in which I perform .astype(float) for each numerical field. On the same instance, I installed a conda env with python 3.6, and this works well. Sample: ``` def streaming_data_collate_fn(batch): df = pd.DataFrame.from_dict(batch) feat_vals = torch.FloatTensor(np.nan_t...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
42
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.3096867799758911, -1.0151026248931885, -0.654405415058136, 1.5715867280960083, -0.2311745584011078, -1.1411608457565308, 0.1739429086446762, -0.9530844688415527, 1.66796875, -0.8388645648956299, 0.2485904097557068, -1.5883809328079224, 0.09396015852689743, -0.5771110653877258, -0.6925...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
`collate_fn` is applied after the `torch` formatting step, so I think the only option when working with an `IterableDataset` is to remove the `with_format` call and perform the conversion from Python values to PyTorch tensors in `collate_fn`. The standard `Dataset` supports `with_format("numpy")`, which should make thi...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
48
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2474640607833862, -0.9869871735572815, -0.7190800905227661, 1.585418462753296, -0.23098866641521454, -1.1668672561645508, 0.17919516563415527, -0.9628450274467468, 1.6689121723175049, -0.8653399348258972, 0.28742289543151855, -1.6140029430389404, 0.06157983839511871, -0.5838105082511902...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Thanks! Python 3.10 conda-env: After replacing with_format("torch") with with_format("numpy"), the error went away. However, it was still taking over 2 minutes to load a very small batch of 64 samples with num_workers set to 32. Once I removed with_format call altogether, it is finishing in 11 seconds. Python 3.6 ...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
83
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2701588869094849, -0.997541069984436, -0.6988049149513245, 1.51397705078125, -0.25449058413505554, -1.1525267362594604, 0.17723368108272552, -0.9602234959602356, 1.6439114809036255, -0.8577772378921509, 0.24207332730293274, -1.6130080223083496, 0.08785462379455566, -0.5229309797286987, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Can you share the `datasets` and `torch` versions installed in these conda envs? > Once I removed with_format call altogether, it is finishing in 11 seconds. Hmm, that's surprising. What are your dataset's `.features`?
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
34
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2404839992523193, -1.0219470262527466, -0.7118972539901733, 1.579450011253357, -0.21030646562576294, -1.133574366569519, 0.14982429146766663, -0.9466209411621094, 1.682141661643982, -0.8303287625312805, 0.23833730816841125, -1.617666482925415, 0.05280683934688568, -0.5681951642036438, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Python 3.6: datasets.__version__ 2.4.0 torch.__version__ 1.10.1+cu102 Python 3.10: datasets.__version__ 2.14.0 torch.__version__ 2.0.0 Anonymized features are of the form (subset shown here): { 'string_feature_i': Value(dtype='string', id=None), 'numerical_feature_i': Value(dtype='decimal128(38, 0)', id=N...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
46
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.3298096656799316, -1.0528417825698853, -0.6605807542800903, 1.6113710403442383, -0.2553796172142029, -1.1346224546432495, 0.14628013968467712, -0.9305062890052795, 1.6034176349639893, -0.8199122548103333, 0.21513554453849792, -1.5423533916473389, 0.0535685159265995, -0.5379592776298523,...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
One more thing, in python 3.10 based kernel, interestingly increasing num_workers seem to be increasing the runtime of iterating I was trying out. In python 3.10 kernel execution, I do not even see multiple CPU cores spiking unlike in 3.6. 512 batch size on 32 workers executes in 2.4 seconds on python 3.6 kernel, wh...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
61
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2805341482162476, -1.0410552024841309, -0.6732608079910278, 1.5353178977966309, -0.254350870847702, -1.1552599668502808, 0.1737479567527771, -0.9447788596153259, 1.6417279243469238, -0.8380560278892517, 0.23958228528499603, -1.5837104320526123, 0.048319969326257706, -0.5743476748466492,...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
**Update**: It seems the latency part is more of a multiprocessing issue with torch and some host specific issue, and I had to scourge through relevant pytorch issues, when I stumbled across these threads: 1. https://github.com/pytorch/pytorch/issues/102494 2. https://github.com/pytorch/pytorch/issues/102269 3. http...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
11
87
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2661398649215698, -1.0214083194732666, -0.7204177379608154, 1.5396506786346436, -0.22598449885845184, -1.1263196468353271, 0.09709709882736206, -0.9557428359985352, 1.616188406944275, -0.767447292804718, 0.2418806105852127, -1.6014881134033203, 0.014191004447638988, -0.5700944066047668,...
https://github.com/huggingface/datasets/issues/6078
resume_download with streaming=True
Currently, it's not possible to efficiently resume streaming after an error. Eventually, we plan to support this for Parquet (see https://github.com/huggingface/datasets/issues/5380).
### Describe the bug I used: ``` dataset = load_dataset( "oscar-corpus/OSCAR-2201", token=True, language="fr", streaming=True, split="train" ) ``` Unfortunately, the server had a problem during the training process. I saved the step my training stopped at. But how can I resume download f...
12
21
resume_download with streaming=True ### Describe the bug I used: ``` dataset = load_dataset( "oscar-corpus/OSCAR-2201", token=True, language="fr", streaming=True, split="train" ) ``` Unfortunately, the server had a problem during the training process. I saved the step my training stoppe...
[ -1.2152302265167236, -0.9335026144981384, -0.6795473694801331, 1.4461760520935059, -0.1448366641998291, -1.1738520860671997, 0.18506090342998505, -1.0231943130493164, 1.5911850929260254, -0.8055459260940552, 0.3376910090446472, -1.6882258653640747, -0.015008553862571716, -0.554406762123107...
https://github.com/huggingface/datasets/issues/6077
Mapping gets stuck at 99%
The `MAX_MAP_BATCH_SIZE = 1_000_000_000` hack is bad as it loads the entire dataset into RAM when performing `.map`. Instead, it's best to use `.iter(batch_size)` to iterate over the data batches and compute `mean` for each column. (`stddev` can be computed in another pass). Also, these arrays are big, so it makes s...
### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize the features of the dataset, ...
13
62
Mapping gets stuck at 99% ### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize t...
[ -1.2467619180679321, -0.9692441821098328, -0.7773353457450867, 1.3650438785552979, -0.10279692709445953, -1.2577497959136963, 0.0820453092455864, -1.0771145820617676, 1.6699507236480713, -0.7752376198768616, 0.2535783052444458, -1.6741188764572144, 0.0001330627128481865, -0.519218504428863...
https://github.com/huggingface/datasets/issues/6077
Mapping gets stuck at 99%
Hi @mariosasko ! I agree, it's an ugly hack, but it was convenient since the resulting `mean_std` could be cached by the library. For my large dataset (which doesn't fit in RAM), I'm actually using something similar to what you suggested. I got rid of the first mapping in the above scripts and replaced it with an it...
### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize the features of the dataset, ...
13
67
Mapping gets stuck at 99% ### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize t...
[ -1.2467619180679321, -0.9692441821098328, -0.7773353457450867, 1.3650438785552979, -0.10279692709445953, -1.2577497959136963, 0.0820453092455864, -1.0771145820617676, 1.6699507236480713, -0.7752376198768616, 0.2535783052444458, -1.6741188764572144, 0.0001330627128481865, -0.519218504428863...
https://github.com/huggingface/datasets/issues/6077
Mapping gets stuck at 99%
Have you tried to reduce `batch_size`/`writer_batch_size` in the 2nd `.map`? Also, can you interrupt the process when it gets stuck and share the error stack trace?
### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize the features of the dataset, ...
13
26
Mapping gets stuck at 99% ### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize t...
[ -1.2467619180679321, -0.9692441821098328, -0.7773353457450867, 1.3650438785552979, -0.10279692709445953, -1.2577497959136963, 0.0820453092455864, -1.0771145820617676, 1.6699507236480713, -0.7752376198768616, 0.2535783052444458, -1.6741188764572144, 0.0001330627128481865, -0.519218504428863...
https://github.com/huggingface/datasets/issues/6077
Mapping gets stuck at 99%
I think `batch_size/writer_batch_size` is already at its lowest in the 2nd `.map` since `batched=False` implies `batch_size=1` and `len(ds) = 1000 = writer_batch_size`. Here is also a bunch of stack traces when I interrupted the process: <details> <summary>stack trace 1</summary> ```python (pyg)[d623204@ro...
### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize the features of the dataset, ...
13
1,454
Mapping gets stuck at 99% ### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize t...
[ -1.2467619180679321, -0.9692441821098328, -0.7773353457450867, 1.3650438785552979, -0.10279692709445953, -1.2577497959136963, 0.0820453092455864, -1.0771145820617676, 1.6699507236480713, -0.7752376198768616, 0.2535783052444458, -1.6741188764572144, 0.0001330627128481865, -0.519218504428863...
https://github.com/huggingface/datasets/issues/6075
Error loading music files using `load_dataset`
This code behaves as expected on my local machine or in Colab. Which version of `soundfile` do you have installed? MP3 requires `soundfile>=0.12.1`.
### Describe the bug I tried to load a music file using `datasets.load_dataset()` from the repository - https://huggingface.co/datasets/susnato/pop2piano_real_music_test I got the following error - ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/susnato/anaconda3/en...
14
23
Error loading music files using `load_dataset` ### Describe the bug I tried to load a music file using `datasets.load_dataset()` from the repository - https://huggingface.co/datasets/susnato/pop2piano_real_music_test I got the following error - ``` Traceback (most recent call last): File "<stdin>", line 1...
[ -1.2145147323608398, -0.921959638595581, -0.7346850633621216, 1.3431165218353271, -0.05712354928255081, -1.3600084781646729, 0.1440393477678299, -1.0172046422958374, 1.5832189321517944, -0.7139288187026978, 0.3244938552379608, -1.6303914785385132, -0.20887859165668488, -0.43720120191574097...
https://github.com/huggingface/datasets/issues/6073
version2.3.2 load_dataset()data_files can't include .xxxx in path
Version 2.3.2 is over one year old, so please use the latest release (2.14.0) to get the expected behavior. Version 2.3.2 does not contain some fixes we made to fix resolving hidden files/directories (starting with a dot).
### Describe the bug First, I cd workdir. Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"}) that couldn't work and <FileNotFoundError: Unable to find '/a/b/c/.d/train/train.jsonl' at /a/b/c/.d/> And I debug, it is fine in version2.1.2...
15
37
version2.3.2 load_dataset()data_files can't include .xxxx in path ### Describe the bug First, I cd workdir. Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"}) that couldn't work and <FileNotFoundError: Unable to find '/a/b/c/.d/train/tr...
[ -1.1433175802230835, -0.7783908247947693, -0.4615917205810547, 1.541563630104065, -0.12804095447063446, -1.145755648612976, 0.23625068366527557, -1.0691317319869995, 1.5106480121612549, -0.679083526134491, 0.15528185665607452, -1.651146411895752, 0.1933295875787735, -0.5603920221328735, ...
https://github.com/huggingface/datasets/issues/6071
storage_options provided to load_dataset not fully piping through since datasets 2.14.0
Hi ! Thanks for reporting, I opened a PR to fix this What filesystem are you using ?
### Describe the bug Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that need those options to be set. I think this is because of the new `_prepare_path_and_sto...
16
18
storage_options provided to load_dataset not fully piping through since datasets 2.14.0 ### Describe the bug Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that ...
[ -1.1659444570541382, -0.9065359830856323, -0.6957008242607117, 1.4630454778671265, -0.13110680878162384, -1.281360149383545, 0.20910535752773285, -1.0998634099960327, 1.6670114994049072, -0.7348120808601379, 0.32562580704689026, -1.6995145082473755, 0.009933054447174072, -0.575554907321929...
https://github.com/huggingface/datasets/issues/6071
storage_options provided to load_dataset not fully piping through since datasets 2.14.0
Hi @lhoestq ! Thank you so much πŸ™Œ It's a bit of a custom setup, but in practice I am using a [pyarrow.fs.S3FileSystem](https://arrow.apache.org/docs/python/generated/pyarrow.fs.S3FileSystem.html) (wrapped in a `fsspec.implementations.arrow.ArrowFSWrapper` [to make it](https://arrow.apache.org/docs/python/filesyste...
### Describe the bug Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that need those options to be set. I think this is because of the new `_prepare_path_and_sto...
16
86
storage_options provided to load_dataset not fully piping through since datasets 2.14.0 ### Describe the bug Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that ...
[ -1.1659444570541382, -0.9065359830856323, -0.6957008242607117, 1.4630454778671265, -0.13110680878162384, -1.281360149383545, 0.20910535752773285, -1.0998634099960327, 1.6670114994049072, -0.7348120808601379, 0.32562580704689026, -1.6995145082473755, 0.009933054447174072, -0.575554907321929...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
You can list the dataset's columns with `ds.column_names` before `.map` to check whether the dataset has an `image` column. If it doesn't, then this is a bug. Otherwise, please paste the line with the `.map` call.
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
17
36
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.17133367061615, -0.9671884775161743, -0.9597904086112976, 1.3718503713607788, -0.1919693946838379, -1.2990189790725708, 0.16572007536888123, -1.1901322603225708, 1.7549738883972168, -0.969964325428009, 0.27610474824905396, -1.708359956741333, 0.08246801793575287, -0.5398345589637756, ...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
This is the piece of code I am running: ``` data_transforms = utils.get_data_augmentation(args) image_dataset = utils.load_image_dataset(args.dataset) def resize(examples): examples["pixel_values"] = [image.convert("RGB").resize((300, 300)) for image in examples["image"]] return examples def preproce...
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
17
1,035
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.32200288772583, -0.9935063719749451, -0.7597574591636658, 1.4973571300506592, -0.1484074741601944, -1.2504184246063232, 0.17417672276496887, -1.1684865951538086, 1.676836371421814, -0.8942831754684448, 0.34366580843925476, -1.6277977228164673, 0.02760835736989975, -0.595337450504303, ...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
The `remove_columns=["image"]` argument in the `.map` call removes the `image` column from the output, so drop this argument to preserve it.
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
17
21
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.1705082654953003, -0.9761772155761719, -0.9978625178337097, 1.4372361898422241, -0.19130155444145203, -1.3484487533569336, 0.14734625816345215, -1.2010046243667603, 1.7176117897033691, -0.9727316498756409, 0.27451565861701965, -1.7099891901016235, 0.08859211206436157, -0.573822081089019...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
The problem is not with the removal of the image key. The bug is why only the labels are sent to be process, instead of all the featues or dictionary keys. P.S. I just dropped the removal argument as you've suggested, but that didn't solve the problem, because only the labels are being sent to be processed
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
17
57
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.1786153316497803, -0.9939639568328857, -0.9947094917297363, 1.3766529560089111, -0.24062664806842804, -1.2966949939727783, 0.09172168374061584, -1.222991704940796, 1.6651229858398438, -0.9775608777999878, 0.26388615369796753, -1.6799390316009521, 0.0680551677942276, -0.5430256724357605,...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
All the `image_dataset.column_names` after the `map` call should also be present in `preprocess_train `/`preprocess_val` unless (input) `columns` in `set_transform` are specified. If that's not the case, we need a full reproducer (not snippets) with the environment info.
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
17
37
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.1579222679138184, -0.939723789691925, -0.9462810754776001, 1.4220147132873535, -0.1891966313123703, -1.2969825267791748, 0.16110125184059143, -1.1809947490692139, 1.7300562858581543, -1.0112223625183105, 0.29581353068351746, -1.6760975122451782, 0.04168058559298515, -0.615085780620575, ...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
I have resolved the error after including a collate function as indicated in the Quick Start session of the Datasets docs.: Here is what I did: ``` data_transforms = utils.get_data_augmentation(args) image_dataset = utils.load_image_dataset(args.dataset) def preprocess_train(example_batch): example_batch[...
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
17
139
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.339882731437683, -0.9851165413856506, -0.8010908961296082, 1.4915612936019897, -0.19218501448631287, -1.2128950357437134, 0.16310033202171326, -1.1704480648040771, 1.6644611358642578, -0.8980790972709656, 0.3183603584766388, -1.6021389961242676, 0.006425078026950359, -0.5804527997970581...
https://github.com/huggingface/datasets/issues/6066
AttributeError: '_tqdm_cls' object has no attribute '_lock'
Hi ! I opened https://github.com/huggingface/datasets/pull/6067 to add the missing `_lock` We'll do a patch release soon, but feel free to install `datasets` from source in the meantime
### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p...
18
27
AttributeError: '_tqdm_cls' object has no attribute '_lock' ### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...
[ -1.3521512746810913, -0.8465626835823059, -0.5868315100669861, 1.283677339553833, 0.03677382320165634, -0.8991362452507019, 0.05178180709481239, -1.024348497390747, 1.9279330968856812, -0.5640751123428345, 0.19435115158557892, -1.4659757614135742, -0.12366081774234772, -0.3550596237182617,...
https://github.com/huggingface/datasets/issues/6066
AttributeError: '_tqdm_cls' object has no attribute '_lock'
I have tested the latest main, it does not work. I add more logs to reproduce this issue, it looks like a multi threading bug: ```python @contextmanager def ensure_lock(tqdm_class, lock_name=""): """get (create if necessary) and then restore `tqdm_class`'s lock""" import os import threading ...
### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p...
18
184
AttributeError: '_tqdm_cls' object has no attribute '_lock' ### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...
[ -1.3521512746810913, -0.8465626835823059, -0.5868315100669861, 1.283677339553833, 0.03677382320165634, -0.8991362452507019, 0.05178180709481239, -1.024348497390747, 1.9279330968856812, -0.5640751123428345, 0.19435115158557892, -1.4659757614135742, -0.12366081774234772, -0.3550596237182617,...
https://github.com/huggingface/datasets/issues/6066
AttributeError: '_tqdm_cls' object has no attribute '_lock'
Maybe it is a bug of tqdm? I think simply use `try ... except AttributeError ...` wraps `del tqdm_class._lock` should work.
### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p...
18
21
AttributeError: '_tqdm_cls' object has no attribute '_lock' ### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...
[ -1.3521512746810913, -0.8465626835823059, -0.5868315100669861, 1.283677339553833, 0.03677382320165634, -0.8991362452507019, 0.05178180709481239, -1.024348497390747, 1.9279330968856812, -0.5640751123428345, 0.19435115158557892, -1.4659757614135742, -0.12366081774234772, -0.3550596237182617,...
https://github.com/huggingface/datasets/issues/6066
AttributeError: '_tqdm_cls' object has no attribute '_lock'
Yes it looks like a bug on their end indeed, do you want to open a PR on tqdm ? Let me see if I can find a workaround in the meantime
### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p...
18
32
AttributeError: '_tqdm_cls' object has no attribute '_lock' ### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...
[ -1.3521512746810913, -0.8465626835823059, -0.5868315100669861, 1.283677339553833, 0.03677382320165634, -0.8991362452507019, 0.05178180709481239, -1.024348497390747, 1.9279330968856812, -0.5640751123428345, 0.19435115158557892, -1.4659757614135742, -0.12366081774234772, -0.3550596237182617,...
https://github.com/huggingface/datasets/issues/6060
Dataset.map() execute twice when in PyTorch DDP mode
Sorry for asking a duplicate question about `num_proc`, I searched the forum and find the solution. But I still can't make the trick with `torch.distributed.barrier()` to only map at the main process work. The [post on forum]( https://discuss.huggingface.co/t/slow-processing-with-map-when-using-deepspeed-or-fairscal...
### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W...
19
40
Dataset.map() execute twice when in PyTorch DDP mode ### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only...
[ -1.3185489177703857, -0.9868546724319458, -0.5761328935623169, 1.4660086631774902, -0.17576950788497925, -1.1413180828094482, 0.1571463942527771, -1.0458317995071411, 1.5937193632125854, -0.8820594549179077, 0.28995469212532043, -1.597267985343933, 0.055676743388175964, -0.522131085395813,...
https://github.com/huggingface/datasets/issues/6060
Dataset.map() execute twice when in PyTorch DDP mode
If it does the `map` twice then it means the hash of your map function is not some same between your two processes. Can you make sure your map functions have the same hash in different processes ? ```python from datasets.fingerprint import Hasher print(Hasher.hash(lambda x: cut_reorder_keys(x, num_stations_li...
### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W...
19
95
Dataset.map() execute twice when in PyTorch DDP mode ### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only...
[ -1.3185489177703857, -0.9868546724319458, -0.5761328935623169, 1.4660086631774902, -0.17576950788497925, -1.1413180828094482, 0.1571463942527771, -1.0458317995071411, 1.5937193632125854, -0.8820594549179077, 0.28995469212532043, -1.597267985343933, 0.055676743388175964, -0.522131085395813,...
https://github.com/huggingface/datasets/issues/6060
Dataset.map() execute twice when in PyTorch DDP mode
Thanks for help! I find the fingerprint between processes don't have same hash: ``` Rank 0: Gpu 0 cut_reorder_keys fingerprint c7f47f40e9a67657 Rank 0: Gpu 0 random_shift fingerprint 240a0ce79831e7d4 Rank 1: Gpu 1 cut_reorder_keys fingerprint 20edd3d9cf284001 Rank 1: Gpu 1 random_shift fingerprint 819f7c1c18e773...
### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W...
19
90
Dataset.map() execute twice when in PyTorch DDP mode ### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only...
[ -1.3185489177703857, -0.9868546724319458, -0.5761328935623169, 1.4660086631774902, -0.17576950788497925, -1.1413180828094482, 0.1571463942527771, -1.0458317995071411, 1.5937193632125854, -0.8820594549179077, 0.28995469212532043, -1.597267985343933, 0.055676743388175964, -0.522131085395813,...
https://github.com/huggingface/datasets/issues/6058
laion-coco download error
This can also mean one of the files was not downloaded correctly. We log an erroneous file's name before raising the reader's error, so this is how you can find the problematic file. Then, you should delete it and call `load_dataset` again. (I checked all the uploaded files, and they seem to be valid Parquet file...
### Describe the bug The full trace: ``` /home/bian/anaconda3/envs/sd/lib/python3.10/site-packages/datasets/load.py:1744: FutureWarning: 'ignore_verifications' was de precated in favor of 'verification_mode' in version 2.9.1 and will be removed in 3.0.0. You can remove this warning by passing 'verification_mode=no...
20
67
laion-coco download error ### Describe the bug The full trace: ``` /home/bian/anaconda3/envs/sd/lib/python3.10/site-packages/datasets/load.py:1744: FutureWarning: 'ignore_verifications' was de precated in favor of 'verification_mode' in version 2.9.1 and will be removed in 3.0.0. You can remove this warning by p...
[ -1.228498935699463, -0.9487924575805664, -0.7148016691207886, 1.4198180437088013, -0.07668789476156235, -1.150172472000122, 0.09865113347768784, -1.0615023374557495, 1.4952726364135742, -0.6555337905883789, 0.2628440856933594, -1.7097371816635132, -0.08578065782785416, -0.5050936937332153,...
https://github.com/huggingface/datasets/issues/6057
Why is the speed difference of gen example so big?
Hi! It's hard to explain this behavior without more information. Can you profile the slower version with the following code ```python import cProfile, pstats from datasets import load_dataset with cProfile.Profile() as profiler: ds = load_dataset(...) stats = pstats.Stats(profiler).sort_stats("cumtime"...
```python def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir): with open(metadata_path, 'r') as file: metadata = json.load(file) for idx, item in enumerate(metadata): image_path = item.get('image_path') text_content = item.get('tex...
21
44
Why is the speed difference of gen example so big? ```python def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir): with open(metadata_path, 'r') as file: metadata = json.load(file) for idx, item in enumerate(metadata): image_path = item.get('i...
[ -1.216934323310852, -0.9637471437454224, -0.751177191734314, 1.5022461414337158, -0.20957061648368835, -1.2565745115280151, 0.1558200716972351, -1.0771048069000244, 1.6094040870666504, -0.8362681269645691, 0.3122471272945404, -1.6983448266983032, -0.053490594029426575, -0.6072121262550354,...
https://github.com/huggingface/datasets/issues/6051
Skipping shard in the remote repo and resume upload
Hi! `_select_contiguous` fetches a (zero-copy) slice of the dataset's Arrow table to build a shard, so I don't think this part is the problem. To me, the issue seems to be the step where we embed external image files' bytes (a lot of file reads). You can use `.map` with multiprocessing to perform this step before `push...
### Describe the bug For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading. From my understanding, the problem is in this part of the code: arrow_dataset.py ```python for index, shard in logging.tqdm( enume...
23
111
Skipping shard in the remote repo and resume upload ### Describe the bug For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading. From my understanding, the problem is in this part of the code: arrow_dataset.py ```python...
[ -1.2337567806243896, -0.90196293592453, -0.678459882736206, 1.487384557723999, -0.22792723774909973, -1.3088724613189697, 0.11573110520839691, -1.0518202781677246, 1.6316242218017578, -0.8330173492431641, 0.36152344942092896, -1.6441075801849365, 0.030849676579236984, -0.6168583035469055, ...
https://github.com/huggingface/datasets/issues/6051
Skipping shard in the remote repo and resume upload
Hi, thanks, this solution saves some time. But can't we avoid embedding all external image files bytes with each push, skipping the images that have already been pushed into the repo? Edit: Ok I missed the part of cache it manually on the disk the first time, this solves the problem. Thank you
### Describe the bug For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading. From my understanding, the problem is in this part of the code: arrow_dataset.py ```python for index, shard in logging.tqdm( enume...
23
53
Skipping shard in the remote repo and resume upload ### Describe the bug For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading. From my understanding, the problem is in this part of the code: arrow_dataset.py ```python...
[ -1.228704571723938, -0.9043618440628052, -0.6839578747749329, 1.5109764337539673, -0.23591430485248566, -1.3173547983169556, 0.1201028823852539, -1.0454732179641724, 1.6178052425384521, -0.835773766040802, 0.34961482882499695, -1.637833833694458, 0.06114544719457626, -0.597659170627594, ...
https://github.com/huggingface/datasets/issues/6048
when i use datasets.load_dataset, i encounter the http connect error!
The `audiofolder` loader is not available in version `2.3.2`, hence the error. Please run the `pip install -U datasets` command to update the `datasets` installation to make `load_dataset("audiofolder", ...)` work.
### Describe the bug `common_voice_test = load_dataset("audiofolder", data_dir="./dataset/",cache_dir="./cache",split=datasets.Split.TEST)` when i run the code above, i got the error as below: -------------------------------------------- ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/...
24
30
when i use datasets.load_dataset, i encounter the http connect error! ### Describe the bug `common_voice_test = load_dataset("audiofolder", data_dir="./dataset/",cache_dir="./cache",split=datasets.Split.TEST)` when i run the code above, i got the error as below: -------------------------------------------- Connec...
[ -1.1496117115020752, -0.7972938418388367, -0.5717766880989075, 1.3664467334747314, -0.026249855756759644, -1.354591727256775, 0.0974016860127449, -0.8682535290718079, 1.5897915363311768, -0.8269336223602295, 0.35338395833969116, -1.7552576065063477, 0.027949972078204155, -0.568594634532928...
https://github.com/huggingface/datasets/issues/6043
Compression kwargs have no effect when saving datasets as csv
Hello @exs-avianello, I have reproduced the bug successfully and have understood the problem. But I am confused regarding this part of the statement, "`pandas.DataFrame.to_csv` is always called with a buf-like `path_or_buf`". Can you please elaborate on it? Thanks!
### Describe the bug Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed. A warning is raised if explicitly providing a `compression` kwarg, ...
25
38
Compression kwargs have no effect when saving datasets as csv ### Describe the bug Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed. A w...
[ -1.1168360710144043, -0.900812566280365, -0.7744753956794739, 1.420461654663086, -0.10985736548900604, -1.2823952436447144, 0.10812487453222275, -1.1060352325439453, 1.7183713912963867, -0.8180946111679077, 0.30754318833351135, -1.7111681699752808, -0.037234142422676086, -0.628463268280029...
https://github.com/huggingface/datasets/issues/6043
Compression kwargs have no effect when saving datasets as csv
Hi @aryanxk02 ! Sure, what I actually meant is that when passing a path-like `path_or_buf` here https://github.com/huggingface/datasets/blob/14f6edd9222e577dccb962ed5338b79b73502fa5/src/datasets/arrow_dataset.py#L4708-L4714 it gets converted to a file object behind the scenes here https://github.com/huggingfa...
### Describe the bug Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed. A warning is raised if explicitly providing a `compression` kwarg, ...
25
52
Compression kwargs have no effect when saving datasets as csv ### Describe the bug Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed. A w...
[ -1.1168360710144043, -0.900812566280365, -0.7744753956794739, 1.420461654663086, -0.10985736548900604, -1.2823952436447144, 0.10812487453222275, -1.1060352325439453, 1.7183713912963867, -0.8180946111679077, 0.30754318833351135, -1.7111681699752808, -0.037234142422676086, -0.628463268280029...
https://github.com/huggingface/datasets/issues/6043
Compression kwargs have no effect when saving datasets as csv
@exs-avianello When `path_or_buf` is set to None, the `to_csv()` method will return the CSV data as a string instead of saving it to a file. Hence the compression doesn't take place. I think setting `path_or_buf=self.path_or_buf` should work. What you say?
### Describe the bug Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed. A warning is raised if explicitly providing a `compression` kwarg, ...
25
40
Compression kwargs have no effect when saving datasets as csv ### Describe the bug Attempting to save a dataset as a compressed csv file, the compression kwargs provided to `.to_csv()` that get piped to panda's `pandas.DataFrame.to_csv` do not have any effect - resulting in the dataset not getting compressed. A w...
[ -1.1168360710144043, -0.900812566280365, -0.7744753956794739, 1.420461654663086, -0.10985736548900604, -1.2823952436447144, 0.10812487453222275, -1.1060352325439453, 1.7183713912963867, -0.8180946111679077, 0.30754318833351135, -1.7111681699752808, -0.037234142422676086, -0.628463268280029...
https://github.com/huggingface/datasets/issues/6038
File "/home/zhizhou/anaconda3/envs/pytorch/lib/python3.10/site-packages/datasets/builder.py", line 992, in _download_and_prepare if str(split_generator.split_info.name).lower() == "all": AttributeError: 'str' object has no attribute 'split_info'. Did you mean: 'splitlines'?
Instead of writing the loading script, you can use the built-in loader to [load JSON files](https://huggingface.co/docs/datasets/loading#json): ```python from datasets import load_dataset ds = load_dataset("json", data_files={"train": os.path.join(data_dir["train"]), "dev": os.path.join(data_dir["dev"])}) ```
Hi, I use the code below to load local file ``` def _split_generators(self, dl_manager): # TODO: This method is tasked with downloading/extracting the data and defining the splits depending on the configuration # If several configurations are possible (listed in BUILDER_CONFIGS), the configurati...
26
29
File "/home/zhizhou/anaconda3/envs/pytorch/lib/python3.10/site-packages/datasets/builder.py", line 992, in _download_and_prepare if str(split_generator.split_info.name).lower() == "all": AttributeError: 'str' object has no attribute 'split_info'. Did you mean: 'splitlines'? Hi, I use the code below to load loca...
[ -1.2361313104629517, -0.9707532525062561, -0.7087602615356445, 1.4899070262908936, -0.20801934599876404, -1.2016712427139282, 0.14858272671699524, -1.0610185861587524, 1.639115333557129, -0.8069785833358765, 0.30962151288986206, -1.6301939487457275, 0.004292638972401619, -0.565640330314636...
https://github.com/huggingface/datasets/issues/6037
Documentation links to examples are broken
These docs are outdated (version 1.2.1 is over two years old). Please refer to [this](https://huggingface.co/docs/datasets/dataset_script) version instead. Initially, we hosted datasets in this repo, but now you can find them [on the HF Hub](https://huggingface.co/datasets) (e.g. the [`ag_news`](https://huggingface....
### Describe the bug The links at the bottom of [add_dataset](https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html) to examples of specific datasets are all broken, for example - text classification: [ag_news](https://github.com/huggingface/datasets/blob/master/datasets/ag_news/ag_news.py) (original data ...
27
38
Documentation links to examples are broken ### Describe the bug The links at the bottom of [add_dataset](https://huggingface.co/docs/datasets/v1.2.1/add_dataset.html) to examples of specific datasets are all broken, for example - text classification: [ag_news](https://github.com/huggingface/datasets/blob/master...
[ -1.201838731765747, -0.8922097682952881, -0.6922095417976379, 1.4508154392242432, -0.05030794069170952, -1.3228124380111694, 0.06119988486170769, -0.9711858630180359, 1.5745917558670044, -0.6364167332649231, 0.32652443647384644, -1.6934266090393066, -0.08025117218494415, -0.499392360448837...
https://github.com/huggingface/datasets/issues/6034
load_dataset hangs on WSL
Even if a dataset is cached, we still make requests to check whether the cache is up-to-date. [This](https://huggingface.co/docs/datasets/v2.13.1/en/loading#offline) section in the docs explains how to avoid them and directly load the cached version.
### Describe the bug load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local cache is available so I am not...
28
33
load_dataset hangs on WSL ### Describe the bug load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local ca...
[ -1.2228678464889526, -0.9821550846099854, -0.7424820065498352, 1.4652351140975952, -0.06262974441051483, -1.1958858966827393, 0.06592405587434769, -1.0115690231323242, 1.7043321132659912, -0.717893660068512, 0.24771787226200104, -1.6565674543380737, -0.04175431653857231, -0.518666923046112...
https://github.com/huggingface/datasets/issues/6034
load_dataset hangs on WSL
Thanks - that works! However it doesn't resolve the original issue (but I am not sure if it is a WSL problem)
### Describe the bug load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local cache is available so I am not...
28
22
load_dataset hangs on WSL ### Describe the bug load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local ca...
[ -1.2606147527694702, -1.025841474533081, -0.7423518300056458, 1.4611889123916626, -0.08388272672891617, -1.1834825277328491, 0.060363709926605225, -1.0104409456253052, 1.7091079950332642, -0.726340115070343, 0.27476975321769714, -1.6232661008834839, 0.004051377065479755, -0.540695071220398...
https://github.com/huggingface/datasets/issues/6034
load_dataset hangs on WSL
We use `requests` to make HTTP requests (and `aiohttp` in the streaming mode), so I don't think we can provide much help regarding the socket issue (it probably has something to do with WSL).
### Describe the bug load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local cache is available so I am not...
28
34
load_dataset hangs on WSL ### Describe the bug load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local ca...
[ -1.2554733753204346, -1.0141922235488892, -0.7413086295127869, 1.4800934791564941, -0.10498398542404175, -1.1705819368362427, 0.08941688388586044, -1.0425829887390137, 1.704733967781067, -0.7188597321510315, 0.2837793231010437, -1.6243011951446533, -0.002514083869755268, -0.527570307254791...
https://github.com/huggingface/datasets/issues/6032
DownloadConfig.proxies not work when load_dataset_builder calling HfApi.dataset_info
`HfApi` comes from the `huggingface_hub` package. You can use [this](https://huggingface.co/docs/huggingface_hub/v0.16.3/en/package_reference/utilities#huggingface_hub.configure_http_backend) utility to change the `huggingface_hub`'s `Session` proxies (see the example). We plan to implement https://github.com/huggin...
### Describe the bug ```python download_config = DownloadConfig(proxies={'https': '<my proxy>'}) builder = load_dataset_builder(..., download_config=download_config) ``` But, when getting the dataset_info from HfApi, the http requests not using the proxies. ### Steps to reproduce the bug 1. Setup proxies i...
29
32
DownloadConfig.proxies not work when load_dataset_builder calling HfApi.dataset_info ### Describe the bug ```python download_config = DownloadConfig(proxies={'https': '<my proxy>'}) builder = load_dataset_builder(..., download_config=download_config) ``` But, when getting the dataset_info from HfApi, the htt...
[ -1.1104506254196167, -0.900575578212738, -0.676376461982727, 1.509926438331604, -0.10628755390644073, -1.2705026865005493, 0.15166087448596954, -1.0402348041534424, 1.57485032081604, -0.6813511252403259, 0.24466240406036377, -1.692107081413269, -0.0123829310759902, -0.5783460140228271, -...
https://github.com/huggingface/datasets/issues/6025
Using a dataset for a use other than it was intended for.
I've opened a PR with a fix. In the meantime, you can avoid the error by deleting `task_templates` with `dataset.info.task_templates = None` before the `interleave_datasets` call. `
### Describe the bug Hi, I want to use the rotten tomatoes dataset but for a task other than classification, but when I interleave the dataset, it throws ```'ValueError: Column label is not present in features.'```. It seems that the label_col must be there in the dataset for some reason? Here is the full stacktra...
31
27
Using a dataset for a use other than it was intended for. ### Describe the bug Hi, I want to use the rotten tomatoes dataset but for a task other than classification, but when I interleave the dataset, it throws ```'ValueError: Column label is not present in features.'```. It seems that the label_col must be there i...
[ -1.1893870830535889, -0.8744972348213196, -0.6626822352409363, 1.4390053749084473, -0.15856659412384033, -1.3036210536956787, 0.06782419234514236, -1.0595897436141968, 1.5827667713165283, -0.763502836227417, 0.24396421015262604, -1.686629056930542, -0.07698502391576767, -0.513750433921814,...
https://github.com/huggingface/datasets/issues/6020
Inconsistent "The features can't be aligned" error when combining map, multiprocessing, and variable length outputs
This scenario currently requires explicitly passing the target features (to avoid the error): ```python import datasets ... features = dataset.features features["output"] = = [{"test": datasets.Value("int64")}] test2 = dataset.map(lambda row, idx: test_func(row, idx), with_indices=True, num_proc=32, features...
### Describe the bug I'm using a dataset with map and multiprocessing to run a function that returned a variable length list of outputs. This output list may be empty. Normally this is handled fine, but there is an edge case that crops up when using multiprocessing. In some cases, an empty list result ends up in a dat...
33
36
Inconsistent "The features can't be aligned" error when combining map, multiprocessing, and variable length outputs ### Describe the bug I'm using a dataset with map and multiprocessing to run a function that returned a variable length list of outputs. This output list may be empty. Normally this is handled fine, bu...
[ -1.2154974937438965, -1.0155515670776367, -0.7719004154205322, 1.5813425779342651, -0.2697145640850067, -1.105621576309204, 0.07095621526241302, -1.0311253070831299, 1.575103759765625, -0.7578650116920471, 0.27627527713775635, -1.6430742740631104, 0.032659441232681274, -0.7017241716384888,...
https://github.com/huggingface/datasets/issues/6109
Problems in downloading Amazon reviews from HF
Thanks for reporting, @610v4nn1. Indeed, the source data files are no longer available. We have contacted the authors of the dataset and they report that Amazon has decided to stop distributing the multilingual reviews dataset. We are adding a notification about this issue to the dataset card. See: https://hug...
### Describe the bug I have a script downloading `amazon_reviews_multi`. When the download starts, I get ``` Downloading data files: 0%| | 0/1 [00:00<?, ?it/s] Downloading data: 243B [00:00, 1.43MB/s] Downloading data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:01<00:00, 1.54s/it] Extracting data files: 100%...
35
49
Problems in downloading Amazon reviews from HF ### Describe the bug I have a script downloading `amazon_reviews_multi`. When the download starts, I get ``` Downloading data files: 0%| | 0/1 [00:00<?, ?it/s] Downloading data: 243B [00:00, 1.43MB/s] Downloading data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:...
[ -1.1749063730239868, -0.8698778748512268, -0.8900013566017151, 1.4243459701538086, -0.13573047518730164, -1.2226994037628174, 0.15154093503952026, -1.2511365413665771, 1.432878017425537, -0.6011939644813538, 0.4040875732898712, -1.727347731590271, 0.04589912295341492, -0.4552827477455139, ...
https://github.com/huggingface/datasets/issues/6108
Loading local datasets got strangely stuck
Yesterday I waited for more than 12 hours to make sure it was really **stuck** instead of proceeding too slow.
### Describe the bug I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as: ```python ds = load_dataset("json", data_files=LIST_OF_FILE_PATHS, num_proc=...
36
20
Loading local datasets got strangely stuck ### Describe the bug I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as: ```python ds = load_dataset("js...
[ -1.1857736110687256, -0.9407454133033752, -0.7441576719284058, 1.4746325016021729, -0.11657063663005829, -1.2399708032608032, 0.11755771934986115, -1.0479322671890259, 1.7225706577301025, -0.7684071063995361, 0.38269880414009094, -1.6014268398284912, 0.03821048140525818, -0.620981097221374...
https://github.com/huggingface/datasets/issues/6108
Loading local datasets got strangely stuck
I've had similar weird issues with `load_dataset` as well. Not multiple files, but dataset is quite big, about 50G.
### Describe the bug I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as: ```python ds = load_dataset("json", data_files=LIST_OF_FILE_PATHS, num_proc=...
36
19
Loading local datasets got strangely stuck ### Describe the bug I try to use `load_dataset()` to load several local `.jsonl` files as a dataset. Every line of these files is a json structure only containing one key `text` (yeah it is a dataset for NLP model). The code snippet is as: ```python ds = load_dataset("js...
[ -1.1857736110687256, -0.9407454133033752, -0.7441576719284058, 1.4746325016021729, -0.11657063663005829, -1.2399708032608032, 0.11755771934986115, -1.0479322671890259, 1.7225706577301025, -0.7684071063995361, 0.38269880414009094, -1.6014268398284912, 0.03821048140525818, -0.620981097221374...
https://github.com/huggingface/datasets/issues/6099
How do i get "amazon_us_reviews
Seems like the problem isn't with the library, but the dataset itself hosted on AWS S3. Its [homepage](https://s3.amazonaws.com/amazon-reviews-pds/readme.html) returns an `AccessDenied` XML response, which is the same thing you get if you try to log the `record` that triggers the exception ```python try: ex...
### Feature request I have been trying to load 'amazon_us_dataset" but unable to do so. `amazon_us_reviews = load_dataset('amazon_us_reviews')` `print(amazon_us_reviews)` > [ValueError: Config name is missing. Please pick one among the available configs: ['Wireless_v1_00', 'Watches_v1_00', 'Video_Games_v1...
39
67
How do i get "amazon_us_reviews ### Feature request I have been trying to load 'amazon_us_dataset" but unable to do so. `amazon_us_reviews = load_dataset('amazon_us_reviews')` `print(amazon_us_reviews)` > [ValueError: Config name is missing. Please pick one among the available configs: ['Wireless_v1_00'...
[ -1.210225224494934, -1.2076855897903442, -0.5899178385734558, 1.6656270027160645, 0.004276080057024956, -1.407726764678955, 0.13049446046352386, -1.0258605480194092, 1.3568669557571411, -0.8192351460456848, 0.10499061644077301, -1.3370153903961182, -0.003387860022485256, -0.735708951950073...
https://github.com/huggingface/datasets/issues/6089
AssertionError: daemonic processes are not allowed to have children
We could add a "threads" parallel backend to `datasets.parallel.parallel_backend` to support downloading with threads but note that `download_and_extract` also decompresses archives, and this is a CPU-intensive task, which is not ideal for (Python) threads (good for IO-intensive tasks).
### Describe the bug When I load_dataset with num_proc > 0 in a deamon process, I got an error: ```python File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract return self.extract(self.download(url_or_urls)) ^^^^^^^^^^^^^^^^^ File "/Users...
42
38
AssertionError: daemonic processes are not allowed to have children ### Describe the bug When I load_dataset with num_proc > 0 in a deamon process, I got an error: ```python File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract return self.extr...
[ -1.3304325342178345, -0.843843400478363, -0.6240853667259216, 1.2293169498443604, 0.023741507902741432, -0.9174615740776062, 0.06877093762159348, -1.0991508960723877, 1.8150602579116821, -0.6466710567474365, 0.21736067533493042, -1.551055669784546, -0.1861899197101593, -0.48568251729011536...
https://github.com/huggingface/datasets/issues/6089
AssertionError: daemonic processes are not allowed to have children
> We could add a "threads" parallel backend to `datasets.parallel.parallel_backend` to support downloading with threads but note that `download_and_extract` also decompresses archives, and this is a CPU-intensive task, which is not ideal for (Python) threads (good for IO-intensive tasks). Great! Download takes more ...
### Describe the bug When I load_dataset with num_proc > 0 in a deamon process, I got an error: ```python File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract return self.extract(self.download(url_or_urls)) ^^^^^^^^^^^^^^^^^ File "/Users...
42
58
AssertionError: daemonic processes are not allowed to have children ### Describe the bug When I load_dataset with num_proc > 0 in a deamon process, I got an error: ```python File "/Users/codingl2k1/Work/datasets/src/datasets/download/download_manager.py", line 564, in download_and_extract return self.extr...
[ -1.3304325342178345, -0.843843400478363, -0.6240853667259216, 1.2293169498443604, 0.023741507902741432, -0.9174615740776062, 0.06877093762159348, -1.0991508960723877, 1.8150602579116821, -0.6466710567474365, 0.21736067533493042, -1.551055669784546, -0.1861899197101593, -0.48568251729011536...
https://github.com/huggingface/datasets/issues/6086
Support `fsspec` in `Dataset.to_<format>` methods
I'm assuming this should just cover `to_csv`, `to_parquet`, and `to_json`, right? As `to_list` and `to_dict` just return Python objects, `to_pandas` returns a `pandas.DataFrame` and `to_sql` just inserts into a SQL DB, is that right?
Supporting this should be fairly easy. Requested on the forum [here](https://discuss.huggingface.co/t/how-can-i-convert-a-loaded-dataset-in-to-a-parquet-file-and-save-it-to-the-s3/48353).
44
34
Support `fsspec` in `Dataset.to_<format>` methods Supporting this should be fairly easy. Requested on the forum [here](https://discuss.huggingface.co/t/how-can-i-convert-a-loaded-dataset-in-to-a-parquet-file-and-save-it-to-the-s3/48353). I'm assuming this should just cover `to_csv`, `to_parquet`, and `to_json`, ...
[ -1.0311846733093262, -0.7468673586845398, -0.7434991002082825, 1.6008455753326416, -0.025689486414194107, -1.4986146688461304, 0.25253990292549133, -1.046440601348877, 1.7896806001663208, -0.9305419921875, 0.49904119968414307, -1.5891073942184448, 0.06446303427219391, -0.7518453001976013, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
When the process starts to hang, can you interrupt it with CTRL + C and paste the error stack trace here?
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
21
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2703862190246582, -1.0200567245483398, -0.7282028198242188, 1.5863037109375, -0.22804024815559387, -1.1206165552139282, 0.14635612070560455, -0.9283672571182251, 1.6327399015426636, -0.8532999157905579, 0.25511327385902405, -1.5935802459716797, 0.06626027077436447, -0.5541383624076843, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Thanks @mariosasko for your prompt response, here's the stack trace: ``` KeyboardInterrupt Traceback (most recent call last) Cell In[12], line 4 2 t = time.time() 3 iter_ = 0 ----> 4 for batch in train_dataloader: 5 #batch_proc = streaming_obj.collect_streaming_dat...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
308
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2974315881729126, -1.0190706253051758, -0.664889931678772, 1.5647151470184326, -0.20822866261005402, -1.1625877618789673, 0.14594830572605133, -0.9586323499679565, 1.6330128908157349, -0.8328601121902466, 0.23161981999874115, -1.5734355449676514, 0.07717040926218033, -0.5745692253112793...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Update: If i let it run, it eventually fails with: ``` RuntimeError Traceback (most recent call last) Cell In[16], line 4 2 t = time.time() 3 iter_ = 0 ----> 4 for batch in train_dataloader: 5 #batch_proc = streaming_obj.collect_streaming_data_batch(batch) ...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
416
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.270894169807434, -1.0062732696533203, -0.6609833240509033, 1.563029408454895, -0.22833463549613953, -1.170157790184021, 0.1303490400314331, -0.9474954009056091, 1.6263947486877441, -0.8306624889373779, 0.2395327091217041, -1.5603952407836914, 0.07044895738363266, -0.571514368057251, -...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
PyTorch tensors cannot store `Decimal` objects. Casting the column with decimals to `float` should fix the issue.
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
17
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2665154933929443, -1.0163578987121582, -0.7126186490058899, 1.5815691947937012, -0.22405150532722473, -1.126129150390625, 0.14313547313213348, -0.9426345229148865, 1.658698558807373, -0.8402796387672424, 0.2593388259410858, -1.6109161376953125, 0.07901172339916229, -0.5566502213478088, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
I already have cast in collate_fn, in which I perform .astype(float) for each numerical field. On the same instance, I installed a conda env with python 3.6, and this works well. Sample: ``` def streaming_data_collate_fn(batch): df = pd.DataFrame.from_dict(batch) feat_vals = torch.FloatTensor(np.nan_t...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
42
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.3096867799758911, -1.0151026248931885, -0.654405415058136, 1.5715867280960083, -0.2311745584011078, -1.1411608457565308, 0.1739429086446762, -0.9530844688415527, 1.66796875, -0.8388645648956299, 0.2485904097557068, -1.5883809328079224, 0.09396015852689743, -0.5771110653877258, -0.6925...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
`collate_fn` is applied after the `torch` formatting step, so I think the only option when working with an `IterableDataset` is to remove the `with_format` call and perform the conversion from Python values to PyTorch tensors in `collate_fn`. The standard `Dataset` supports `with_format("numpy")`, which should make thi...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
48
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2474640607833862, -0.9869871735572815, -0.7190800905227661, 1.585418462753296, -0.23098866641521454, -1.1668672561645508, 0.17919516563415527, -0.9628450274467468, 1.6689121723175049, -0.8653399348258972, 0.28742289543151855, -1.6140029430389404, 0.06157983839511871, -0.5838105082511902...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Thanks! Python 3.10 conda-env: After replacing with_format("torch") with with_format("numpy"), the error went away. However, it was still taking over 2 minutes to load a very small batch of 64 samples with num_workers set to 32. Once I removed with_format call altogether, it is finishing in 11 seconds. Python 3.6 ...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
83
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2701588869094849, -0.997541069984436, -0.6988049149513245, 1.51397705078125, -0.25449058413505554, -1.1525267362594604, 0.17723368108272552, -0.9602234959602356, 1.6439114809036255, -0.8577772378921509, 0.24207332730293274, -1.6130080223083496, 0.08785462379455566, -0.5229309797286987, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Can you share the `datasets` and `torch` versions installed in these conda envs? > Once I removed with_format call altogether, it is finishing in 11 seconds. Hmm, that's surprising. What are your dataset's `.features`?
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
34
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2404839992523193, -1.0219470262527466, -0.7118972539901733, 1.579450011253357, -0.21030646562576294, -1.133574366569519, 0.14982429146766663, -0.9466209411621094, 1.682141661643982, -0.8303287625312805, 0.23833730816841125, -1.617666482925415, 0.05280683934688568, -0.5681951642036438, ...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
Python 3.6: datasets.__version__ 2.4.0 torch.__version__ 1.10.1+cu102 Python 3.10: datasets.__version__ 2.14.0 torch.__version__ 2.0.0 Anonymized features are of the form (subset shown here): { 'string_feature_i': Value(dtype='string', id=None), 'numerical_feature_i': Value(dtype='decimal128(38, 0)', id=N...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
46
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.3298096656799316, -1.0528417825698853, -0.6605807542800903, 1.6113710403442383, -0.2553796172142029, -1.1346224546432495, 0.14628013968467712, -0.9305062890052795, 1.6034176349639893, -0.8199122548103333, 0.21513554453849792, -1.5423533916473389, 0.0535685159265995, -0.5379592776298523,...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
One more thing, in python 3.10 based kernel, interestingly increasing num_workers seem to be increasing the runtime of iterating I was trying out. In python 3.10 kernel execution, I do not even see multiple CPU cores spiking unlike in 3.6. 512 batch size on 32 workers executes in 2.4 seconds on python 3.6 kernel, wh...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
61
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2805341482162476, -1.0410552024841309, -0.6732608079910278, 1.5353178977966309, -0.254350870847702, -1.1552599668502808, 0.1737479567527771, -0.9447788596153259, 1.6417279243469238, -0.8380560278892517, 0.23958228528499603, -1.5837104320526123, 0.048319969326257706, -0.5743476748466492,...
https://github.com/huggingface/datasets/issues/6079
Iterating over DataLoader based on HF datasets is stuck forever
**Update**: It seems the latency part is more of a multiprocessing issue with torch and some host specific issue, and I had to scourge through relevant pytorch issues, when I stumbled across these threads: 1. https://github.com/pytorch/pytorch/issues/102494 2. https://github.com/pytorch/pytorch/issues/102269 3. http...
### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is working for python 3.6 based conda environment seamlessly. What shou...
45
87
Iterating over DataLoader based on HF datasets is stuck forever ### Describe the bug I am using Amazon Sagemaker notebook (Amazon Linux 2) with python 3.10 based Conda environment. I have a dataset in parquet format locally. When I try to iterate over it, the loader is stuck forever. Note that the same code is wo...
[ -1.2661398649215698, -1.0214083194732666, -0.7204177379608154, 1.5396506786346436, -0.22598449885845184, -1.1263196468353271, 0.09709709882736206, -0.9557428359985352, 1.616188406944275, -0.767447292804718, 0.2418806105852127, -1.6014881134033203, 0.014191004447638988, -0.5700944066047668,...
https://github.com/huggingface/datasets/issues/6078
resume_download with streaming=True
Currently, it's not possible to efficiently resume streaming after an error. Eventually, we plan to support this for Parquet (see https://github.com/huggingface/datasets/issues/5380).
### Describe the bug I used: ``` dataset = load_dataset( "oscar-corpus/OSCAR-2201", token=True, language="fr", streaming=True, split="train" ) ``` Unfortunately, the server had a problem during the training process. I saved the step my training stopped at. But how can I resume download f...
46
21
resume_download with streaming=True ### Describe the bug I used: ``` dataset = load_dataset( "oscar-corpus/OSCAR-2201", token=True, language="fr", streaming=True, split="train" ) ``` Unfortunately, the server had a problem during the training process. I saved the step my training stoppe...
[ -1.2152302265167236, -0.9335026144981384, -0.6795473694801331, 1.4461760520935059, -0.1448366641998291, -1.1738520860671997, 0.18506090342998505, -1.0231943130493164, 1.5911850929260254, -0.8055459260940552, 0.3376910090446472, -1.6882258653640747, -0.015008553862571716, -0.554406762123107...
https://github.com/huggingface/datasets/issues/6077
Mapping gets stuck at 99%
The `MAX_MAP_BATCH_SIZE = 1_000_000_000` hack is bad as it loads the entire dataset into RAM when performing `.map`. Instead, it's best to use `.iter(batch_size)` to iterate over the data batches and compute `mean` for each column. (`stddev` can be computed in another pass). Also, these arrays are big, so it makes s...
### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize the features of the dataset, ...
47
62
Mapping gets stuck at 99% ### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize t...
[ -1.2467619180679321, -0.9692441821098328, -0.7773353457450867, 1.3650438785552979, -0.10279692709445953, -1.2577497959136963, 0.0820453092455864, -1.0771145820617676, 1.6699507236480713, -0.7752376198768616, 0.2535783052444458, -1.6741188764572144, 0.0001330627128481865, -0.519218504428863...
https://github.com/huggingface/datasets/issues/6077
Mapping gets stuck at 99%
Hi @mariosasko ! I agree, it's an ugly hack, but it was convenient since the resulting `mean_std` could be cached by the library. For my large dataset (which doesn't fit in RAM), I'm actually using something similar to what you suggested. I got rid of the first mapping in the above scripts and replaced it with an it...
### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize the features of the dataset, ...
47
67
Mapping gets stuck at 99% ### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize t...
[ -1.2467619180679321, -0.9692441821098328, -0.7773353457450867, 1.3650438785552979, -0.10279692709445953, -1.2577497959136963, 0.0820453092455864, -1.0771145820617676, 1.6699507236480713, -0.7752376198768616, 0.2535783052444458, -1.6741188764572144, 0.0001330627128481865, -0.519218504428863...
https://github.com/huggingface/datasets/issues/6077
Mapping gets stuck at 99%
Have you tried to reduce `batch_size`/`writer_batch_size` in the 2nd `.map`? Also, can you interrupt the process when it gets stuck and share the error stack trace?
### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize the features of the dataset, ...
47
26
Mapping gets stuck at 99% ### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize t...
[ -1.2467619180679321, -0.9692441821098328, -0.7773353457450867, 1.3650438785552979, -0.10279692709445953, -1.2577497959136963, 0.0820453092455864, -1.0771145820617676, 1.6699507236480713, -0.7752376198768616, 0.2535783052444458, -1.6741188764572144, 0.0001330627128481865, -0.519218504428863...
https://github.com/huggingface/datasets/issues/6077
Mapping gets stuck at 99%
I think `batch_size/writer_batch_size` is already at its lowest in the 2nd `.map` since `batched=False` implies `batch_size=1` and `len(ds) = 1000 = writer_batch_size`. Here is also a bunch of stack traces when I interrupted the process: <details> <summary>stack trace 1</summary> ```python (pyg)[d623204@ro...
### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize the features of the dataset, ...
47
1,454
Mapping gets stuck at 99% ### Describe the bug Hi ! I'm currently working with a large (~150GB) unnormalized dataset at work. The dataset is available on a read-only filesystem internally, and I use a [loading script](https://huggingface.co/docs/datasets/dataset_script) to retreive it. I want to normalize t...
[ -1.2467619180679321, -0.9692441821098328, -0.7773353457450867, 1.3650438785552979, -0.10279692709445953, -1.2577497959136963, 0.0820453092455864, -1.0771145820617676, 1.6699507236480713, -0.7752376198768616, 0.2535783052444458, -1.6741188764572144, 0.0001330627128481865, -0.519218504428863...
https://github.com/huggingface/datasets/issues/6075
Error loading music files using `load_dataset`
This code behaves as expected on my local machine or in Colab. Which version of `soundfile` do you have installed? MP3 requires `soundfile>=0.12.1`.
### Describe the bug I tried to load a music file using `datasets.load_dataset()` from the repository - https://huggingface.co/datasets/susnato/pop2piano_real_music_test I got the following error - ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/susnato/anaconda3/en...
48
23
Error loading music files using `load_dataset` ### Describe the bug I tried to load a music file using `datasets.load_dataset()` from the repository - https://huggingface.co/datasets/susnato/pop2piano_real_music_test I got the following error - ``` Traceback (most recent call last): File "<stdin>", line 1...
[ -1.2145147323608398, -0.921959638595581, -0.7346850633621216, 1.3431165218353271, -0.05712354928255081, -1.3600084781646729, 0.1440393477678299, -1.0172046422958374, 1.5832189321517944, -0.7139288187026978, 0.3244938552379608, -1.6303914785385132, -0.20887859165668488, -0.43720120191574097...
https://github.com/huggingface/datasets/issues/6073
version2.3.2 load_dataset()data_files can't include .xxxx in path
Version 2.3.2 is over one year old, so please use the latest release (2.14.0) to get the expected behavior. Version 2.3.2 does not contain some fixes we made to fix resolving hidden files/directories (starting with a dot).
### Describe the bug First, I cd workdir. Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"}) that couldn't work and <FileNotFoundError: Unable to find '/a/b/c/.d/train/train.jsonl' at /a/b/c/.d/> And I debug, it is fine in version2.1.2...
49
37
version2.3.2 load_dataset()data_files can't include .xxxx in path ### Describe the bug First, I cd workdir. Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"}) that couldn't work and <FileNotFoundError: Unable to find '/a/b/c/.d/train/tr...
[ -1.1433175802230835, -0.7783908247947693, -0.4615917205810547, 1.541563630104065, -0.12804095447063446, -1.145755648612976, 0.23625068366527557, -1.0691317319869995, 1.5106480121612549, -0.679083526134491, 0.15528185665607452, -1.651146411895752, 0.1933295875787735, -0.5603920221328735, ...
https://github.com/huggingface/datasets/issues/6071
storage_options provided to load_dataset not fully piping through since datasets 2.14.0
Hi ! Thanks for reporting, I opened a PR to fix this What filesystem are you using ?
### Describe the bug Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that need those options to be set. I think this is because of the new `_prepare_path_and_sto...
50
18
storage_options provided to load_dataset not fully piping through since datasets 2.14.0 ### Describe the bug Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that ...
[ -1.1659444570541382, -0.9065359830856323, -0.6957008242607117, 1.4630454778671265, -0.13110680878162384, -1.281360149383545, 0.20910535752773285, -1.0998634099960327, 1.6670114994049072, -0.7348120808601379, 0.32562580704689026, -1.6995145082473755, 0.009933054447174072, -0.575554907321929...
https://github.com/huggingface/datasets/issues/6071
storage_options provided to load_dataset not fully piping through since datasets 2.14.0
Hi @lhoestq ! Thank you so much πŸ™Œ It's a bit of a custom setup, but in practice I am using a [pyarrow.fs.S3FileSystem](https://arrow.apache.org/docs/python/generated/pyarrow.fs.S3FileSystem.html) (wrapped in a `fsspec.implementations.arrow.ArrowFSWrapper` [to make it](https://arrow.apache.org/docs/python/filesyste...
### Describe the bug Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that need those options to be set. I think this is because of the new `_prepare_path_and_sto...
50
86
storage_options provided to load_dataset not fully piping through since datasets 2.14.0 ### Describe the bug Since the latest release of `datasets` (`2.14.0`), custom filesystem `storage_options` passed to `load_dataset()` do not seem to propagate through all the way - leading to problems if loading data files that ...
[ -1.1659444570541382, -0.9065359830856323, -0.6957008242607117, 1.4630454778671265, -0.13110680878162384, -1.281360149383545, 0.20910535752773285, -1.0998634099960327, 1.6670114994049072, -0.7348120808601379, 0.32562580704689026, -1.6995145082473755, 0.009933054447174072, -0.575554907321929...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
You can list the dataset's columns with `ds.column_names` before `.map` to check whether the dataset has an `image` column. If it doesn't, then this is a bug. Otherwise, please paste the line with the `.map` call.
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
51
36
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.17133367061615, -0.9671884775161743, -0.9597904086112976, 1.3718503713607788, -0.1919693946838379, -1.2990189790725708, 0.16572007536888123, -1.1901322603225708, 1.7549738883972168, -0.969964325428009, 0.27610474824905396, -1.708359956741333, 0.08246801793575287, -0.5398345589637756, ...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
This is the piece of code I am running: ``` data_transforms = utils.get_data_augmentation(args) image_dataset = utils.load_image_dataset(args.dataset) def resize(examples): examples["pixel_values"] = [image.convert("RGB").resize((300, 300)) for image in examples["image"]] return examples def preproce...
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
51
1,035
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.32200288772583, -0.9935063719749451, -0.7597574591636658, 1.4973571300506592, -0.1484074741601944, -1.2504184246063232, 0.17417672276496887, -1.1684865951538086, 1.676836371421814, -0.8942831754684448, 0.34366580843925476, -1.6277977228164673, 0.02760835736989975, -0.595337450504303, ...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
The `remove_columns=["image"]` argument in the `.map` call removes the `image` column from the output, so drop this argument to preserve it.
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
51
21
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.1705082654953003, -0.9761772155761719, -0.9978625178337097, 1.4372361898422241, -0.19130155444145203, -1.3484487533569336, 0.14734625816345215, -1.2010046243667603, 1.7176117897033691, -0.9727316498756409, 0.27451565861701965, -1.7099891901016235, 0.08859211206436157, -0.573822081089019...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
The problem is not with the removal of the image key. The bug is why only the labels are sent to be process, instead of all the featues or dictionary keys. P.S. I just dropped the removal argument as you've suggested, but that didn't solve the problem, because only the labels are being sent to be processed
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
51
57
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.1786153316497803, -0.9939639568328857, -0.9947094917297363, 1.3766529560089111, -0.24062664806842804, -1.2966949939727783, 0.09172168374061584, -1.222991704940796, 1.6651229858398438, -0.9775608777999878, 0.26388615369796753, -1.6799390316009521, 0.0680551677942276, -0.5430256724357605,...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
All the `image_dataset.column_names` after the `map` call should also be present in `preprocess_train `/`preprocess_val` unless (input) `columns` in `set_transform` are specified. If that's not the case, we need a full reproducer (not snippets) with the environment info.
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
51
37
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.1579222679138184, -0.939723789691925, -0.9462810754776001, 1.4220147132873535, -0.1891966313123703, -1.2969825267791748, 0.16110125184059143, -1.1809947490692139, 1.7300562858581543, -1.0112223625183105, 0.29581353068351746, -1.6760975122451782, 0.04168058559298515, -0.615085780620575, ...
https://github.com/huggingface/datasets/issues/6069
KeyError: dataset has no key "image"
I have resolved the error after including a collate function as indicated in the Quick Start session of the Datasets docs.: Here is what I did: ``` data_transforms = utils.get_data_augmentation(args) image_dataset = utils.load_image_dataset(args.dataset) def preprocess_train(example_batch): example_batch[...
### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset. When I printed out the example_batch ...
51
139
KeyError: dataset has no key "image" ### Describe the bug I've loaded a local image dataset with: `ds = laod_dataset("imagefolder", data_dir=path-to-data)` And defined a transform to process the data, following the Datasets docs. However, I get a keyError error, indicating there's no "image" key in my dataset...
[ -1.339882731437683, -0.9851165413856506, -0.8010908961296082, 1.4915612936019897, -0.19218501448631287, -1.2128950357437134, 0.16310033202171326, -1.1704480648040771, 1.6644611358642578, -0.8980790972709656, 0.3183603584766388, -1.6021389961242676, 0.006425078026950359, -0.5804527997970581...
https://github.com/huggingface/datasets/issues/6066
AttributeError: '_tqdm_cls' object has no attribute '_lock'
Hi ! I opened https://github.com/huggingface/datasets/pull/6067 to add the missing `_lock` We'll do a patch release soon, but feel free to install `datasets` from source in the meantime
### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p...
52
27
AttributeError: '_tqdm_cls' object has no attribute '_lock' ### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...
[ -1.3521512746810913, -0.8465626835823059, -0.5868315100669861, 1.283677339553833, 0.03677382320165634, -0.8991362452507019, 0.05178180709481239, -1.024348497390747, 1.9279330968856812, -0.5640751123428345, 0.19435115158557892, -1.4659757614135742, -0.12366081774234772, -0.3550596237182617,...
https://github.com/huggingface/datasets/issues/6066
AttributeError: '_tqdm_cls' object has no attribute '_lock'
I have tested the latest main, it does not work. I add more logs to reproduce this issue, it looks like a multi threading bug: ```python @contextmanager def ensure_lock(tqdm_class, lock_name=""): """get (create if necessary) and then restore `tqdm_class`'s lock""" import os import threading ...
### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p...
52
184
AttributeError: '_tqdm_cls' object has no attribute '_lock' ### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...
[ -1.3521512746810913, -0.8465626835823059, -0.5868315100669861, 1.283677339553833, 0.03677382320165634, -0.8991362452507019, 0.05178180709481239, -1.024348497390747, 1.9279330968856812, -0.5640751123428345, 0.19435115158557892, -1.4659757614135742, -0.12366081774234772, -0.3550596237182617,...
https://github.com/huggingface/datasets/issues/6066
AttributeError: '_tqdm_cls' object has no attribute '_lock'
Maybe it is a bug of tqdm? I think simply use `try ... except AttributeError ...` wraps `del tqdm_class._lock` should work.
### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p...
52
21
AttributeError: '_tqdm_cls' object has no attribute '_lock' ### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...
[ -1.3521512746810913, -0.8465626835823059, -0.5868315100669861, 1.283677339553833, 0.03677382320165634, -0.8991362452507019, 0.05178180709481239, -1.024348497390747, 1.9279330968856812, -0.5640751123428345, 0.19435115158557892, -1.4659757614135742, -0.12366081774234772, -0.3550596237182617,...
https://github.com/huggingface/datasets/issues/6066
AttributeError: '_tqdm_cls' object has no attribute '_lock'
Yes it looks like a bug on their end indeed, do you want to open a PR on tqdm ? Let me see if I can find a workaround in the meantime
### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-p...
52
32
AttributeError: '_tqdm_cls' object has no attribute '_lock' ### Describe the bug ```python File "/Users/codingl2k1/.pyenv/versions/3.11.4/lib/python3.11/site-packages/datasets/load.py", line 1034, in get_module data_files = DataFilesDict.from_patterns( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...
[ -1.3521512746810913, -0.8465626835823059, -0.5868315100669861, 1.283677339553833, 0.03677382320165634, -0.8991362452507019, 0.05178180709481239, -1.024348497390747, 1.9279330968856812, -0.5640751123428345, 0.19435115158557892, -1.4659757614135742, -0.12366081774234772, -0.3550596237182617,...
https://github.com/huggingface/datasets/issues/6060
Dataset.map() execute twice when in PyTorch DDP mode
Sorry for asking a duplicate question about `num_proc`, I searched the forum and find the solution. But I still can't make the trick with `torch.distributed.barrier()` to only map at the main process work. The [post on forum]( https://discuss.huggingface.co/t/slow-processing-with-map-when-using-deepspeed-or-fairscal...
### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W...
53
40
Dataset.map() execute twice when in PyTorch DDP mode ### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only...
[ -1.3185489177703857, -0.9868546724319458, -0.5761328935623169, 1.4660086631774902, -0.17576950788497925, -1.1413180828094482, 0.1571463942527771, -1.0458317995071411, 1.5937193632125854, -0.8820594549179077, 0.28995469212532043, -1.597267985343933, 0.055676743388175964, -0.522131085395813,...
https://github.com/huggingface/datasets/issues/6060
Dataset.map() execute twice when in PyTorch DDP mode
If it does the `map` twice then it means the hash of your map function is not some same between your two processes. Can you make sure your map functions have the same hash in different processes ? ```python from datasets.fingerprint import Hasher print(Hasher.hash(lambda x: cut_reorder_keys(x, num_stations_li...
### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W...
53
95
Dataset.map() execute twice when in PyTorch DDP mode ### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only...
[ -1.3185489177703857, -0.9868546724319458, -0.5761328935623169, 1.4660086631774902, -0.17576950788497925, -1.1413180828094482, 0.1571463942527771, -1.0458317995071411, 1.5937193632125854, -0.8820594549179077, 0.28995469212532043, -1.597267985343933, 0.055676743388175964, -0.522131085395813,...
https://github.com/huggingface/datasets/issues/6060
Dataset.map() execute twice when in PyTorch DDP mode
Thanks for help! I find the fingerprint between processes don't have same hash: ``` Rank 0: Gpu 0 cut_reorder_keys fingerprint c7f47f40e9a67657 Rank 0: Gpu 0 random_shift fingerprint 240a0ce79831e7d4 Rank 1: Gpu 1 cut_reorder_keys fingerprint 20edd3d9cf284001 Rank 1: Gpu 1 random_shift fingerprint 819f7c1c18e773...
### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only execute map at the main process doesn't always work. W...
53
90
Dataset.map() execute twice when in PyTorch DDP mode ### Describe the bug I use `torchrun --standalone --nproc_per_node=2 train.py` to start training. And write the code following the [docs](https://huggingface.co/docs/datasets/process#distributed-usage). The trick about using `torch.distributed.barrier()` to only...
[ -1.3185489177703857, -0.9868546724319458, -0.5761328935623169, 1.4660086631774902, -0.17576950788497925, -1.1413180828094482, 0.1571463942527771, -1.0458317995071411, 1.5937193632125854, -0.8820594549179077, 0.28995469212532043, -1.597267985343933, 0.055676743388175964, -0.522131085395813,...
https://github.com/huggingface/datasets/issues/6058
laion-coco download error
This can also mean one of the files was not downloaded correctly. We log an erroneous file's name before raising the reader's error, so this is how you can find the problematic file. Then, you should delete it and call `load_dataset` again. (I checked all the uploaded files, and they seem to be valid Parquet file...
### Describe the bug The full trace: ``` /home/bian/anaconda3/envs/sd/lib/python3.10/site-packages/datasets/load.py:1744: FutureWarning: 'ignore_verifications' was de precated in favor of 'verification_mode' in version 2.9.1 and will be removed in 3.0.0. You can remove this warning by passing 'verification_mode=no...
54
67
laion-coco download error ### Describe the bug The full trace: ``` /home/bian/anaconda3/envs/sd/lib/python3.10/site-packages/datasets/load.py:1744: FutureWarning: 'ignore_verifications' was de precated in favor of 'verification_mode' in version 2.9.1 and will be removed in 3.0.0. You can remove this warning by p...
[ -1.228498935699463, -0.9487924575805664, -0.7148016691207886, 1.4198180437088013, -0.07668789476156235, -1.150172472000122, 0.09865113347768784, -1.0615023374557495, 1.4952726364135742, -0.6555337905883789, 0.2628440856933594, -1.7097371816635132, -0.08578065782785416, -0.5050936937332153,...
https://github.com/huggingface/datasets/issues/6057
Why is the speed difference of gen example so big?
Hi! It's hard to explain this behavior without more information. Can you profile the slower version with the following code ```python import cProfile, pstats from datasets import load_dataset with cProfile.Profile() as profiler: ds = load_dataset(...) stats = pstats.Stats(profiler).sort_stats("cumtime"...
```python def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir): with open(metadata_path, 'r') as file: metadata = json.load(file) for idx, item in enumerate(metadata): image_path = item.get('image_path') text_content = item.get('tex...
55
44
Why is the speed difference of gen example so big? ```python def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir): with open(metadata_path, 'r') as file: metadata = json.load(file) for idx, item in enumerate(metadata): image_path = item.get('i...
[ -1.216934323310852, -0.9637471437454224, -0.751177191734314, 1.5022461414337158, -0.20957061648368835, -1.2565745115280151, 0.1558200716972351, -1.0771048069000244, 1.6094040870666504, -0.8362681269645691, 0.3122471272945404, -1.6983448266983032, -0.053490594029426575, -0.6072121262550354,...
https://github.com/huggingface/datasets/issues/6051
Skipping shard in the remote repo and resume upload
Hi! `_select_contiguous` fetches a (zero-copy) slice of the dataset's Arrow table to build a shard, so I don't think this part is the problem. To me, the issue seems to be the step where we embed external image files' bytes (a lot of file reads). You can use `.map` with multiprocessing to perform this step before `push...
### Describe the bug For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading. From my understanding, the problem is in this part of the code: arrow_dataset.py ```python for index, shard in logging.tqdm( enume...
57
111
Skipping shard in the remote repo and resume upload ### Describe the bug For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading. From my understanding, the problem is in this part of the code: arrow_dataset.py ```python...
[ -1.2337567806243896, -0.90196293592453, -0.678459882736206, 1.487384557723999, -0.22792723774909973, -1.3088724613189697, 0.11573110520839691, -1.0518202781677246, 1.6316242218017578, -0.8330173492431641, 0.36152344942092896, -1.6441075801849365, 0.030849676579236984, -0.6168583035469055, ...
https://github.com/huggingface/datasets/issues/6051
Skipping shard in the remote repo and resume upload
Hi, thanks, this solution saves some time. But can't we avoid embedding all external image files bytes with each push, skipping the images that have already been pushed into the repo? Edit: Ok I missed the part of cache it manually on the disk the first time, this solves the problem. Thank you
### Describe the bug For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading. From my understanding, the problem is in this part of the code: arrow_dataset.py ```python for index, shard in logging.tqdm( enume...
57
53
Skipping shard in the remote repo and resume upload ### Describe the bug For some reason when I try to resume the upload of my dataset, it is very slow to reach the index of the shard from which to resume the uploading. From my understanding, the problem is in this part of the code: arrow_dataset.py ```python...
[ -1.228704571723938, -0.9043618440628052, -0.6839578747749329, 1.5109764337539673, -0.23591430485248566, -1.3173547983169556, 0.1201028823852539, -1.0454732179641724, 1.6178052425384521, -0.835773766040802, 0.34961482882499695, -1.637833833694458, 0.06114544719457626, -0.597659170627594, ...