html_url
stringlengths
51
51
title
stringlengths
6
280
comments
stringlengths
67
24.7k
body
stringlengths
51
36.2k
__index_level_0__
int64
1
1.17k
comment_length
int64
16
1.45k
text
stringlengths
190
38.3k
embeddings
list
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
It appears that the parquet files at `hf://datasets/lambdalabs/pokemon-blip-captions` don't have this metadata, and it is defined in the dataset_infos.json instead (legacy). You can get the right schema with the HF metadata this way: ```python from datasets import load_dataset_builder features = load_dataset_...
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
45
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
Btw in the future we might add support for an dedicated Image extension type in Arrow so that you won't need to add the schema metadata anymore ;)
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
28
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
Thanks @Wauplin @lhoestq for the quick reply :)! I tried your approach by passing the huggingface schema to the dask writer ``` from datasets import Features, Image, Value df = dd.read_parquet(f"hf://datasets/lambdalabs/pokemon-blip-captions",index=False) features = Features({"image": Image(), "text": Value(...
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
152
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
You don't need the dataset_infos.json file as long as you have the schema with HF metadata ;) I could also check that it works fine myself on the git revision without the dataset_infos.json file. What made you think it didn't work ?
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
42
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
> You don't need the dataset_infos.json file as long as you have the schema with HF metadata ;) I could also check that it works fine myself on the git revision without the dataset_infos.json file. > > What made you think it didn't work ? Those are two identical dataset repos where both were pushed with dask with...
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
116
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
I see ! I think it's a bug on our side - it should work without the metadata - let me investigate
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
22
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
Alright, it's fixed: https://huggingface.co/datasets/philippemo/dummy_dataset_without_schema It shows the image correctly now - even without the extra metadata :)
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
17
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
Thanks @lhoestq! I tested pushing a dataset again without the metadata and it works perfectly! I appreciate the help
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
19
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
Hi @lhoestq, I'v tried pushing another dataset again and I think the issue reappeared again: ``` df = dd.read_parquet(f"hf://datasets/lambdalabs/pokemon-blip-captions") features = datasets.Features({"image": datasets.Image(), "text": datasets.Value("string")}) schema = features.arrow_schema dd.to_parquet(df...
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
102
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
We're doing some changes in the way we're handling image parquet datasets right now. We'll include the fix from https://github.com/huggingface/datasets/pull/5921 in the new datasets-server version in the coming days
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
29
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
alright thanks for the update :), would that be part of the new release of datasets or is it something separate? if so, where can I track it?
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
28
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5869
Image Encoding Issue when submitting a Parquet Dataset
Once the new version of `datasets` is released (tomorrow probably) we'll open an issue on https://github.com/huggingface/datasets-server to update to this version :)
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
136
22
Image Encoding Issue when submitting a Parquet Dataset ### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (...
[ -1.1553672552108765, -0.8704099059104919, -0.8205088376998901, 1.4618592262268066, -0.13831359148025513, -1.3031227588653564, 0.08137144148349762, -1.077457070350647, 1.7212998867034912, -0.7961846590042114, 0.25587376952171326, -1.7589480876922607, -0.05134357139468193, -0.533249974250793...
https://github.com/huggingface/datasets/issues/5868
Is it possible to change a cached file and 're-cache' it instead of re-generating?
Arrow files/primitives (tables and arrays) are immutable, so re-generating them is the only option, I'm afraid.
### Feature request Hi, I have a huge cached file using `map`(over 500GB), and I want to change an attribution of each element, is there possible to do it using some method instead of re-generating, because `map` takes over 24 hours ### Motivation For large datasets, I think it is very important because we always f...
137
16
Is it possible to change a cached file and 're-cache' it instead of re-generating? ### Feature request Hi, I have a huge cached file using `map`(over 500GB), and I want to change an attribution of each element, is there possible to do it using some method instead of re-generating, because `map` takes over 24 hours ...
[ -1.1673550605773926, -0.9624342322349548, -0.8973989486694336, 1.5511939525604248, -0.15229423344135284, -1.2601709365844727, -0.014970440417528152, -0.9520003199577332, 1.6296809911727905, -0.8596575856208801, 0.3625255823135376, -1.636365532875061, 0.0979856550693512, -0.5727867484092712...
https://github.com/huggingface/datasets/issues/5864
Slow iteration over Torch tensors
I am highly interested performance of dataset so I ran your example as a curious user. ```python train_dataset.cast_column("x", Array3D(shape=img_shape, dtype="float32")) ``` have return values and "x" is a new column, it shoulde be ```python ds=train_dataset.cast_column("img", Array3D(shape=(3,32,32), dtype="flo...
### Describe the bug I have a problem related to this [issue](https://github.com/huggingface/datasets/issues/5841): I get a way slower iteration when using a Torch dataloader if I use vanilla Numpy tensors or if I first apply a ToTensor transform to the input. In particular, it takes 5 seconds to iterate over the vani...
139
112
Slow iteration over Torch tensors ### Describe the bug I have a problem related to this [issue](https://github.com/huggingface/datasets/issues/5841): I get a way slower iteration when using a Torch dataloader if I use vanilla Numpy tensors or if I first apply a ToTensor transform to the input. In particular, it take...
[ -0.9561881422996521, -0.947793185710907, -0.7728447914123535, 1.6186071634292603, -0.10586841404438019, -1.2148075103759766, 0.30790024995803833, -0.9178924560546875, 1.4549853801727295, -0.717391312122345, 0.01934342086315155, -1.6736693382263184, -0.0020538344979286194, -0.67579203844070...
https://github.com/huggingface/datasets/issues/5862
IndexError: list index out of range with data hosted on Zenodo
This error is also raised when data is hosted on Google Drive: - https://huggingface.co/datasets/docred/discussions/5 - https://huggingface.co/datasets/linnaeus/discussions/3 - https://huggingface.co/datasets/poleval2019_mt/discussions/3 - https://huggingface.co/datasets/reddit_tifu/discussions/2 - https://hugging...
The dataset viewer sometimes raises an `IndexError`: ``` IndexError: list index out of range ``` See: - huggingface/datasets-server#1151 - https://huggingface.co/datasets/reddit/discussions/5 - huggingface/datasets-server#1118 - https://huggingface.co/datasets/krr-oxford/OntoLAMA/discussions/1 - https://hu...
140
26
IndexError: list index out of range with data hosted on Zenodo The dataset viewer sometimes raises an `IndexError`: ``` IndexError: list index out of range ``` See: - huggingface/datasets-server#1151 - https://huggingface.co/datasets/reddit/discussions/5 - huggingface/datasets-server#1118 - https://huggin...
[ -1.005910038948059, -0.9337819218635559, -0.8602526187896729, 1.411230444908142, 0.06302760541439056, -1.333472728729248, 0.07111027836799622, -0.8140448331832886, 1.5044735670089722, -0.5217399597167969, 0.25575631856918335, -1.7397807836532593, -0.236036017537117, -0.5876840353012085, ...
https://github.com/huggingface/datasets/issues/5858
Throw an error when dataset improperly indexed
Thanks for reporting, @sarahwie. Please note that in `datasets` we do not have vectorized operation like `pandas`. Therefore, your equality comparisons above are `False`: - For example: `squad['question']` returns a `list`, and this list is not equal to `"Who was the Norse leader?"` The `False` value is equivale...
### Describe the bug Pandas-style subset indexing on dataset does not throw an error, when maybe it should. Instead returns the first instance of the dataset regardless of index condition. ### Steps to reproduce the bug Steps to reproduce the behavior: 1. `squad = datasets.load_dataset("squad_v2", split="validati...
141
99
Throw an error when dataset improperly indexed ### Describe the bug Pandas-style subset indexing on dataset does not throw an error, when maybe it should. Instead returns the first instance of the dataset regardless of index condition. ### Steps to reproduce the bug Steps to reproduce the behavior: 1. `squad = ...
[ -1.228513240814209, -0.9444810748100281, -0.7796606421470642, 1.574063777923584, -0.16571839153766632, -1.1630815267562866, 0.08318566530942917, -1.0527688264846802, 1.683459997177124, -0.7457794547080994, 0.2045617699623108, -1.6489673852920532, 0.0285106860101223, -0.6358548402786255, ...
https://github.com/huggingface/datasets/issues/5857
Adding chemistry dataset/models in huggingface
Hi! This would be a nice addition to the Hub! You can find the existing chemistry datasets/models on the Hub (using the `chemistry` tag) [here](https://huggingface.co/search/full-text?q=chemistry&type=model&type=dataset). Feel free to ping us here on the Hub if you need help adding the datasets.
### Feature request Huggingface is really amazing platform for open science. In addition to computer vision, video and NLP, would it be of interest to add chemistry/materials science dataset/models in Huggingface? Or, if its already done, can you provide some pointers. We have been working on a comprehensive ben...
142
41
Adding chemistry dataset/models in huggingface ### Feature request Huggingface is really amazing platform for open science. In addition to computer vision, video and NLP, would it be of interest to add chemistry/materials science dataset/models in Huggingface? Or, if its already done, can you provide some pointer...
[ -1.1607859134674072, -0.9598888158798218, -0.8385828733444214, 1.5231138467788696, -0.053224142640829086, -1.287381649017334, 0.050442833453416824, -0.9187561869621277, 1.6461793184280396, -0.752089262008667, 0.3613884150981903, -1.7062150239944458, 0.02108807861804962, -0.6100136637687683...
https://github.com/huggingface/datasets/issues/5856
Error loading natural_questions
Hi! You can avoid this error by using the preprocessed version: ```python import datasets ds = datasets.load_dataset('natural_questions') ``` PS: Once we finish https://github.com/huggingface/datasets/pull/5364, this error will no longer be a problem.
### Describe the bug When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9: ```python import datasets datasets.load_dataset('natural_questions',beam_runner='DirectRunner') ``` It failed with following info: `pyarrow.lib.ArrowNotImplementedError: Nested data conversions not impl...
143
31
Error loading natural_questions ### Describe the bug When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9: ```python import datasets datasets.load_dataset('natural_questions',beam_runner='DirectRunner') ``` It failed with following info: `pyarrow.lib.ArrowNotImplementedError...
[ -1.1990299224853516, -0.9553800225257874, -0.6237490177154541, 1.434041976928711, -0.17947255074977875, -1.2862924337387085, 0.11571666598320007, -1.0527979135513306, 1.5706970691680908, -0.7776874303817749, 0.20714278519153595, -1.6421008110046387, -0.09185336530208588, -0.478563755750656...
https://github.com/huggingface/datasets/issues/5856
Error loading natural_questions
> Hi! You can avoid this error by using the preprocessed version: > > ```python > import datasets > ds = datasets.load_dataset('natural_questions') > ``` > > PS: Once we finish #5364, this error will no longer be a problem. Thanks, wish #5364 finish early
### Describe the bug When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9: ```python import datasets datasets.load_dataset('natural_questions',beam_runner='DirectRunner') ``` It failed with following info: `pyarrow.lib.ArrowNotImplementedError: Nested data conversions not impl...
143
44
Error loading natural_questions ### Describe the bug When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9: ```python import datasets datasets.load_dataset('natural_questions',beam_runner='DirectRunner') ``` It failed with following info: `pyarrow.lib.ArrowNotImplementedError...
[ -1.1990299224853516, -0.9553800225257874, -0.6237490177154541, 1.434041976928711, -0.17947255074977875, -1.2862924337387085, 0.11571666598320007, -1.0527979135513306, 1.5706970691680908, -0.7776874303817749, 0.20714278519153595, -1.6421008110046387, -0.09185336530208588, -0.478563755750656...
https://github.com/huggingface/datasets/issues/5855
`to_tf_dataset` consumes too much memory
Cc @amyeroberts @Rocketknight1 Indded I think it's because it does something like this under the hood when there's no multiprocessing: ```python tf_dataset = tf_dataset.shuffle(len(dataset)) ``` PS: with multiprocessing it appears to be different: ```python indices = np.arange(len(dataset)) if shuffle:...
### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github....
144
41
`to_tf_dataset` consumes too much memory ### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle ...
[ -1.21280837059021, -0.9416200518608093, -0.6731261014938354, 1.4203523397445679, -0.16164042055606842, -1.3120754957199097, 0.2671286165714264, -1.0843902826309204, 1.663701057434082, -0.8069621324539185, 0.256039023399353, -1.6323555707931519, 0.015697987750172615, -0.6138535141944885, ...
https://github.com/huggingface/datasets/issues/5855
`to_tf_dataset` consumes too much memory
Hi @massquantity, the dataset being shuffled there is not the full dataset. If you look at [the line above](https://github.com/huggingface/datasets/blob/main/src/datasets/utils/tf_utils.py#L182), the dataset is actually just a single indices array at that point, and that array is the only thing that gets fully loaded i...
### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github....
144
134
`to_tf_dataset` consumes too much memory ### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle ...
[ -1.21280837059021, -0.9416200518608093, -0.6731261014938354, 1.4203523397445679, -0.16164042055606842, -1.3120754957199097, 0.2671286165714264, -1.0843902826309204, 1.663701057434082, -0.8069621324539185, 0.256039023399353, -1.6323555707931519, 0.015697987750172615, -0.6138535141944885, ...
https://github.com/huggingface/datasets/issues/5855
`to_tf_dataset` consumes too much memory
Hi @Rocketknight1, you're right and I also noticed that only indices are used in shuffling. My data has shape (50000000, 10), but really the problem doesn't relate to a specific dataset. Simply running the following code costs me 10GB of memory. ```python from datasets import Dataset def gen(): for i in ran...
### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github....
144
160
`to_tf_dataset` consumes too much memory ### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle ...
[ -1.21280837059021, -0.9416200518608093, -0.6731261014938354, 1.4203523397445679, -0.16164042055606842, -1.3120754957199097, 0.2671286165714264, -1.0843902826309204, 1.663701057434082, -0.8069621324539185, 0.256039023399353, -1.6323555707931519, 0.015697987750172615, -0.6138535141944885, ...
https://github.com/huggingface/datasets/issues/5855
`to_tf_dataset` consumes too much memory
Thanks for that reproduction script - I've confirmed the same issue is occurring for me. Investigating it now!
### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github....
144
18
`to_tf_dataset` consumes too much memory ### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle ...
[ -1.21280837059021, -0.9416200518608093, -0.6731261014938354, 1.4203523397445679, -0.16164042055606842, -1.3120754957199097, 0.2671286165714264, -1.0843902826309204, 1.663701057434082, -0.8069621324539185, 0.256039023399353, -1.6323555707931519, 0.015697987750172615, -0.6138535141944885, ...
https://github.com/huggingface/datasets/issues/5855
`to_tf_dataset` consumes too much memory
Update: The memory usage is occurring in creation of the index and shuffle buffer. You can reproduce it very simply with: ```python import tensorflow as tf indices = tf.range(50_000_000, dtype=tf.int64) dataset = tf.data.Dataset.from_tensor_slices(indices) dataset = dataset.shuffle(len(dataset)) print(next(iter...
### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github....
144
99
`to_tf_dataset` consumes too much memory ### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle ...
[ -1.21280837059021, -0.9416200518608093, -0.6731261014938354, 1.4203523397445679, -0.16164042055606842, -1.3120754957199097, 0.2671286165714264, -1.0843902826309204, 1.663701057434082, -0.8069621324539185, 0.256039023399353, -1.6323555707931519, 0.015697987750172615, -0.6138535141944885, ...
https://github.com/huggingface/datasets/issues/5854
Can not load audiofolder dataset on kaggle
Hi! `audiofolder` requires `datasets>=2.5.0`, so please update the `datasets`' installation (`pip install -U datasets`) in the environment to resolve the issue.
### Describe the bug It's crash log: FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingfac...
145
21
Can not load audiofolder dataset on kaggle ### Describe the bug It's crash log: FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at...
[ -1.1507515907287598, -0.9056360721588135, -0.711188793182373, 1.461253046989441, -0.15764841437339783, -1.2411704063415527, 0.06300104409456253, -0.9573824405670166, 1.6098647117614746, -0.7448211908340454, 0.24856036901474, -1.7515056133270264, -0.0716467946767807, -0.551703155040741, -...
https://github.com/huggingface/datasets/issues/5854
Can not load audiofolder dataset on kaggle
> Hi! `audiofolder` requires `datasets>=2.5.0`, so please update the `datasets`' installation (`pip install -U datasets`) in the environment to resolve the issue. I don't think it is a problem of the version. It runs ok on colab or local machine. Only on kaggle will has this bug.
### Describe the bug It's crash log: FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingfac...
145
47
Can not load audiofolder dataset on kaggle ### Describe the bug It's crash log: FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at...
[ -1.154200792312622, -0.9217584133148193, -0.7177584767341614, 1.4420685768127441, -0.15082411468029022, -1.2470622062683105, 0.0723164901137352, -0.9657632112503052, 1.6368587017059326, -0.7386761903762817, 0.2630559504032135, -1.738788366317749, -0.06320306658744812, -0.5161727666854858, ...
https://github.com/huggingface/datasets/issues/5854
Can not load audiofolder dataset on kaggle
Based on your dataset info, the installed version is `2.1.0`, which does not include `audiofolder`. By default, Kaggle preinstalls `datasets` into a new env, but the version it installs is outdated and does not contain newer features such as `audiofolder`
### Describe the bug It's crash log: FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingfac...
145
40
Can not load audiofolder dataset on kaggle ### Describe the bug It's crash log: FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at...
[ -1.1274595260620117, -0.9151424765586853, -0.7192152738571167, 1.441693902015686, -0.15230119228363037, -1.22844660282135, 0.08878697454929352, -0.9621649384498596, 1.590338110923767, -0.7543883323669434, 0.26574140787124634, -1.7408254146575928, -0.06892996281385422, -0.539057731628418, ...
https://github.com/huggingface/datasets/issues/5847
Streaming IterableDataset not working with translation pipeline
[`KeyDataset`](https://github.com/huggingface/transformers/blob/7f8b909189547944617741d8d3c6c84504701693/src/transformers/pipelines/pt_utils.py#L296) doesn't support iterable datasets, so you either need to implement a version that does (and also indexing nested (translation) fields): ```python from torch.utils.dat...
### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that IterableDataset has no len(). ### Steps to reproduce the bug CODE: ``` ...
146
128
Streaming IterableDataset not working with translation pipeline ### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that Iterable...
[ -1.2637747526168823, -0.9374301433563232, -0.7347829937934875, 1.4584630727767944, -0.18755613267421722, -1.1666781902313232, 0.17908449470996857, -1.0838751792907715, 1.640811562538147, -0.759229302406311, 0.2735395133495331, -1.6184109449386597, 0.04956751689314842, -0.5483786463737488, ...
https://github.com/huggingface/datasets/issues/5847
Streaming IterableDataset not working with translation pipeline
@mariosasko The map snippet didn't quite work, but gave me enough of a clue to get it working. The following snippet does work: ``` def en_translation(x): return {"en":x['translation']['en']} ks = ds.map(en_translation, remove_columns=['translation']) test=[] for x in iter(ks): test.append(x['en']) xx=...
### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that IterableDataset has no len(). ### Steps to reproduce the bug CODE: ``` ...
146
219
Streaming IterableDataset not working with translation pipeline ### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that Iterable...
[ -1.2637747526168823, -0.9374301433563232, -0.7347829937934875, 1.4584630727767944, -0.18755613267421722, -1.1666781902313232, 0.17908449470996857, -1.0838751792907715, 1.640811562538147, -0.759229302406311, 0.2735395133495331, -1.6184109449386597, 0.04956751689314842, -0.5483786463737488, ...
https://github.com/huggingface/datasets/issues/5847
Streaming IterableDataset not working with translation pipeline
So perhaps there's no bug exactly, but I would love to see two things: 1) improve the documentation to better understand what's really getting returned. 2) update the example provided of using transformer pipeline with a dataset to include the oddball case that translation appears to be.
### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that IterableDataset has no len(). ### Steps to reproduce the bug CODE: ``` ...
146
47
Streaming IterableDataset not working with translation pipeline ### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that Iterable...
[ -1.2637747526168823, -0.9374301433563232, -0.7347829937934875, 1.4584630727767944, -0.18755613267421722, -1.1666781902313232, 0.17908449470996857, -1.0838751792907715, 1.640811562538147, -0.759229302406311, 0.2735395133495331, -1.6184109449386597, 0.04956751689314842, -0.5483786463737488, ...
https://github.com/huggingface/datasets/issues/5847
Streaming IterableDataset not working with translation pipeline
Hi, for the original snippet, the issue is that `streaming` datasets are not countable (they have no len) and therefore `KeyDataset` cannot work with them ( KeyDataset is a dataset and therefore requires a length). I modified slightly the original snippet to make it work: ```python from transformers import pi...
### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that IterableDataset has no len(). ### Steps to reproduce the bug CODE: ``` ...
146
166
Streaming IterableDataset not working with translation pipeline ### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that Iterable...
[ -1.2637747526168823, -0.9374301433563232, -0.7347829937934875, 1.4584630727767944, -0.18755613267421722, -1.1666781902313232, 0.17908449470996857, -1.0838751792907715, 1.640811562538147, -0.759229302406311, 0.2735395133495331, -1.6184109449386597, 0.04956751689314842, -0.5483786463737488, ...
https://github.com/huggingface/datasets/issues/5847
Streaming IterableDataset not working with translation pipeline
Maybe moving `KeyDataset` later on in the guide and specify it's mostly for streaming then? Or is it also necessary for batch_size>1 (which is what the current doc implies)?
### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that IterableDataset has no len(). ### Steps to reproduce the bug CODE: ``` ...
146
29
Streaming IterableDataset not working with translation pipeline ### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that Iterable...
[ -1.2637747526168823, -0.9374301433563232, -0.7347829937934875, 1.4584630727767944, -0.18755613267421722, -1.1666781902313232, 0.17908449470996857, -1.0838751792907715, 1.640811562538147, -0.759229302406311, 0.2735395133495331, -1.6184109449386597, 0.04956751689314842, -0.5483786463737488, ...
https://github.com/huggingface/datasets/issues/5847
Streaming IterableDataset not working with translation pipeline
Hmm Iterator (`yield`) : - Not countable - Super flexible - Cannot use `num_workers>1` (threading requires indexing at the correct location, iterators require to iterate in order,so each thread would iterate over the full thing being genuinely a bad idea) - Can batch - tqdm doesn't show a nice progress bar (it ...
### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that IterableDataset has no len(). ### Steps to reproduce the bug CODE: ``` ...
146
188
Streaming IterableDataset not working with translation pipeline ### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that Iterable...
[ -1.2637747526168823, -0.9374301433563232, -0.7347829937934875, 1.4584630727767944, -0.18755613267421722, -1.1666781902313232, 0.17908449470996857, -1.0838751792907715, 1.640811562538147, -0.759229302406311, 0.2735395133495331, -1.6184109449386597, 0.04956751689314842, -0.5483786463737488, ...
https://github.com/huggingface/datasets/issues/5846
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
This is due to the slow resolution of the data files: https://github.com/huggingface/datasets/issues/5537. We plan to switch to `huggingface_hub`'s `HfFileSystem` soon to make the resolution faster (will be up to 20x faster once we merge https://github.com/huggingface/huggingface_hub/pull/1443)
### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds. ### Environment info - `datasets` version: 2.1...
147
35
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow! ### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one o...
[ -1.126883864402771, -0.9017356038093567, -0.7633296847343445, 1.486032247543335, -0.11794661730527878, -1.3176013231277466, 0.11445632576942444, -0.9202184677124023, 1.6097322702407837, -0.7836844325065613, 0.353589802980423, -1.73179030418396, -0.009220968931913376, -0.6028525829315186, ...
https://github.com/huggingface/datasets/issues/5846
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
You're right, when I try to parse more than 50GB of text data, I also get very slow, usually taking hours or even tens of hours.
### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds. ### Environment info - `datasets` version: 2.1...
147
26
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow! ### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one o...
[ -1.2048887014389038, -0.8930887579917908, -0.7761214375495911, 1.4448931217193604, -0.17262375354766846, -1.313103437423706, 0.11856652051210403, -0.9846729636192322, 1.634273648262024, -0.8477228879928589, 0.3353094756603241, -1.7674273252487183, 0.044337861239910126, -0.5383802652359009,...
https://github.com/huggingface/datasets/issues/5846
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
> You're right, when I try to parse more than 50GB of text data, I also get very slow, usually taking hours or even tens of hours. That's unrelated to the problem discussed in this issue.
### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds. ### Environment info - `datasets` version: 2.1...
147
36
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow! ### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one o...
[ -1.2141969203948975, -0.9122954607009888, -0.7814145684242249, 1.449281930923462, -0.1722673773765564, -1.3275378942489624, 0.10293480008840561, -0.9914499521255493, 1.6627647876739502, -0.8441689014434814, 0.31415659189224243, -1.7814815044403076, 0.06872977316379547, -0.5377781987190247,...
https://github.com/huggingface/datasets/issues/5846
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
> > You're right, when I try to parse more than 50GB of text data, I also get very slow, usually taking hours or even tens of hours. > > That's unrelated to the problem discussed in this issue. Sorry, I misunderstood it.
### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds. ### Environment info - `datasets` version: 2.1...
147
43
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow! ### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one o...
[ -1.2186570167541504, -0.894500195980072, -0.7827442288398743, 1.4336414337158203, -0.18983258306980133, -1.3164923191070557, 0.08473926037549973, -1.0351999998092651, 1.6580960750579834, -0.8356509804725647, 0.30576083064079285, -1.7782236337661743, 0.0691995918750763, -0.5293062329292297,...
https://github.com/huggingface/datasets/issues/5841
Abusurdly slow on iteration
Hi ! You can try to use the [Image](https://huggingface.co/docs/datasets/v2.12.0/en/package_reference/main_classes#datasets.Image) type which [decodes images on-the-fly](https://huggingface.co/docs/datasets/v2.12.0/en/about_dataset_features#image-feature) into pytorch tensors :) ```python ds = Dataset.from_dict({"t...
### Describe the bug I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment: ```python a=torch.randn(100,224) a=torch.stack([a] * 10000) a.shape # %% ds=Dataset.from_d...
148
226
Abusurdly slow on iteration ### Describe the bug I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment: ```python a=torch.randn(100,224) a=torch.stack([a] * 10000) a.sh...
[ -1.2770248651504517, -0.8749205470085144, -0.7151722311973572, 1.4440370798110962, -0.22366788983345032, -1.2704261541366577, 0.19402125477790833, -1.0399311780929565, 1.7257866859436035, -0.8413228988647461, 0.34074726700782776, -1.6853406429290771, 0.12786391377449036, -0.534410238265991...
https://github.com/huggingface/datasets/issues/5841
Abusurdly slow on iteration
Anyway, regarding the speed difference between numpy and pytorch, I think the issue is that we first convert numpy sub-arrays to pytorch and then consolidate into one tensor, while we should to the opposite. Indeed converting a numpy array to pytorch has a fix cost that seems to cause a slow down. The current pipeline ...
### Describe the bug I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment: ```python a=torch.randn(100,224) a=torch.stack([a] * 10000) a.shape # %% ds=Dataset.from_d...
148
91
Abusurdly slow on iteration ### Describe the bug I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment: ```python a=torch.randn(100,224) a=torch.stack([a] * 10000) a.sh...
[ -1.2770248651504517, -0.8749205470085144, -0.7151722311973572, 1.4440370798110962, -0.22366788983345032, -1.2704261541366577, 0.19402125477790833, -1.0399311780929565, 1.7257866859436035, -0.8413228988647461, 0.34074726700782776, -1.6853406429290771, 0.12786391377449036, -0.534410238265991...
https://github.com/huggingface/datasets/issues/5841
Abusurdly slow on iteration
I have a similar issue: iterating over a dataset takes 5s without applying any transform, but takes ~30s after applying a transform. Here is the minimum code to reproduce the problem ```python import numpy as np from datasets import Dataset, DatasetDict, load_dataset, Array3D, Image, Features from torch.utils.da...
### Describe the bug I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment: ```python a=torch.randn(100,224) a=torch.stack([a] * 10000) a.shape # %% ds=Dataset.from_d...
148
187
Abusurdly slow on iteration ### Describe the bug I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment: ```python a=torch.randn(100,224) a=torch.stack([a] * 10000) a.sh...
[ -1.2770248651504517, -0.8749205470085144, -0.7151722311973572, 1.4440370798110962, -0.22366788983345032, -1.2704261541366577, 0.19402125477790833, -1.0399311780929565, 1.7257866859436035, -0.8413228988647461, 0.34074726700782776, -1.6853406429290771, 0.12786391377449036, -0.534410238265991...
https://github.com/huggingface/datasets/issues/5842
Remove columns in interable dataset
This method has been recently added to the `IterableDataset`, so you need to update the `datasets`' installation (`pip install -U datasets`) to use it.
### Feature request Right now, remove_columns() produces a NotImplementedError for iterable style datasets ### Motivation It would be great to have the same functionality irrespective of whether one is using an iterable or a map-style dataset ### Your contribution hope and courage.
150
24
Remove columns in interable dataset ### Feature request Right now, remove_columns() produces a NotImplementedError for iterable style datasets ### Motivation It would be great to have the same functionality irrespective of whether one is using an iterable or a map-style dataset ### Your contribution hope and cou...
[ -1.188256859779358, -0.9099082350730896, -0.8461495041847229, 1.46682870388031, -0.10237018764019012, -1.3405834436416626, 0.14300890266895294, -1.020204782485962, 1.7700542211532593, -0.7978855967521667, 0.3247756063938141, -1.661026120185852, 0.018075844272971153, -0.6116166114807129, ...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
As the name says, `load_from_disk` load the data from your disk. If the data is hosted on S3, it is first downloaded locally and then loaded from your disk. There is a discussion on streaming data from S3 here though: #5281
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
41
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.1314722299575806, -1.0132465362548828, -0.8381555080413818, 1.601589322090149, -0.22572031617164612, -1.320214033126831, 0.25365862250328064, -1.107954740524292, 1.7764413356781006, -1.0004515647888184, 0.37760037183761597, -1.665255069732666, 0.09470492601394653, -0.7143040299415588, ...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
@lhoestq Thanks for your comment. I have checked out the discussion before and attempted at replicating the mentioned changes in the main branch (#5580). What I found was that if a dataset is saved using `save_to_disk`, it cannot be read by `load_dataset`. The error message asks me to to use `load_from_disk` instead....
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
65
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.108872890472412, -0.9334377646446228, -0.8395171165466309, 1.5438768863677979, -0.20451661944389343, -1.3188121318817139, 0.22357897460460663, -1.1496220827102661, 1.727474570274353, -0.9264642000198364, 0.34790247678756714, -1.6536660194396973, 0.14242373406887054, -0.6958389282226562,...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
Using `push_to_hub` you can save the dataset on the HF Hub as parquet files, and reload it / stream it using `load_dataset` :) If you want to save your dataset somewhere else you can use `.to_parquet` to get a parquet file. If your dataset is big it's usually recommended to shard it into multi parquet files (around ...
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
59
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.0931892395019531, -0.9550338387489319, -0.8697564601898193, 1.601189374923706, -0.20103377103805542, -1.3057785034179688, 0.1850281059741974, -1.0832709074020386, 1.755860447883606, -0.9503853917121887, 0.4491075575351715, -1.6701313257217407, 0.15975995361804962, -0.6955724954605103, ...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
@lhoestq I tried the method you mentioned. This the current scenario I'm facing: - The parquet file can be read from disk and streaming can be enabled. - The parquet file can be read from `s3` (local MinIO). - When `streaming=True` is enabled for `s3`, I get the error mentioned below: ``` File ~/.../lib/pytho...
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
106
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.18739652633667, -0.9611495137214661, -0.7330560088157654, 1.516315221786499, -0.20272262394428253, -1.3028978109359741, 0.24334430694580078, -1.1313461065292358, 1.6697351932525635, -0.8155365586280823, 0.3690471351146698, -1.6336036920547485, 0.02870561182498932, -0.7040586471557617, ...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
Streaming from S3 is still experimental, there might be a few bugs unfortunately. Can you share the full stack trace ?
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
21
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.1281518936157227, -0.956210732460022, -0.8254011273384094, 1.5410455465316772, -0.2120295614004135, -1.3372948169708252, 0.2579007148742676, -1.1193286180496216, 1.7654820680618286, -0.9240826368331909, 0.34024980664253235, -1.6349722146987915, 0.11261370778083801, -0.6820954084396362, ...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
@lhoestq Sure, here you go: ```python TypeError Traceback (most recent call last) Cell In[8], line 1 ----> 1 dataset = load_dataset("parquet", data_files=["s3://<bucket name>/<data folder>/data-parquet"], storage_options=fs.storage_options, streaming=True) File ~/.../datasets/...
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
1,037
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.2041233777999878, -1.0062037706375122, -0.6313180327415466, 1.437657117843628, -0.18664082884788513, -1.2890902757644653, 0.2077275812625885, -1.1264944076538086, 1.6255488395690918, -0.8505526781082153, 0.2045203596353531, -1.7202439308166504, -0.027755051851272583, -0.6888807415962219...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
Is `"data-parquet"` a file ? In `data_files` you should pass the paths to the parquet files (not to a directory). Glob patterns are not supported yet for S3 URLs. The bug seems to happen because your provided data file has no extension. Because of that it tries to infer it from the file content, but fails because `_...
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
63
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.1007708311080933, -0.9273959994316101, -0.8560488224029541, 1.587786316871643, -0.18420720100402832, -1.3287254571914673, 0.19336439669132233, -1.1795574426651, 1.8246614933013916, -0.943767249584198, 0.3852311670780182, -1.6260333061218262, 0.16734789311885834, -0.7044389247894287, -...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
@lhoestq Thank you for your answer. Saving the file with `.parquet` extension solved the issue! This is really great! Really appreciate all the help! Let me know if I should close the issue or feel free to close it if you want.
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
42
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.1393764019012451, -0.9299445152282715, -0.8536918759346008, 1.56565523147583, -0.20641542971134186, -1.3232091665267944, 0.1775520145893097, -1.209576964378357, 1.7437632083892822, -0.9298161864280701, 0.3524990677833557, -1.5843348503112793, 0.16504746675491333, -0.6726346015930176, ...
https://github.com/huggingface/datasets/issues/5838
Streaming support for `load_from_disk`
Cool ! I'm glad it worked out :) Sure feel free to close the issue, since the original question about streaming with load_from_disk has been answered anyway
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
152
27
Streaming support for `load_from_disk` ### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very...
[ -1.1667375564575195, -0.9901631474494934, -0.866985559463501, 1.560174584388733, -0.20836511254310608, -1.3153737783432007, 0.2408045083284378, -1.1217237710952759, 1.729688286781311, -0.9517273902893066, 0.30514436960220337, -1.655367136001587, 0.15841801464557648, -0.6709691286087036, ...
https://github.com/huggingface/datasets/issues/5837
Use DeepSpeed load myself " .csv " dataset.
Hi ! Doing `load_dataset("path/to/data.csv")` is not supported yet, but you can do ```python ds = load_dataset("csv", data_files=["path/to/data.csv"]) ```
### Describe the bug When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", line 1767, in load_dataset builder_instan...
153
18
Use DeepSpeed load myself " .csv " dataset. ### Describe the bug When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", ...
[ -1.1078572273254395, -0.9771057963371277, -0.6886123418807983, 1.417290449142456, -0.07183212786912918, -1.2041537761688232, 0.13848580420017242, -1.0561885833740234, 1.5372542142868042, -0.702926516532898, 0.16869564354419708, -1.7251721620559692, -0.2262987643480301, -0.4913672208786011,...
https://github.com/huggingface/datasets/issues/5837
Use DeepSpeed load myself " .csv " dataset.
The other question: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", line 1767, in load_dataset builder_instance = load_dataset_builder( File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/da...
### Describe the bug When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", line 1767, in load_dataset builder_instan...
153
229
Use DeepSpeed load myself " .csv " dataset. ### Describe the bug When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", ...
[ -1.1713345050811768, -0.9852048754692078, -0.6814395189285278, 1.3524361848831177, -0.07349710166454315, -1.2071709632873535, 0.13374069333076477, -1.0489364862442017, 1.5145426988601685, -0.6892326474189758, 0.1229124665260315, -1.6937755346298218, -0.23447763919830322, -0.462883621454238...
https://github.com/huggingface/datasets/issues/5834
Is uint8 supported?
Hi ! The numpy formatting detaults to int64 and float32 - but you can use uint8 using ```python ds = ds.with_format("numpy", dtype=np.uint8) ```
### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well. Is there a way ...
154
23
Is uint8 supported? ### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as ...
[ -1.2034748792648315, -0.887102484703064, -0.7158054113388062, 1.5004172325134277, -0.15594178438186646, -1.2199872732162476, 0.1520499736070633, -1.0931395292282104, 1.7031718492507935, -0.7893821597099304, 0.375015527009964, -1.6874333620071411, -0.007075731176882982, -0.5761370062828064,...
https://github.com/huggingface/datasets/issues/5834
Is uint8 supported?
Thank you! By setting `ds.with_format("numpy", dtype=np.uint8)`, the dataset returns the data in `uint8`. However, `with_format` and `set_format` seem to cast the data on-the-fly. I want to reduce the dataset size by using `uint8` instead of `int64` and I observe no difference between using `int64` and `uint8` for...
### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well. Is there a way ...
154
74
Is uint8 supported? ### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as ...
[ -1.213528037071228, -0.8775548338890076, -0.7121641635894775, 1.4739090204238892, -0.16309478878974915, -1.2679214477539062, 0.1714174598455429, -1.1260467767715454, 1.7571793794631958, -0.8333533406257629, 0.3388484716415405, -1.6772592067718506, 0.005824077874422073, -0.6078845262527466,...
https://github.com/huggingface/datasets/issues/5834
Is uint8 supported?
If the feature type is `Value("uint8")` then it's written an uint8 on disk using the uint8 Arrow dtype. e.g. ```python ds = Dataset.from_dict({"a": range(10)}, features=Features({"a": Value("uint8")})) ds.data.nbytes # 10 ```
### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well. Is there a way ...
154
30
Is uint8 supported? ### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as ...
[ -1.1952255964279175, -0.9149463176727295, -0.7227889895439148, 1.5222022533416748, -0.17281107604503632, -1.211870551109314, 0.14803460240364075, -1.0799599885940552, 1.7145726680755615, -0.7580557465553284, 0.39515259861946106, -1.6825087070465088, -0.00803280621767044, -0.594983041286468...
https://github.com/huggingface/datasets/issues/5834
Is uint8 supported?
Oh, I understand now. The data was stored in `uint8` from the beginning (when the dataset returns `int64`). Thank you for your time! My question is fully resolved.
### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well. Is there a way ...
154
28
Is uint8 supported? ### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as ...
[ -1.2012821435928345, -0.8838594555854797, -0.7114223837852478, 1.4705657958984375, -0.13130874931812286, -1.212508201599121, 0.12983840703964233, -1.0950819253921509, 1.7106516361236572, -0.7925498485565186, 0.3421580195426941, -1.699402928352356, -0.021122673526406288, -0.5970733165740967...
https://github.com/huggingface/datasets/issues/5833
Unable to push dataset - `create_pr` problem
Thanks for reporting, @agombert. In this case, I think the root issue is authentication: before pushing to Hub, you should authenticate. See our docs: https://huggingface.co/docs/datasets/upload_dataset#upload-with-python > 2. To upload a dataset on the Hub in Python, you need to log in to your Hugging Face account...
### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ```python dataset.push_to_hub("agomberto/Fr...
155
50
Unable to push dataset - `create_pr` problem ### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ...
[ -1.1540454626083374, -0.7597602605819702, -0.6611561179161072, 1.4835621118545532, -0.0010568797588348389, -1.429667592048645, 0.05472240969538689, -0.9580106735229492, 1.5048177242279053, -0.7912386655807495, 0.3948341906070709, -1.7450616359710693, 0.05644043907523155, -0.588965237140655...
https://github.com/huggingface/datasets/issues/5833
Unable to push dataset - `create_pr` problem
Hey @albertvillanova well I actually did :D <img width="1079" alt="Capture d’écran 2023-05-09 à 18 02 58" src="https://github.com/huggingface/datasets/assets/17645711/e091aa20-06b1-4dd3-bfdb-35e832c66f8d">
### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ```python dataset.push_to_hub("agomberto/Fr...
155
17
Unable to push dataset - `create_pr` problem ### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ...
[ -1.1540454626083374, -0.7597602605819702, -0.6611561179161072, 1.4835621118545532, -0.0010568797588348389, -1.429667592048645, 0.05472240969538689, -0.9580106735229492, 1.5048177242279053, -0.7912386655807495, 0.3948341906070709, -1.7450616359710693, 0.05644043907523155, -0.588965237140655...
https://github.com/huggingface/datasets/issues/5833
Unable to push dataset - `create_pr` problem
That is weird that you get a Forbidden error if you are properly authenticated... Today we had a big outage issue affecting the Hugging Face Hub. Could you please retry to push_to_hub your dataset? Maybe that was the cause...
### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ```python dataset.push_to_hub("agomberto/Fr...
155
39
Unable to push dataset - `create_pr` problem ### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ...
[ -1.1540454626083374, -0.7597602605819702, -0.6611561179161072, 1.4835621118545532, -0.0010568797588348389, -1.429667592048645, 0.05472240969538689, -0.9580106735229492, 1.5048177242279053, -0.7912386655807495, 0.3948341906070709, -1.7450616359710693, 0.05644043907523155, -0.588965237140655...
https://github.com/huggingface/datasets/issues/5833
Unable to push dataset - `create_pr` problem
Login successful but also got this error "Forbidden: pass `create_pr=1` as a query parameter to create a Pull Request"
### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ```python dataset.push_to_hub("agomberto/Fr...
155
19
Unable to push dataset - `create_pr` problem ### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ...
[ -1.1540454626083374, -0.7597602605819702, -0.6611561179161072, 1.4835621118545532, -0.0010568797588348389, -1.429667592048645, 0.05472240969538689, -0.9580106735229492, 1.5048177242279053, -0.7912386655807495, 0.3948341906070709, -1.7450616359710693, 0.05644043907523155, -0.588965237140655...
https://github.com/huggingface/datasets/issues/5833
Unable to push dataset - `create_pr` problem
Make sure your API token has a `write` role. I had the same issue as you with the `read` token. Creating a `write` token and using that solved the issue.
### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ```python dataset.push_to_hub("agomberto/Fr...
155
30
Unable to push dataset - `create_pr` problem ### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ...
[ -1.1540454626083374, -0.7597602605819702, -0.6611561179161072, 1.4835621118545532, -0.0010568797588348389, -1.429667592048645, 0.05472240969538689, -0.9580106735229492, 1.5048177242279053, -0.7912386655807495, 0.3948341906070709, -1.7450616359710693, 0.05644043907523155, -0.588965237140655...
https://github.com/huggingface/datasets/issues/5833
Unable to push dataset - `create_pr` problem
> Make sure your API token has a `write` role. I had the same issue as you with the `read` token. Creating a `write` token and using that solved the issue. I generate a token with write role. It works! thank you so much.
### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ```python dataset.push_to_hub("agomberto/Fr...
155
44
Unable to push dataset - `create_pr` problem ### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ...
[ -1.1540454626083374, -0.7597602605819702, -0.6611561179161072, 1.4835621118545532, -0.0010568797588348389, -1.429667592048645, 0.05472240969538689, -0.9580106735229492, 1.5048177242279053, -0.7912386655807495, 0.3948341906070709, -1.7450616359710693, 0.05644043907523155, -0.588965237140655...
https://github.com/huggingface/datasets/issues/5833
Unable to push dataset - `create_pr` problem
@dmitrijsk amazing thanks so much ! The error should be clearer when the token is read-only – I wasted a lot of time there..
### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ```python dataset.push_to_hub("agomberto/Fr...
155
24
Unable to push dataset - `create_pr` problem ### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ...
[ -1.1540454626083374, -0.7597602605819702, -0.6611561179161072, 1.4835621118545532, -0.0010568797588348389, -1.429667592048645, 0.05472240969538689, -0.9580106735229492, 1.5048177242279053, -0.7912386655807495, 0.3948341906070709, -1.7450616359710693, 0.05644043907523155, -0.588965237140655...
https://github.com/huggingface/datasets/issues/5831
[Bug]504 Server Error when loading dataset which was already cached
I am experiencing the same problem with the following environment: * `datasets` version: 2.11.0 * Platform: `Linux 5.19.0-41-generic x86_64 GNU/Linux` * Python version: `3.8.5` * Huggingface_hub version: 0.13.3 * PyArrow version: `11.0.0` * Pandas version: `1.5.3` Trying to get some diagnostics, I got the fo...
### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it again using the same machine, I got this error: ``` Traceback (most rece...
157
120
[Bug]504 Server Error when loading dataset which was already cached ### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it agai...
[ -1.1827178001403809, -0.8930608630180359, -0.7200658321380615, 1.3524177074432373, -0.07334086298942566, -1.2466628551483154, 0.04284411296248436, -1.0322849750518799, 1.5563263893127441, -0.6889245510101318, 0.17516101896762848, -1.7303858995437622, -0.2097388356924057, -0.497710257768631...
https://github.com/huggingface/datasets/issues/5831
[Bug]504 Server Error when loading dataset which was already cached
same problem with a private dataset repo, seems the huggingface hub server got some connection problem?
### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it again using the same machine, I got this error: ``` Traceback (most rece...
157
16
[Bug]504 Server Error when loading dataset which was already cached ### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it agai...
[ -1.1827178001403809, -0.8930608630180359, -0.7200658321380615, 1.3524177074432373, -0.07334086298942566, -1.2466628551483154, 0.04284411296248436, -1.0322849750518799, 1.5563263893127441, -0.6889245510101318, 0.17516101896762848, -1.7303858995437622, -0.2097388356924057, -0.497710257768631...
https://github.com/huggingface/datasets/issues/5831
[Bug]504 Server Error when loading dataset which was already cached
@SingL3 You can avoid this error by setting the [`HF_DATASETS_OFFLINE`](https://huggingface.co/docs/datasets/v2.12.0/en/loading#offline) env variable to 1. By default, if an internet connection is available, we check whether the cache of a cached dataset is up-to-date. @lucidBrot `datasets`' cache is still not alig...
### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it again using the same machine, I got this error: ``` Traceback (most rece...
157
48
[Bug]504 Server Error when loading dataset which was already cached ### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it agai...
[ -1.1827178001403809, -0.8930608630180359, -0.7200658321380615, 1.3524177074432373, -0.07334086298942566, -1.2466628551483154, 0.04284411296248436, -1.0322849750518799, 1.5563263893127441, -0.6889245510101318, 0.17516101896762848, -1.7303858995437622, -0.2097388356924057, -0.497710257768631...
https://github.com/huggingface/datasets/issues/5831
[Bug]504 Server Error when loading dataset which was already cached
Today we had a big issue affecting the Hugging Face Hub, thus all the `504 Server Error: Gateway Time-out` errors. It is fixed now and loading your datasets should work as expected.
### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it again using the same machine, I got this error: ``` Traceback (most rece...
157
32
[Bug]504 Server Error when loading dataset which was already cached ### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it agai...
[ -1.1827178001403809, -0.8930608630180359, -0.7200658321380615, 1.3524177074432373, -0.07334086298942566, -1.2466628551483154, 0.04284411296248436, -1.0322849750518799, 1.5563263893127441, -0.6889245510101318, 0.17516101896762848, -1.7303858995437622, -0.2097388356924057, -0.497710257768631...
https://github.com/huggingface/datasets/issues/5831
[Bug]504 Server Error when loading dataset which was already cached
Hi, @albertvillanova. If there is a locally cached version of datasets or something cache using huggingface_hub, when a network problem(either client or server) occurs, is it a better way to fallback to use the current cached version rather than raise a exception and exit?
### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it again using the same machine, I got this error: ``` Traceback (most rece...
157
44
[Bug]504 Server Error when loading dataset which was already cached ### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it agai...
[ -1.1827178001403809, -0.8930608630180359, -0.7200658321380615, 1.3524177074432373, -0.07334086298942566, -1.2466628551483154, 0.04284411296248436, -1.0322849750518799, 1.5563263893127441, -0.6889245510101318, 0.17516101896762848, -1.7303858995437622, -0.2097388356924057, -0.497710257768631...
https://github.com/huggingface/datasets/issues/5829
(mach-o file, but is an incompatible architecture (have 'arm64', need 'x86_64'))
That is weird. I can't reproduce it again after reboot. ```python In [2]: import platform In [3]: platform.platform() Out[3]: 'macOS-13.2-arm64-arm-64bit' In [4]: from datasets import load_dataset ...: ...: jazzy = load_dataset("nomic-ai/gpt4all-j-prompt-generations", revision='v1.2-jazzy') Found cach...
### Describe the bug M2 MBP can't run ```python from datasets import load_dataset jazzy = load_dataset("nomic-ai/gpt4all-j-prompt-generations", revision='v1.2-jazzy') ``` ### Steps to reproduce the bug 1. Use M2 MBP 2. Python 3.10.10 from pyenv 3. Run ``` from datasets import load_dataset jazzy = load_...
158
42
(mach-o file, but is an incompatible architecture (have 'arm64', need 'x86_64')) ### Describe the bug M2 MBP can't run ```python from datasets import load_dataset jazzy = load_dataset("nomic-ai/gpt4all-j-prompt-generations", revision='v1.2-jazzy') ``` ### Steps to reproduce the bug 1. Use M2 MBP 2. Python 3...
[ -1.2452958822250366, -0.8774450421333313, -0.686089277267456, 1.5009102821350098, -0.1302803009748459, -1.2884726524353027, 0.1107284277677536, -1.034258246421814, 1.5793386697769165, -0.8080062866210938, 0.3421730101108551, -1.63355553150177, -0.0004932684823870659, -0.599759578704834, ...
https://github.com/huggingface/datasets/issues/5828
Stream data concatenation issue
Hi! You can call `map` as follows to avoid the error: ```python augmented_dataset_cln = dataset_cln['train'].map(augment_dataset, features=dataset_cln['train'].features) ```
### Describe the bug I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset. ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string', id=None), 'audio': {'array': Sequence(feature=Value(dtype='float32', id=...
159
17
Stream data concatenation issue ### Describe the bug I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset. ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string', id=None), 'audio': {'array': Sequence(...
[ -1.3739922046661377, -1.1050032377243042, -0.708509087562561, 1.662628173828125, -0.27772530913352966, -0.9798835515975952, 0.11023248732089996, -1.0560520887374878, 1.5177412033081055, -0.753578245639801, 0.24591761827468872, -1.5724157094955444, -0.014820529147982597, -0.7283430099487305...
https://github.com/huggingface/datasets/issues/5828
Stream data concatenation issue
Hi! I have run into the same problem with you. Could you please let me know how you solve it? Thanks!
### Describe the bug I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset. ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string', id=None), 'audio': {'array': Sequence(feature=Value(dtype='float32', id=...
159
21
Stream data concatenation issue ### Describe the bug I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset. ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string', id=None), 'audio': {'array': Sequence(...
[ -1.3739922046661377, -1.1050032377243042, -0.708509087562561, 1.662628173828125, -0.27772530913352966, -0.9798835515975952, 0.11023248732089996, -1.0560520887374878, 1.5177412033081055, -0.753578245639801, 0.24591761827468872, -1.5724157094955444, -0.014820529147982597, -0.7283430099487305...
https://github.com/huggingface/datasets/issues/5827
load json dataset interrupt when dtype cast problem occured
Indeed the JSON dataset builder raises an error when it encounters an unexpected type. There's an old PR open to add away to ignore such elements though, if it can help: https://github.com/huggingface/datasets/pull/2838
### Describe the bug i have a json like this: [ {"id": 1, "name": 1}, {"id": 2, "name": "Nan"}, {"id": 3, "name": 3}, .... ] ,which have several problematic rows data like row 2, then i load it with datasets.load_dataset('json', data_files=['xx.json'], split='train'), it will report like this: ...
160
32
load json dataset interrupt when dtype cast problem occured ### Describe the bug i have a json like this: [ {"id": 1, "name": 1}, {"id": 2, "name": "Nan"}, {"id": 3, "name": 3}, .... ] ,which have several problematic rows data like row 2, then i load it with datasets.load_dataset('json', data...
[ -1.2560925483703613, -0.9595218896865845, -0.7310112714767456, 1.4875977039337158, -0.19069238007068634, -1.1312564611434937, 0.042799677699804306, -1.043371558189392, 1.718193531036377, -0.6977907419204712, 0.19077137112617493, -1.6873589754104614, 0.020541850477457047, -0.606010735034942...
https://github.com/huggingface/datasets/issues/5825
FileNotFound even though exists
Hi! This would only work if `bigscience/xP3` was a no-code dataset, but it isn't (it has a Python builder script). But this should work: ```python load_dataset("json", data_files="https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab....
### Describe the bug I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong? ``` Downloading builder script: 100% 2.82k/2.8...
161
28
FileNotFound even though exists ### Describe the bug I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong? ``` Downloadi...
[ -1.2246572971343994, -0.7424915432929993, -0.5879114270210266, 1.5263761281967163, 0.044311460107564926, -1.3627488613128662, 0.0702437236905098, -0.8897339105606079, 1.5228021144866943, -0.792717456817627, 0.4206331670284271, -1.6202031373977661, 0.06020117178559303, -0.5875002145767212, ...
https://github.com/huggingface/datasets/issues/5825
FileNotFound even though exists
> I see, it's not compatible w/ regex right? e.g. `load_dataset("json", data_files="https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/*")` It should work for patterns that "reference" the local filesystem, but to make this work with the Hub, we must implement https://github.com/huggingface/datasets/issu...
### Describe the bug I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong? ``` Downloading builder script: 100% 2.82k/2.8...
161
106
FileNotFound even though exists ### Describe the bug I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong? ``` Downloadi...
[ -1.2246572971343994, -0.7424915432929993, -0.5879114270210266, 1.5263761281967163, 0.044311460107564926, -1.3627488613128662, 0.0702437236905098, -0.8897339105606079, 1.5228021144866943, -0.792717456817627, 0.4206331670284271, -1.6202031373977661, 0.06020117178559303, -0.5875002145767212, ...
https://github.com/huggingface/datasets/issues/5820
Incomplete docstring for `BuilderConfig`
Thanks for reporting! You are more than welcome to improve `BuilderConfig`'s docstring. This class serves an identical purpose as `tensorflow_datasets`'s `BuilderConfig`, and its docstring is [here](https://github.com/tensorflow/datasets/blob/a95e38b5bb018312c3d3720619c2a8ef83ebf57f/tensorflow_datasets/core/dataset_...
Hi guys ! I stumbled upon this docstring while working on a project. Some of the attributes have missing descriptions. https://github.com/huggingface/datasets/blob/bc5fef5b6d91f009e4101684adcb374df2c170f6/src/datasets/builder.py#L104-L117
164
34
Incomplete docstring for `BuilderConfig` Hi guys ! I stumbled upon this docstring while working on a project. Some of the attributes have missing descriptions. https://github.com/huggingface/datasets/blob/bc5fef5b6d91f009e4101684adcb374df2c170f6/src/datasets/builder.py#L104-L117 Thanks for reporting! You are m...
[ -1.3261981010437012, -0.8913604617118835, -0.7317317128181458, 1.399688959121704, -0.09366767853498459, -1.2431700229644775, 0.09195101261138916, -1.0481846332550049, 1.6627223491668701, -0.6732012033462524, 0.27303406596183777, -1.6899288892745972, -0.025011997669935226, -0.53314298391342...
https://github.com/huggingface/datasets/issues/5819
Cannot pickle error in Dataset.from_generator()
Hi! It should work if you put `model = torch.compile(model)` inside the `generate_data` function. If a referenced object is outside, it needs to be pickable, and that's not the case for the compiled models (or functions).
### Describe the bug I'm trying to use Dataset.from_generator() to generate a large dataset. ### Steps to reproduce the bug Code to reproduce: ``` from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig import torch from tqdm import tqdm from datasets import load_dataset tokenizer...
165
36
Cannot pickle error in Dataset.from_generator() ### Describe the bug I'm trying to use Dataset.from_generator() to generate a large dataset. ### Steps to reproduce the bug Code to reproduce: ``` from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig import torch from tqdm import tq...
[ -1.312134861946106, -1.0129865407943726, -0.5742440819740295, 1.5412966012954712, -0.1450708657503128, -1.1299595832824707, 0.22083961963653564, -1.0590128898620605, 1.5314139127731323, -0.8891088962554932, 0.3064503073692322, -1.626840353012085, 0.027890335768461227, -0.5555327534675598, ...
https://github.com/huggingface/datasets/issues/5819
Cannot pickle error in Dataset.from_generator()
> Hi! It should work if you put `model = torch.compile(model)` inside the `generate_data` function. If a referenced object is outside, it needs to be pickable, and that's not the case for the compiled models (or functions). Hi! Thank you for your reply! Everything works perfectly with your suggestion! Closing the...
### Describe the bug I'm trying to use Dataset.from_generator() to generate a large dataset. ### Steps to reproduce the bug Code to reproduce: ``` from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig import torch from tqdm import tqdm from datasets import load_dataset tokenizer...
165
52
Cannot pickle error in Dataset.from_generator() ### Describe the bug I'm trying to use Dataset.from_generator() to generate a large dataset. ### Steps to reproduce the bug Code to reproduce: ``` from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig import torch from tqdm import tq...
[ -1.312134861946106, -1.0129865407943726, -0.5742440819740295, 1.5412966012954712, -0.1450708657503128, -1.1299595832824707, 0.22083961963653564, -1.0590128898620605, 1.5314139127731323, -0.8891088962554932, 0.3064503073692322, -1.626840353012085, 0.027890335768461227, -0.5555327534675598, ...
https://github.com/huggingface/datasets/issues/5818
Ability to update a dataset
In this case, I think we can avoid the `PermissionError` by unpacking the underlying `ConcatenationTable` and saving only the newly added data blocks (in new files).
### Feature request The ability to load a dataset, add or change something, and save it back to disk. Maybe it's possible, but I can't work out how to do it, e.g. this fails: ```py import datasets dataset = datasets.load_from_disk("data/test1") dataset = dataset.add_item({"text": "A new item"}) dataset.sav...
166
26
Ability to update a dataset ### Feature request The ability to load a dataset, add or change something, and save it back to disk. Maybe it's possible, but I can't work out how to do it, e.g. this fails: ```py import datasets dataset = datasets.load_from_disk("data/test1") dataset = dataset.add_item({"text...
[ -1.1833677291870117, -1.0026443004608154, -0.8135682940483093, 1.444919228553772, -0.15244489908218384, -1.2608505487442017, 0.15001557767391205, -1.05531644821167, 1.690877079963684, -0.8565082550048828, 0.32282742857933044, -1.6625043153762817, 0.006127838045358658, -0.6394720077514648, ...
https://github.com/huggingface/datasets/issues/5818
Ability to update a dataset
Thanks @stevhliu and @mariosasko , so saving to individual files then loading them later, concatenating again and saving again is the recommended way. Good to know. Question that I hope doesn't sound rude: is this sort of thing (processing a dataset that doesn't fit in memory) outside of `datasets`'s core area of fo...
### Feature request The ability to load a dataset, add or change something, and save it back to disk. Maybe it's possible, but I can't work out how to do it, e.g. this fails: ```py import datasets dataset = datasets.load_from_disk("data/test1") dataset = dataset.add_item({"text": "A new item"}) dataset.sav...
166
89
Ability to update a dataset ### Feature request The ability to load a dataset, add or change something, and save it back to disk. Maybe it's possible, but I can't work out how to do it, e.g. this fails: ```py import datasets dataset = datasets.load_from_disk("data/test1") dataset = dataset.add_item({"text...
[ -1.1918529272079468, -1.032199740409851, -0.7797137498855591, 1.3972488641738892, -0.16604754328727722, -1.2583508491516113, 0.12762945890426636, -1.0819259881973267, 1.7094367742538452, -0.8580541610717773, 0.27730515599250793, -1.664820909500122, -0.0029700882732868195, -0.64612269401550...
https://github.com/huggingface/datasets/issues/5817
Setting `num_proc` errors when `.map` returns additional items.
Hi ! Unfortunately I couldn't reproduce on my side locally and with datasets 2.11 and python 3.10.11 on colab. What version of `multiprocess` are you using ?
### Describe the bug I'm using a map function that returns more rows than are passed in. If I try to use `num_proc` I get: ``` File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kw...
167
27
Setting `num_proc` errors when `.map` returns additional items. ### Describe the bug I'm using a map function that returns more rows than are passed in. If I try to use `num_proc` I get: ``` File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper ...
[ -1.1727386713027954, -0.9139394760131836, -0.6932677626609802, 1.4026706218719482, -0.1811823844909668, -1.2461044788360596, 0.11010365188121796, -1.0843526124954224, 1.4787546396255493, -0.6602264046669006, 0.26320338249206543, -1.6500051021575928, -0.14681124687194824, -0.462516814470291...
https://github.com/huggingface/datasets/issues/5817
Setting `num_proc` errors when `.map` returns additional items.
I've got `multiprocess` version `0.70.14`. I've done some more testing and the error only occurs in PyCharm's Python Console. It seems to be [this PyCharm bug](https://youtrack.jetbrains.com/issue/PY-51922/Multiprocessing-bug.-Can-only-run-in-debugger.), I'll close this.
### Describe the bug I'm using a map function that returns more rows than are passed in. If I try to use `num_proc` I get: ``` File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kw...
167
29
Setting `num_proc` errors when `.map` returns additional items. ### Describe the bug I'm using a map function that returns more rows than are passed in. If I try to use `num_proc` I get: ``` File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper ...
[ -1.1727386713027954, -0.9139394760131836, -0.6932677626609802, 1.4026706218719482, -0.1811823844909668, -1.2461044788360596, 0.11010365188121796, -1.0843526124954224, 1.4787546396255493, -0.6602264046669006, 0.26320338249206543, -1.6500051021575928, -0.14681124687194824, -0.462516814470291...
https://github.com/huggingface/datasets/issues/5817
Setting `num_proc` errors when `.map` returns additional items.
For other users facing this, my workaround is to conditionally set `num_proc` so I can work interactively in the PyCharm Python Console while developing, then when I'm ready to run on the whole dataset, run it as a script and use multiprocessing. ```py mapped_ds = ds.map( my_map_function, batched=True, ...
### Describe the bug I'm using a map function that returns more rows than are passed in. If I try to use `num_proc` I get: ``` File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kw...
167
58
Setting `num_proc` errors when `.map` returns additional items. ### Describe the bug I'm using a map function that returns more rows than are passed in. If I try to use `num_proc` I get: ``` File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper ...
[ -1.1727386713027954, -0.9139394760131836, -0.6932677626609802, 1.4026706218719482, -0.1811823844909668, -1.2461044788360596, 0.11010365188121796, -1.0843526124954224, 1.4787546396255493, -0.6602264046669006, 0.26320338249206543, -1.6500051021575928, -0.14681124687194824, -0.462516814470291...
https://github.com/huggingface/datasets/issues/5815
Easy way to create a Kaggle dataset from a Huggingface dataset?
Hi @hrbigelow , I'm no expert for such a question so I'll ping @lhoestq from the `datasets` library (also this issue could be moved there if someone with permission can do it :) )
I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset. While Kaggle does provide the option to create a dataset from a URI, that URI must point to a single file. For example: ![image](https://user...
168
34
Easy way to create a Kaggle dataset from a Huggingface dataset? I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset. While Kaggle does provide the option to create a dataset from a URI, that URI m...
[ -1.2396824359893799, -0.9163654446601868, -0.7634274959564209, 1.4499483108520508, -0.1522327959537506, -1.2989747524261475, 0.029951062053442, -1.106934905052185, 1.6886955499649048, -0.8171396851539612, 0.34949982166290283, -1.727960467338562, 0.008662544190883636, -0.6666302680969238, ...
https://github.com/huggingface/datasets/issues/5815
Easy way to create a Kaggle dataset from a Huggingface dataset?
Hi ! Many datasets are made of several files, and how they are parsed often requires a python script. Because of that, datasets like wmt14 are not available as a single file on HF. Though you can create this file using `datasets`: ```python from datasets import load_dataset ds = load_dataset("wmt14", "de-en", sp...
I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset. While Kaggle does provide the option to create a dataset from a URI, that URI must point to a single file. For example: ![image](https://user...
168
102
Easy way to create a Kaggle dataset from a Huggingface dataset? I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset. While Kaggle does provide the option to create a dataset from a URI, that URI m...
[ -1.2066450119018555, -0.9107058048248291, -0.7251673340797424, 1.4683332443237305, -0.1572166085243225, -1.2567741870880127, 0.08483322709798813, -1.0845699310302734, 1.7319858074188232, -0.8124799132347107, 0.322973370552063, -1.685076117515564, 0.01419223751872778, -0.5806387066841125, ...
https://github.com/huggingface/datasets/issues/5815
Easy way to create a Kaggle dataset from a Huggingface dataset?
I've requested support for creating a Kaggle dataset from an imported HF dataset repo on their "forum" here: https://www.kaggle.com/discussions/product-feedback/427142 (upvotes appreciated 🙂)
I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset. While Kaggle does provide the option to create a dataset from a URI, that URI must point to a single file. For example: ![image](https://user...
168
22
Easy way to create a Kaggle dataset from a Huggingface dataset? I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset. While Kaggle does provide the option to create a dataset from a URI, that URI m...
[ -1.2180674076080322, -0.915431022644043, -0.7667586207389832, 1.4365451335906982, -0.1447499394416809, -1.328983187675476, 0.01761925034224987, -1.082245945930481, 1.6956593990325928, -0.8153727650642395, 0.3435761034488678, -1.7329462766647339, 0.0023516323417425156, -0.627507746219635, ...
https://github.com/huggingface/datasets/issues/5811
load_dataset: TypeError: 'NoneType' object is not callable, on local dataset filename changes
This error means a `DatasetBuilder` subclass that generates the dataset could not be found inside the script, so make sure `dushowxa-characters/dushowxa-characters.py `is a valid dataset script (assuming `path_or_dataset` is `dushowxa-characters`) Also, we should improve the error to make it more obvious what the pr...
### Describe the bug I've adapted Databrick's [train_dolly.py](/databrickslabs/dolly/blob/master/train_dolly.py) to train using a local dataset, which has been working. Upon changing the filenames of the `.json` & `.py` files in my local dataset directory, `dataset = load_dataset(path_or_dataset)["train"]` throws th...
169
45
load_dataset: TypeError: 'NoneType' object is not callable, on local dataset filename changes ### Describe the bug I've adapted Databrick's [train_dolly.py](/databrickslabs/dolly/blob/master/train_dolly.py) to train using a local dataset, which has been working. Upon changing the filenames of the `.json` & `.py` f...
[ -1.1963320970535278, -0.8724532127380371, -0.6726217269897461, 1.4031168222427368, -0.09186489135026932, -1.2873766422271729, 0.10977230966091156, -1.0881808996200562, 1.6891809701919556, -0.8217951059341431, 0.24431470036506653, -1.6554174423217773, -0.0670352652668953, -0.615050435066223...
https://github.com/huggingface/datasets/issues/5809
wiki_dpr details for Open Domain Question Answering tasks
Hi ! I don't remember exactly how it was done, but maybe you have to embed `f"{title}<sep>{text}"` ? Using a HF tokenizer it corresponds to doing ```python tokenized = tokenizer(titles, texts) ```
Hey guys! Thanks for creating the wiki_dpr dataset! I am currently trying to combine wiki_dpr and my own datasets. but I don't know how to make the embedding value the same way as wiki_dpr. As an experiment, I embeds the text of id="7" of wiki_dpr, but this result was very different from wiki_dpr.
170
32
wiki_dpr details for Open Domain Question Answering tasks Hey guys! Thanks for creating the wiki_dpr dataset! I am currently trying to combine wiki_dpr and my own datasets. but I don't know how to make the embedding value the same way as wiki_dpr. As an experiment, I embeds the text of id="7" of wiki_dpr, bu...
[ -1.2218024730682373, -0.8924668431282043, -0.7989919185638428, 1.584542989730835, -0.21870309114456177, -1.2377710342407227, 0.11410284042358398, -1.123988389968872, 1.6763991117477417, -1.0001040697097778, 0.3618963360786438, -1.6350091695785522, 0.03576774522662163, -0.541400671005249, ...
https://github.com/huggingface/datasets/issues/5806
Return the name of the currently loaded file in the load_dataset function.
Implementing this makes sense (e.g., `tensorflow_datasets`' imagefolder returns image filenames). Also, in Datasets 3.0, we plan only to store the bytes of an image/audio, not its path, so this feature would be useful when the path info is still needed.
### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Motivation When training large language models, machine problems may interrupt...
171
40
Return the name of the currently loaded file in the load_dataset function. ### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Mo...
[ -1.1434634923934937, -0.9244081377983093, -0.8602347373962402, 1.4389992952346802, -0.1318759173154831, -1.3050371408462524, 0.09337573498487473, -1.0374008417129517, 1.6407697200775146, -0.8867209553718567, 0.23797853291034698, -1.6980918645858765, -0.031181447207927704, -0.59579062461853...
https://github.com/huggingface/datasets/issues/5806
Return the name of the currently loaded file in the load_dataset function.
Hey @mariosasko, Can I work on this issue, this one seems interesting to implement. I have contributed to jupyterlab recently, and would love to contribute here as well.
### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Motivation When training large language models, machine problems may interrupt...
171
28
Return the name of the currently loaded file in the load_dataset function. ### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Mo...
[ -1.1340118646621704, -0.9239808320999146, -0.9057530760765076, 1.4740118980407715, -0.15070407092571259, -1.250565767288208, 0.07254789769649506, -1.060838222503662, 1.5978522300720215, -0.8674106597900391, 0.2628788948059082, -1.6827445030212402, 0.0029409388080239296, -0.6156407594680786...
https://github.com/huggingface/datasets/issues/5806
Return the name of the currently loaded file in the load_dataset function.
@tsabbir96 if you are planning to start working on this, you can take on this issue by writing a comment with only the keyword: #self-assign
### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Motivation When training large language models, machine problems may interrupt...
171
25
Return the name of the currently loaded file in the load_dataset function. ### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Mo...
[ -1.1221041679382324, -0.9181694984436035, -0.92864990234375, 1.4660353660583496, -0.12687699496746063, -1.285989761352539, 0.08314664661884308, -1.0642856359481812, 1.6114219427108765, -0.855737566947937, 0.26281461119651794, -1.6873915195465088, 0.000348774716258049, -0.6051350831985474, ...
https://github.com/huggingface/datasets/issues/5806
Return the name of the currently loaded file in the load_dataset function.
@albertvillanova thank you for letting me contribute here. @albertvillanova @mariosasko As I am totally new to this repo, could you tell me something more about this issue or perhaps give me some idea on how I can proceed with it? Thanks!
### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Motivation When training large language models, machine problems may interrupt...
171
41
Return the name of the currently loaded file in the load_dataset function. ### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Mo...
[ -1.1838386058807373, -0.9403164386749268, -0.8811663389205933, 1.478714108467102, -0.12408968061208725, -1.216644048690796, 0.10859201848506927, -1.043196439743042, 1.6114376783370972, -0.8631000518798828, 0.23932495713233948, -1.675797462463379, -0.01909806951880455, -0.5746715068817139, ...
https://github.com/huggingface/datasets/issues/5806
Return the name of the currently loaded file in the load_dataset function.
Hello there, is this issue resolved? @tsabbir96 are you still working on it? Otherwise I would love to give it a try
### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Motivation When training large language models, machine problems may interrupt...
171
22
Return the name of the currently loaded file in the load_dataset function. ### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Mo...
[ -1.1220901012420654, -0.9188336133956909, -0.9206020832061768, 1.4837565422058105, -0.12952572107315063, -1.2754708528518677, 0.07025479525327682, -1.0292587280273438, 1.6465485095977783, -0.8408465385437012, 0.25097814202308655, -1.6787211894989014, -0.008177501149475574, -0.5914601683616...
https://github.com/huggingface/datasets/issues/5806
Return the name of the currently loaded file in the load_dataset function.
Hey @mariosasko, I've taken the time to take a look at how we load the datasets usually. My main question now is about the final solution. So the idea is that whenever we load the datasets we also add a new column in the _generate_tables() method from the builders called filename (or file_name) that should be relate...
### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Motivation When training large language models, machine problems may interrupt...
171
74
Return the name of the currently loaded file in the load_dataset function. ### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Mo...
[ -1.1459382772445679, -0.9234645366668701, -0.870688796043396, 1.4424853324890137, -0.07883188873529434, -1.279450535774231, 0.08283881843090057, -1.0544235706329346, 1.643027663230896, -0.8828725814819336, 0.26749086380004883, -1.7021139860153198, -0.008590882644057274, -0.5644901990890503...
https://github.com/huggingface/datasets/issues/5805
Improve `Create a dataset` tutorial
I can work on this. The link to the tutorial seems to be broken though @polinaeterna.
Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading. 1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created from directory with data of required f...
172
16
Improve `Create a dataset` tutorial Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading. 1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created...
[ -1.1967639923095703, -0.8706130385398865, -0.7348169684410095, 1.4267008304595947, -0.12611046433448792, -1.2761508226394653, 0.038396887481212616, -1.041788101196289, 1.7662724256515503, -0.8197718262672424, 0.30907607078552246, -1.7420363426208496, 0.11263249814510345, -0.602516174316406...
https://github.com/huggingface/datasets/issues/5805
Improve `Create a dataset` tutorial
@isunitha98selvan would be great, thank you! which link are you talking about? I think it should work: https://huggingface.co/docs/datasets/create_dataset
Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading. 1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created from directory with data of required f...
172
18
Improve `Create a dataset` tutorial Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading. 1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created...
[ -1.1815582513809204, -0.8871431350708008, -0.7414290904998779, 1.395148754119873, -0.0902702659368515, -1.2827537059783936, 0.024044565856456757, -1.037650465965271, 1.7516496181488037, -0.793410062789917, 0.2662605345249176, -1.7461603879928589, 0.03457608446478844, -0.5815868377685547, ...
https://github.com/huggingface/datasets/issues/5798
Support parallelized downloading and processing in load_dataset with Spark
Hi ! We're using process pools for parallelism right now. I was wondering if there's a package that implements the same API as a process pool but runs with Spark under the hood ? That or something similar would be cool because users could use whatever distributed framework they want this way. Feel free to ping us wh...
### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```python load_dataset(..., use_spark=True) ``` ### Motivation Further speed up ...
173
81
Support parallelized downloading and processing in load_dataset with Spark ### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```pyth...
[ -1.2188078165054321, -0.9379724264144897, -0.8021289110183716, 1.559159517288208, -0.1595870405435562, -1.3118427991867065, 0.0809711217880249, -1.091163158416748, 1.7647982835769653, -0.8780925869941711, 0.34803804755210876, -1.636702537536621, 0.13220249116420746, -0.6404291391372681, ...
https://github.com/huggingface/datasets/issues/5798
Support parallelized downloading and processing in load_dataset with Spark
Hi, thanks for taking a look and providing your input! I don't know of such packages, and even it exists, I don't think with the process pool API it's possible to run Spark as backend properly; otherwise I understand a unified API would be preferable. The process pool API requires splitting the workload to a fixed n...
### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```python load_dataset(..., use_spark=True) ``` ### Motivation Further speed up ...
173
132
Support parallelized downloading and processing in load_dataset with Spark ### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```pyth...
[ -1.1745531558990479, -0.9043828248977661, -0.7972707152366638, 1.4532036781311035, -0.1395183950662613, -1.327866554260254, 0.08323068916797638, -1.0543800592422485, 1.7299998998641968, -0.8975285887718201, 0.33123713731765747, -1.5892170667648315, 0.07251695543527603, -0.6023445129394531,...
https://github.com/huggingface/datasets/issues/5798
Support parallelized downloading and processing in load_dataset with Spark
I see ! I think we only need an equivalent of `pool.map`. We use it to run download and conversion of data files on disk. That would require less changes in the internal code - and therefore less tests to write ;) We also use `pool.apply_async` in some places with a `Queue` to get progress updates of the running job...
### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```python load_dataset(..., use_spark=True) ``` ### Motivation Further speed up ...
173
84
Support parallelized downloading and processing in load_dataset with Spark ### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```pyth...
[ -1.1881126165390015, -0.9759367108345032, -0.7956238389015198, 1.563268780708313, -0.14781729876995087, -1.4093343019485474, 0.17102780938148499, -1.147376298904419, 1.7974903583526611, -0.9399683475494385, 0.39250436425209045, -1.6497114896774292, 0.12126322835683823, -0.6536846160888672,...
https://github.com/huggingface/datasets/issues/5798
Support parallelized downloading and processing in load_dataset with Spark
For Spark, `rdd.map` (where `rdd` can be created by `sparkContext.parallelize`) is the most similar as `pool.map`, but it requires creating a Spark RDD first that is used for distributing the `iterable` and the actual parallelization is managed by the Spark framework; `pool.map` takes the splits of `iterable` that are ...
### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```python load_dataset(..., use_spark=True) ``` ### Motivation Further speed up ...
173
127
Support parallelized downloading and processing in load_dataset with Spark ### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```pyth...
[ -1.1539030075073242, -0.8316649198532104, -0.8357244729995728, 1.500771164894104, -0.1406068503856659, -1.3223670721054077, 0.1813368946313858, -1.1577095985412598, 1.7817007303237915, -0.8220474123954773, 0.34623152017593384, -1.6578197479248047, 0.11396794021129608, -0.6205869317054749, ...
https://github.com/huggingface/datasets/issues/5798
Support parallelized downloading and processing in load_dataset with Spark
Indeed I think the current use of multiprocessing.Pool in `map_nested` can be rewritten to work like `sparkContext.parallelize` - without splitting the iterable. Maybe from the user's perspective it's ok to let multiprocessing.Pool or spark distribute the load on their own, as long as it takes a list and runs jobs i...
### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```python load_dataset(..., use_spark=True) ``` ### Motivation Further speed up ...
173
56
Support parallelized downloading and processing in load_dataset with Spark ### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```pyth...
[ -1.1622463464736938, -0.911980926990509, -0.8511636853218079, 1.5933550596237183, -0.15241794288158417, -1.3350327014923096, 0.18569372594356537, -1.1170432567596436, 1.7492403984069824, -0.880490779876709, 0.3449287414550781, -1.6263542175292969, 0.08924947679042816, -0.6167879700660706, ...