html_url stringlengths 51 51 | title stringlengths 6 280 | comments stringlengths 67 24.7k | body stringlengths 51 36.2k | __index_level_0__ int64 1 1.17k | comment_length int64 16 1.45k | text stringlengths 190 38.3k | embeddings list |
|---|---|---|---|---|---|---|---|
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | It appears that the parquet files at `hf://datasets/lambdalabs/pokemon-blip-captions` don't have this metadata, and it is defined in the dataset_infos.json instead (legacy).
You can get the right schema with the HF metadata this way:
```python
from datasets import load_dataset_builder
features = load_dataset_... | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 45 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | Btw in the future we might add support for an dedicated Image extension type in Arrow so that you won't need to add the schema metadata anymore ;) | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 28 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | Thanks @Wauplin @lhoestq for the quick reply :)!
I tried your approach by passing the huggingface schema to the dask writer
```
from datasets import Features, Image, Value
df = dd.read_parquet(f"hf://datasets/lambdalabs/pokemon-blip-captions",index=False)
features = Features({"image": Image(), "text": Value(... | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 152 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | You don't need the dataset_infos.json file as long as you have the schema with HF metadata ;)
I could also check that it works fine myself on the git revision without the dataset_infos.json file.
What made you think it didn't work ? | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 42 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | > You don't need the dataset_infos.json file as long as you have the schema with HF metadata ;) I could also check that it works fine myself on the git revision without the dataset_infos.json file.
>
> What made you think it didn't work ?
Those are two identical dataset repos where both were pushed with dask with... | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 116 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | I see ! I think it's a bug on our side - it should work without the metadata - let me investigate | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 22 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | Alright, it's fixed: https://huggingface.co/datasets/philippemo/dummy_dataset_without_schema
It shows the image correctly now - even without the extra metadata :) | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 17 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | Thanks @lhoestq!
I tested pushing a dataset again without the metadata and it works perfectly!
I appreciate the help | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 19 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | Hi @lhoestq,
I'v tried pushing another dataset again and I think the issue reappeared again:
```
df = dd.read_parquet(f"hf://datasets/lambdalabs/pokemon-blip-captions")
features = datasets.Features({"image": datasets.Image(), "text": datasets.Value("string")})
schema = features.arrow_schema
dd.to_parquet(df... | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 102 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | We're doing some changes in the way we're handling image parquet datasets right now. We'll include the fix from https://github.com/huggingface/datasets/pull/5921 in the new datasets-server version in the coming days | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 29 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | alright thanks for the update :), would that be part of the new release of datasets or is it something separate? if so, where can I track it? | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 28 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5869 | Image Encoding Issue when submitting a Parquet Dataset | Once the new version of `datasets` is released (tomorrow probably) we'll open an issue on https://github.com/huggingface/datasets-server to update to this version :) | ### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet... | 136 | 22 | Image Encoding Issue when submitting a Parquet Dataset
### Describe the bug
Hello,
I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details:
We attempted to load an example dataset in Parquet format from the Hugging Face (... | [
-1.1553672552108765,
-0.8704099059104919,
-0.8205088376998901,
1.4618592262268066,
-0.13831359148025513,
-1.3031227588653564,
0.08137144148349762,
-1.077457070350647,
1.7212998867034912,
-0.7961846590042114,
0.25587376952171326,
-1.7589480876922607,
-0.05134357139468193,
-0.533249974250793... |
https://github.com/huggingface/datasets/issues/5868 | Is it possible to change a cached file and 're-cache' it instead of re-generating? | Arrow files/primitives (tables and arrays) are immutable, so re-generating them is the only option, I'm afraid. | ### Feature request
Hi,
I have a huge cached file using `map`(over 500GB), and I want to change an attribution of each element, is there possible to do it using some method instead of re-generating, because `map` takes over 24 hours
### Motivation
For large datasets, I think it is very important because we always f... | 137 | 16 | Is it possible to change a cached file and 're-cache' it instead of re-generating?
### Feature request
Hi,
I have a huge cached file using `map`(over 500GB), and I want to change an attribution of each element, is there possible to do it using some method instead of re-generating, because `map` takes over 24 hours
... | [
-1.1673550605773926,
-0.9624342322349548,
-0.8973989486694336,
1.5511939525604248,
-0.15229423344135284,
-1.2601709365844727,
-0.014970440417528152,
-0.9520003199577332,
1.6296809911727905,
-0.8596575856208801,
0.3625255823135376,
-1.636365532875061,
0.0979856550693512,
-0.5727867484092712... |
https://github.com/huggingface/datasets/issues/5864 | Slow iteration over Torch tensors | I am highly interested performance of dataset so I ran your example as a curious user.
```python
train_dataset.cast_column("x", Array3D(shape=img_shape, dtype="float32"))
```
have return values and "x" is a new column, it shoulde be
```python
ds=train_dataset.cast_column("img", Array3D(shape=(3,32,32), dtype="flo... | ### Describe the bug
I have a problem related to this [issue](https://github.com/huggingface/datasets/issues/5841): I get a way slower iteration when using a Torch dataloader if I use vanilla Numpy tensors or if I first apply a ToTensor transform to the input. In particular, it takes 5 seconds to iterate over the vani... | 139 | 112 | Slow iteration over Torch tensors
### Describe the bug
I have a problem related to this [issue](https://github.com/huggingface/datasets/issues/5841): I get a way slower iteration when using a Torch dataloader if I use vanilla Numpy tensors or if I first apply a ToTensor transform to the input. In particular, it take... | [
-0.9561881422996521,
-0.947793185710907,
-0.7728447914123535,
1.6186071634292603,
-0.10586841404438019,
-1.2148075103759766,
0.30790024995803833,
-0.9178924560546875,
1.4549853801727295,
-0.717391312122345,
0.01934342086315155,
-1.6736693382263184,
-0.0020538344979286194,
-0.67579203844070... |
https://github.com/huggingface/datasets/issues/5862 | IndexError: list index out of range with data hosted on Zenodo | This error is also raised when data is hosted on Google Drive:
- https://huggingface.co/datasets/docred/discussions/5
- https://huggingface.co/datasets/linnaeus/discussions/3
- https://huggingface.co/datasets/poleval2019_mt/discussions/3
- https://huggingface.co/datasets/reddit_tifu/discussions/2
- https://hugging... | The dataset viewer sometimes raises an `IndexError`:
```
IndexError: list index out of range
```
See:
- huggingface/datasets-server#1151
- https://huggingface.co/datasets/reddit/discussions/5
- huggingface/datasets-server#1118
- https://huggingface.co/datasets/krr-oxford/OntoLAMA/discussions/1
- https://hu... | 140 | 26 | IndexError: list index out of range with data hosted on Zenodo
The dataset viewer sometimes raises an `IndexError`:
```
IndexError: list index out of range
```
See:
- huggingface/datasets-server#1151
- https://huggingface.co/datasets/reddit/discussions/5
- huggingface/datasets-server#1118
- https://huggin... | [
-1.005910038948059,
-0.9337819218635559,
-0.8602526187896729,
1.411230444908142,
0.06302760541439056,
-1.333472728729248,
0.07111027836799622,
-0.8140448331832886,
1.5044735670089722,
-0.5217399597167969,
0.25575631856918335,
-1.7397807836532593,
-0.236036017537117,
-0.5876840353012085,
... |
https://github.com/huggingface/datasets/issues/5858 | Throw an error when dataset improperly indexed | Thanks for reporting, @sarahwie.
Please note that in `datasets` we do not have vectorized operation like `pandas`. Therefore, your equality comparisons above are `False`:
- For example: `squad['question']` returns a `list`, and this list is not equal to `"Who was the Norse leader?"`
The `False` value is equivale... | ### Describe the bug
Pandas-style subset indexing on dataset does not throw an error, when maybe it should. Instead returns the first instance of the dataset regardless of index condition.
### Steps to reproduce the bug
Steps to reproduce the behavior:
1. `squad = datasets.load_dataset("squad_v2", split="validati... | 141 | 99 | Throw an error when dataset improperly indexed
### Describe the bug
Pandas-style subset indexing on dataset does not throw an error, when maybe it should. Instead returns the first instance of the dataset regardless of index condition.
### Steps to reproduce the bug
Steps to reproduce the behavior:
1. `squad = ... | [
-1.228513240814209,
-0.9444810748100281,
-0.7796606421470642,
1.574063777923584,
-0.16571839153766632,
-1.1630815267562866,
0.08318566530942917,
-1.0527688264846802,
1.683459997177124,
-0.7457794547080994,
0.2045617699623108,
-1.6489673852920532,
0.0285106860101223,
-0.6358548402786255,
... |
https://github.com/huggingface/datasets/issues/5857 | Adding chemistry dataset/models in huggingface | Hi!
This would be a nice addition to the Hub! You can find the existing chemistry datasets/models on the Hub (using the `chemistry` tag) [here](https://huggingface.co/search/full-text?q=chemistry&type=model&type=dataset).
Feel free to ping us here on the Hub if you need help adding the datasets.
| ### Feature request
Huggingface is really amazing platform for open science.
In addition to computer vision, video and NLP, would it be of interest to add chemistry/materials science dataset/models in Huggingface? Or, if its already done, can you provide some pointers.
We have been working on a comprehensive ben... | 142 | 41 | Adding chemistry dataset/models in huggingface
### Feature request
Huggingface is really amazing platform for open science.
In addition to computer vision, video and NLP, would it be of interest to add chemistry/materials science dataset/models in Huggingface? Or, if its already done, can you provide some pointer... | [
-1.1607859134674072,
-0.9598888158798218,
-0.8385828733444214,
1.5231138467788696,
-0.053224142640829086,
-1.287381649017334,
0.050442833453416824,
-0.9187561869621277,
1.6461793184280396,
-0.752089262008667,
0.3613884150981903,
-1.7062150239944458,
0.02108807861804962,
-0.6100136637687683... |
https://github.com/huggingface/datasets/issues/5856 | Error loading natural_questions | Hi! You can avoid this error by using the preprocessed version:
```python
import datasets
ds = datasets.load_dataset('natural_questions')
```
PS: Once we finish https://github.com/huggingface/datasets/pull/5364, this error will no longer be a problem. | ### Describe the bug
When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9:
```python
import datasets
datasets.load_dataset('natural_questions',beam_runner='DirectRunner')
```
It failed with following info:
`pyarrow.lib.ArrowNotImplementedError: Nested data conversions not impl... | 143 | 31 | Error loading natural_questions
### Describe the bug
When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9:
```python
import datasets
datasets.load_dataset('natural_questions',beam_runner='DirectRunner')
```
It failed with following info:
`pyarrow.lib.ArrowNotImplementedError... | [
-1.1990299224853516,
-0.9553800225257874,
-0.6237490177154541,
1.434041976928711,
-0.17947255074977875,
-1.2862924337387085,
0.11571666598320007,
-1.0527979135513306,
1.5706970691680908,
-0.7776874303817749,
0.20714278519153595,
-1.6421008110046387,
-0.09185336530208588,
-0.478563755750656... |
https://github.com/huggingface/datasets/issues/5856 | Error loading natural_questions | > Hi! You can avoid this error by using the preprocessed version:
>
> ```python
> import datasets
> ds = datasets.load_dataset('natural_questions')
> ```
>
> PS: Once we finish #5364, this error will no longer be a problem.
Thanks, wish #5364 finish early | ### Describe the bug
When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9:
```python
import datasets
datasets.load_dataset('natural_questions',beam_runner='DirectRunner')
```
It failed with following info:
`pyarrow.lib.ArrowNotImplementedError: Nested data conversions not impl... | 143 | 44 | Error loading natural_questions
### Describe the bug
When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9:
```python
import datasets
datasets.load_dataset('natural_questions',beam_runner='DirectRunner')
```
It failed with following info:
`pyarrow.lib.ArrowNotImplementedError... | [
-1.1990299224853516,
-0.9553800225257874,
-0.6237490177154541,
1.434041976928711,
-0.17947255074977875,
-1.2862924337387085,
0.11571666598320007,
-1.0527979135513306,
1.5706970691680908,
-0.7776874303817749,
0.20714278519153595,
-1.6421008110046387,
-0.09185336530208588,
-0.478563755750656... |
https://github.com/huggingface/datasets/issues/5855 | `to_tf_dataset` consumes too much memory | Cc @amyeroberts @Rocketknight1
Indded I think it's because it does something like this under the hood when there's no multiprocessing:
```python
tf_dataset = tf_dataset.shuffle(len(dataset))
```
PS: with multiprocessing it appears to be different:
```python
indices = np.arange(len(dataset))
if shuffle:... | ### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github.... | 144 | 41 | `to_tf_dataset` consumes too much memory
### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle ... | [
-1.21280837059021,
-0.9416200518608093,
-0.6731261014938354,
1.4203523397445679,
-0.16164042055606842,
-1.3120754957199097,
0.2671286165714264,
-1.0843902826309204,
1.663701057434082,
-0.8069621324539185,
0.256039023399353,
-1.6323555707931519,
0.015697987750172615,
-0.6138535141944885,
... |
https://github.com/huggingface/datasets/issues/5855 | `to_tf_dataset` consumes too much memory | Hi @massquantity, the dataset being shuffled there is not the full dataset. If you look at [the line above](https://github.com/huggingface/datasets/blob/main/src/datasets/utils/tf_utils.py#L182), the dataset is actually just a single indices array at that point, and that array is the only thing that gets fully loaded i... | ### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github.... | 144 | 134 | `to_tf_dataset` consumes too much memory
### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle ... | [
-1.21280837059021,
-0.9416200518608093,
-0.6731261014938354,
1.4203523397445679,
-0.16164042055606842,
-1.3120754957199097,
0.2671286165714264,
-1.0843902826309204,
1.663701057434082,
-0.8069621324539185,
0.256039023399353,
-1.6323555707931519,
0.015697987750172615,
-0.6138535141944885,
... |
https://github.com/huggingface/datasets/issues/5855 | `to_tf_dataset` consumes too much memory | Hi @Rocketknight1, you're right and I also noticed that only indices are used in shuffling. My data has shape (50000000, 10), but really the problem doesn't relate to a specific dataset. Simply running the following code costs me 10GB of memory.
```python
from datasets import Dataset
def gen():
for i in ran... | ### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github.... | 144 | 160 | `to_tf_dataset` consumes too much memory
### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle ... | [
-1.21280837059021,
-0.9416200518608093,
-0.6731261014938354,
1.4203523397445679,
-0.16164042055606842,
-1.3120754957199097,
0.2671286165714264,
-1.0843902826309204,
1.663701057434082,
-0.8069621324539185,
0.256039023399353,
-1.6323555707931519,
0.015697987750172615,
-0.6138535141944885,
... |
https://github.com/huggingface/datasets/issues/5855 | `to_tf_dataset` consumes too much memory | Thanks for that reproduction script - I've confirmed the same issue is occurring for me. Investigating it now! | ### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github.... | 144 | 18 | `to_tf_dataset` consumes too much memory
### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle ... | [
-1.21280837059021,
-0.9416200518608093,
-0.6731261014938354,
1.4203523397445679,
-0.16164042055606842,
-1.3120754957199097,
0.2671286165714264,
-1.0843902826309204,
1.663701057434082,
-0.8069621324539185,
0.256039023399353,
-1.6323555707931519,
0.015697987750172615,
-0.6138535141944885,
... |
https://github.com/huggingface/datasets/issues/5855 | `to_tf_dataset` consumes too much memory | Update: The memory usage is occurring in creation of the index and shuffle buffer. You can reproduce it very simply with:
```python
import tensorflow as tf
indices = tf.range(50_000_000, dtype=tf.int64)
dataset = tf.data.Dataset.from_tensor_slices(indices)
dataset = dataset.shuffle(len(dataset))
print(next(iter... | ### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github.... | 144 | 99 | `to_tf_dataset` consumes too much memory
### Describe the bug
Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`.
After some digging, i believe the reason lies in the shuffle ... | [
-1.21280837059021,
-0.9416200518608093,
-0.6731261014938354,
1.4203523397445679,
-0.16164042055606842,
-1.3120754957199097,
0.2671286165714264,
-1.0843902826309204,
1.663701057434082,
-0.8069621324539185,
0.256039023399353,
-1.6323555707931519,
0.015697987750172615,
-0.6138535141944885,
... |
https://github.com/huggingface/datasets/issues/5854 | Can not load audiofolder dataset on kaggle | Hi! `audiofolder` requires `datasets>=2.5.0`, so please update the `datasets`' installation (`pip install -U datasets`) in the environment to resolve the issue. | ### Describe the bug
It's crash log:
FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingfac... | 145 | 21 | Can not load audiofolder dataset on kaggle
### Describe the bug
It's crash log:
FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at... | [
-1.1507515907287598,
-0.9056360721588135,
-0.711188793182373,
1.461253046989441,
-0.15764841437339783,
-1.2411704063415527,
0.06300104409456253,
-0.9573824405670166,
1.6098647117614746,
-0.7448211908340454,
0.24856036901474,
-1.7515056133270264,
-0.0716467946767807,
-0.551703155040741,
-... |
https://github.com/huggingface/datasets/issues/5854 | Can not load audiofolder dataset on kaggle | > Hi! `audiofolder` requires `datasets>=2.5.0`, so please update the `datasets`' installation (`pip install -U datasets`) in the environment to resolve the issue.
I don't think it is a problem of the version. It runs ok on colab or local machine. Only on kaggle will has this bug. | ### Describe the bug
It's crash log:
FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingfac... | 145 | 47 | Can not load audiofolder dataset on kaggle
### Describe the bug
It's crash log:
FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at... | [
-1.154200792312622,
-0.9217584133148193,
-0.7177584767341614,
1.4420685768127441,
-0.15082411468029022,
-1.2470622062683105,
0.0723164901137352,
-0.9657632112503052,
1.6368587017059326,
-0.7386761903762817,
0.2630559504032135,
-1.738788366317749,
-0.06320306658744812,
-0.5161727666854858,
... |
https://github.com/huggingface/datasets/issues/5854 | Can not load audiofolder dataset on kaggle | Based on your dataset info, the installed version is `2.1.0`, which does not include `audiofolder`.
By default, Kaggle preinstalls `datasets` into a new env, but the version it installs is outdated and does not contain newer features such as `audiofolder` | ### Describe the bug
It's crash log:
FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingfac... | 145 | 40 | Can not load audiofolder dataset on kaggle
### Describe the bug
It's crash log:
FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at... | [
-1.1274595260620117,
-0.9151424765586853,
-0.7192152738571167,
1.441693902015686,
-0.15230119228363037,
-1.22844660282135,
0.08878697454929352,
-0.9621649384498596,
1.590338110923767,
-0.7543883323669434,
0.26574140787124634,
-1.7408254146575928,
-0.06892996281385422,
-0.539057731628418,
... |
https://github.com/huggingface/datasets/issues/5847 | Streaming IterableDataset not working with translation pipeline | [`KeyDataset`](https://github.com/huggingface/transformers/blob/7f8b909189547944617741d8d3c6c84504701693/src/transformers/pipelines/pt_utils.py#L296) doesn't support iterable datasets, so you either need to implement a version that does (and also indexing nested (translation) fields):
```python
from torch.utils.dat... | ### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that IterableDataset has no len().
### Steps to reproduce the bug
CODE:
```
... | 146 | 128 | Streaming IterableDataset not working with translation pipeline
### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that Iterable... | [
-1.2637747526168823,
-0.9374301433563232,
-0.7347829937934875,
1.4584630727767944,
-0.18755613267421722,
-1.1666781902313232,
0.17908449470996857,
-1.0838751792907715,
1.640811562538147,
-0.759229302406311,
0.2735395133495331,
-1.6184109449386597,
0.04956751689314842,
-0.5483786463737488,
... |
https://github.com/huggingface/datasets/issues/5847 | Streaming IterableDataset not working with translation pipeline | @mariosasko The map snippet didn't quite work, but gave me enough of a clue to get it working. The following snippet does work:
```
def en_translation(x):
return {"en":x['translation']['en']}
ks = ds.map(en_translation, remove_columns=['translation'])
test=[]
for x in iter(ks):
test.append(x['en'])
xx=... | ### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that IterableDataset has no len().
### Steps to reproduce the bug
CODE:
```
... | 146 | 219 | Streaming IterableDataset not working with translation pipeline
### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that Iterable... | [
-1.2637747526168823,
-0.9374301433563232,
-0.7347829937934875,
1.4584630727767944,
-0.18755613267421722,
-1.1666781902313232,
0.17908449470996857,
-1.0838751792907715,
1.640811562538147,
-0.759229302406311,
0.2735395133495331,
-1.6184109449386597,
0.04956751689314842,
-0.5483786463737488,
... |
https://github.com/huggingface/datasets/issues/5847 | Streaming IterableDataset not working with translation pipeline | So perhaps there's no bug exactly, but I would love to see two things: 1) improve the documentation to better understand what's really getting returned. 2) update the example provided of using transformer pipeline with a dataset to include the oddball case that translation appears to be. | ### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that IterableDataset has no len().
### Steps to reproduce the bug
CODE:
```
... | 146 | 47 | Streaming IterableDataset not working with translation pipeline
### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that Iterable... | [
-1.2637747526168823,
-0.9374301433563232,
-0.7347829937934875,
1.4584630727767944,
-0.18755613267421722,
-1.1666781902313232,
0.17908449470996857,
-1.0838751792907715,
1.640811562538147,
-0.759229302406311,
0.2735395133495331,
-1.6184109449386597,
0.04956751689314842,
-0.5483786463737488,
... |
https://github.com/huggingface/datasets/issues/5847 | Streaming IterableDataset not working with translation pipeline | Hi,
for the original snippet, the issue is that `streaming` datasets are not countable (they have no len) and therefore `KeyDataset` cannot work with them ( KeyDataset is a dataset and therefore requires a length).
I modified slightly the original snippet to make it work:
```python
from transformers import pi... | ### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that IterableDataset has no len().
### Steps to reproduce the bug
CODE:
```
... | 146 | 166 | Streaming IterableDataset not working with translation pipeline
### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that Iterable... | [
-1.2637747526168823,
-0.9374301433563232,
-0.7347829937934875,
1.4584630727767944,
-0.18755613267421722,
-1.1666781902313232,
0.17908449470996857,
-1.0838751792907715,
1.640811562538147,
-0.759229302406311,
0.2735395133495331,
-1.6184109449386597,
0.04956751689314842,
-0.5483786463737488,
... |
https://github.com/huggingface/datasets/issues/5847 | Streaming IterableDataset not working with translation pipeline | Maybe moving `KeyDataset` later on in the guide and specify it's mostly for streaming then? Or is it also necessary for batch_size>1 (which is what the current doc implies)? | ### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that IterableDataset has no len().
### Steps to reproduce the bug
CODE:
```
... | 146 | 29 | Streaming IterableDataset not working with translation pipeline
### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that Iterable... | [
-1.2637747526168823,
-0.9374301433563232,
-0.7347829937934875,
1.4584630727767944,
-0.18755613267421722,
-1.1666781902313232,
0.17908449470996857,
-1.0838751792907715,
1.640811562538147,
-0.759229302406311,
0.2735395133495331,
-1.6184109449386597,
0.04956751689314842,
-0.5483786463737488,
... |
https://github.com/huggingface/datasets/issues/5847 | Streaming IterableDataset not working with translation pipeline | Hmm
Iterator (`yield`) :
- Not countable
- Super flexible
- Cannot use `num_workers>1` (threading requires indexing at the correct location, iterators require to iterate in order,so each thread would iterate over the full thing being genuinely a bad idea)
- Can batch
- tqdm doesn't show a nice progress bar (it ... | ### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that IterableDataset has no len().
### Steps to reproduce the bug
CODE:
```
... | 146 | 188 | Streaming IterableDataset not working with translation pipeline
### Describe the bug
I'm trying to use a streaming dataset for translation inference to avoid downloading the training data.
I'm using a pipeline and a dataset, and following the guidance in the tutorial.
Instead I get an exception that Iterable... | [
-1.2637747526168823,
-0.9374301433563232,
-0.7347829937934875,
1.4584630727767944,
-0.18755613267421722,
-1.1666781902313232,
0.17908449470996857,
-1.0838751792907715,
1.640811562538147,
-0.759229302406311,
0.2735395133495331,
-1.6184109449386597,
0.04956751689314842,
-0.5483786463737488,
... |
https://github.com/huggingface/datasets/issues/5846 | load_dataset('bigcode/the-stack-dedup', streaming=True) very slow! | This is due to the slow resolution of the data files: https://github.com/huggingface/datasets/issues/5537.
We plan to switch to `huggingface_hub`'s `HfFileSystem` soon to make the resolution faster (will be up to 20x faster once we merge https://github.com/huggingface/huggingface_hub/pull/1443)
| ### Describe the bug
Running
```
import datasets
ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True)
```
takes about 2.5 minutes!
I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds.
### Environment info
- `datasets` version: 2.1... | 147 | 35 | load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
### Describe the bug
Running
```
import datasets
ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True)
```
takes about 2.5 minutes!
I would expect this to be near instantaneous. With other datasets, the runtime is one o... | [
-1.126883864402771,
-0.9017356038093567,
-0.7633296847343445,
1.486032247543335,
-0.11794661730527878,
-1.3176013231277466,
0.11445632576942444,
-0.9202184677124023,
1.6097322702407837,
-0.7836844325065613,
0.353589802980423,
-1.73179030418396,
-0.009220968931913376,
-0.6028525829315186,
... |
https://github.com/huggingface/datasets/issues/5846 | load_dataset('bigcode/the-stack-dedup', streaming=True) very slow! | You're right, when I try to parse more than 50GB of text data, I also get very slow, usually taking hours or even tens of hours. | ### Describe the bug
Running
```
import datasets
ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True)
```
takes about 2.5 minutes!
I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds.
### Environment info
- `datasets` version: 2.1... | 147 | 26 | load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
### Describe the bug
Running
```
import datasets
ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True)
```
takes about 2.5 minutes!
I would expect this to be near instantaneous. With other datasets, the runtime is one o... | [
-1.2048887014389038,
-0.8930887579917908,
-0.7761214375495911,
1.4448931217193604,
-0.17262375354766846,
-1.313103437423706,
0.11856652051210403,
-0.9846729636192322,
1.634273648262024,
-0.8477228879928589,
0.3353094756603241,
-1.7674273252487183,
0.044337861239910126,
-0.5383802652359009,... |
https://github.com/huggingface/datasets/issues/5846 | load_dataset('bigcode/the-stack-dedup', streaming=True) very slow! | > You're right, when I try to parse more than 50GB of text data, I also get very slow, usually taking hours or even tens of hours.
That's unrelated to the problem discussed in this issue. | ### Describe the bug
Running
```
import datasets
ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True)
```
takes about 2.5 minutes!
I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds.
### Environment info
- `datasets` version: 2.1... | 147 | 36 | load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
### Describe the bug
Running
```
import datasets
ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True)
```
takes about 2.5 minutes!
I would expect this to be near instantaneous. With other datasets, the runtime is one o... | [
-1.2141969203948975,
-0.9122954607009888,
-0.7814145684242249,
1.449281930923462,
-0.1722673773765564,
-1.3275378942489624,
0.10293480008840561,
-0.9914499521255493,
1.6627647876739502,
-0.8441689014434814,
0.31415659189224243,
-1.7814815044403076,
0.06872977316379547,
-0.5377781987190247,... |
https://github.com/huggingface/datasets/issues/5846 | load_dataset('bigcode/the-stack-dedup', streaming=True) very slow! | > > You're right, when I try to parse more than 50GB of text data, I also get very slow, usually taking hours or even tens of hours.
>
> That's unrelated to the problem discussed in this issue.
Sorry, I misunderstood it. | ### Describe the bug
Running
```
import datasets
ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True)
```
takes about 2.5 minutes!
I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds.
### Environment info
- `datasets` version: 2.1... | 147 | 43 | load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
### Describe the bug
Running
```
import datasets
ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True)
```
takes about 2.5 minutes!
I would expect this to be near instantaneous. With other datasets, the runtime is one o... | [
-1.2186570167541504,
-0.894500195980072,
-0.7827442288398743,
1.4336414337158203,
-0.18983258306980133,
-1.3164923191070557,
0.08473926037549973,
-1.0351999998092651,
1.6580960750579834,
-0.8356509804725647,
0.30576083064079285,
-1.7782236337661743,
0.0691995918750763,
-0.5293062329292297,... |
https://github.com/huggingface/datasets/issues/5841 | Abusurdly slow on iteration | Hi ! You can try to use the [Image](https://huggingface.co/docs/datasets/v2.12.0/en/package_reference/main_classes#datasets.Image) type which [decodes images on-the-fly](https://huggingface.co/docs/datasets/v2.12.0/en/about_dataset_features#image-feature) into pytorch tensors :)
```python
ds = Dataset.from_dict({"t... | ### Describe the bug
I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment:
```python
a=torch.randn(100,224)
a=torch.stack([a] * 10000)
a.shape
# %%
ds=Dataset.from_d... | 148 | 226 | Abusurdly slow on iteration
### Describe the bug
I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment:
```python
a=torch.randn(100,224)
a=torch.stack([a] * 10000)
a.sh... | [
-1.2770248651504517,
-0.8749205470085144,
-0.7151722311973572,
1.4440370798110962,
-0.22366788983345032,
-1.2704261541366577,
0.19402125477790833,
-1.0399311780929565,
1.7257866859436035,
-0.8413228988647461,
0.34074726700782776,
-1.6853406429290771,
0.12786391377449036,
-0.534410238265991... |
https://github.com/huggingface/datasets/issues/5841 | Abusurdly slow on iteration | Anyway, regarding the speed difference between numpy and pytorch, I think the issue is that we first convert numpy sub-arrays to pytorch and then consolidate into one tensor, while we should to the opposite. Indeed converting a numpy array to pytorch has a fix cost that seems to cause a slow down. The current pipeline ... | ### Describe the bug
I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment:
```python
a=torch.randn(100,224)
a=torch.stack([a] * 10000)
a.shape
# %%
ds=Dataset.from_d... | 148 | 91 | Abusurdly slow on iteration
### Describe the bug
I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment:
```python
a=torch.randn(100,224)
a=torch.stack([a] * 10000)
a.sh... | [
-1.2770248651504517,
-0.8749205470085144,
-0.7151722311973572,
1.4440370798110962,
-0.22366788983345032,
-1.2704261541366577,
0.19402125477790833,
-1.0399311780929565,
1.7257866859436035,
-0.8413228988647461,
0.34074726700782776,
-1.6853406429290771,
0.12786391377449036,
-0.534410238265991... |
https://github.com/huggingface/datasets/issues/5841 | Abusurdly slow on iteration | I have a similar issue: iterating over a dataset takes 5s without applying any transform, but takes ~30s after applying a transform.
Here is the minimum code to reproduce the problem
```python
import numpy as np
from datasets import Dataset, DatasetDict, load_dataset, Array3D, Image, Features
from torch.utils.da... | ### Describe the bug
I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment:
```python
a=torch.randn(100,224)
a=torch.stack([a] * 10000)
a.shape
# %%
ds=Dataset.from_d... | 148 | 187 | Abusurdly slow on iteration
### Describe the bug
I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment:
```python
a=torch.randn(100,224)
a=torch.stack([a] * 10000)
a.sh... | [
-1.2770248651504517,
-0.8749205470085144,
-0.7151722311973572,
1.4440370798110962,
-0.22366788983345032,
-1.2704261541366577,
0.19402125477790833,
-1.0399311780929565,
1.7257866859436035,
-0.8413228988647461,
0.34074726700782776,
-1.6853406429290771,
0.12786391377449036,
-0.534410238265991... |
https://github.com/huggingface/datasets/issues/5842 | Remove columns in interable dataset | This method has been recently added to the `IterableDataset`, so you need to update the `datasets`' installation (`pip install -U datasets`) to use it. | ### Feature request
Right now, remove_columns() produces a NotImplementedError for iterable style datasets
### Motivation
It would be great to have the same functionality irrespective of whether one is using an iterable or a map-style dataset
### Your contribution
hope and courage. | 150 | 24 | Remove columns in interable dataset
### Feature request
Right now, remove_columns() produces a NotImplementedError for iterable style datasets
### Motivation
It would be great to have the same functionality irrespective of whether one is using an iterable or a map-style dataset
### Your contribution
hope and cou... | [
-1.188256859779358,
-0.9099082350730896,
-0.8461495041847229,
1.46682870388031,
-0.10237018764019012,
-1.3405834436416626,
0.14300890266895294,
-1.020204782485962,
1.7700542211532593,
-0.7978855967521667,
0.3247756063938141,
-1.661026120185852,
0.018075844272971153,
-0.6116166114807129,
... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | As the name says, `load_from_disk` load the data from your disk. If the data is hosted on S3, it is first downloaded locally and then loaded from your disk.
There is a discussion on streaming data from S3 here though: #5281 | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 41 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.1314722299575806,
-1.0132465362548828,
-0.8381555080413818,
1.601589322090149,
-0.22572031617164612,
-1.320214033126831,
0.25365862250328064,
-1.107954740524292,
1.7764413356781006,
-1.0004515647888184,
0.37760037183761597,
-1.665255069732666,
0.09470492601394653,
-0.7143040299415588,
... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | @lhoestq
Thanks for your comment. I have checked out the discussion before and attempted at replicating the mentioned changes in the main branch (#5580). What I found was that if a dataset is saved using `save_to_disk`, it cannot be read by `load_dataset`. The error message asks me to to use `load_from_disk` instead.... | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 65 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.108872890472412,
-0.9334377646446228,
-0.8395171165466309,
1.5438768863677979,
-0.20451661944389343,
-1.3188121318817139,
0.22357897460460663,
-1.1496220827102661,
1.727474570274353,
-0.9264642000198364,
0.34790247678756714,
-1.6536660194396973,
0.14242373406887054,
-0.6958389282226562,... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | Using `push_to_hub` you can save the dataset on the HF Hub as parquet files, and reload it / stream it using `load_dataset` :)
If you want to save your dataset somewhere else you can use `.to_parquet` to get a parquet file. If your dataset is big it's usually recommended to shard it into multi parquet files (around ... | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 59 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.0931892395019531,
-0.9550338387489319,
-0.8697564601898193,
1.601189374923706,
-0.20103377103805542,
-1.3057785034179688,
0.1850281059741974,
-1.0832709074020386,
1.755860447883606,
-0.9503853917121887,
0.4491075575351715,
-1.6701313257217407,
0.15975995361804962,
-0.6955724954605103,
... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | @lhoestq
I tried the method you mentioned. This the current scenario I'm facing:
- The parquet file can be read from disk and streaming can be enabled.
- The parquet file can be read from `s3` (local MinIO).
- When `streaming=True` is enabled for `s3`, I get the error mentioned below:
```
File ~/.../lib/pytho... | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 106 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.18739652633667,
-0.9611495137214661,
-0.7330560088157654,
1.516315221786499,
-0.20272262394428253,
-1.3028978109359741,
0.24334430694580078,
-1.1313461065292358,
1.6697351932525635,
-0.8155365586280823,
0.3690471351146698,
-1.6336036920547485,
0.02870561182498932,
-0.7040586471557617,
... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | Streaming from S3 is still experimental, there might be a few bugs unfortunately.
Can you share the full stack trace ? | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 21 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.1281518936157227,
-0.956210732460022,
-0.8254011273384094,
1.5410455465316772,
-0.2120295614004135,
-1.3372948169708252,
0.2579007148742676,
-1.1193286180496216,
1.7654820680618286,
-0.9240826368331909,
0.34024980664253235,
-1.6349722146987915,
0.11261370778083801,
-0.6820954084396362,
... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | @lhoestq
Sure, here you go:
```python
TypeError Traceback (most recent call last)
Cell In[8], line 1
----> 1 dataset = load_dataset("parquet", data_files=["s3://<bucket name>/<data folder>/data-parquet"], storage_options=fs.storage_options, streaming=True)
File ~/.../datasets/... | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 1,037 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.2041233777999878,
-1.0062037706375122,
-0.6313180327415466,
1.437657117843628,
-0.18664082884788513,
-1.2890902757644653,
0.2077275812625885,
-1.1264944076538086,
1.6255488395690918,
-0.8505526781082153,
0.2045203596353531,
-1.7202439308166504,
-0.027755051851272583,
-0.6888807415962219... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | Is `"data-parquet"` a file ? In `data_files` you should pass the paths to the parquet files (not to a directory). Glob patterns are not supported yet for S3 URLs.
The bug seems to happen because your provided data file has no extension. Because of that it tries to infer it from the file content, but fails because `_... | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 63 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.1007708311080933,
-0.9273959994316101,
-0.8560488224029541,
1.587786316871643,
-0.18420720100402832,
-1.3287254571914673,
0.19336439669132233,
-1.1795574426651,
1.8246614933013916,
-0.943767249584198,
0.3852311670780182,
-1.6260333061218262,
0.16734789311885834,
-0.7044389247894287,
-... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | @lhoestq
Thank you for your answer. Saving the file with `.parquet` extension solved the issue! This is really great! Really appreciate all the help!
Let me know if I should close the issue or feel free to close it if you want. | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 42 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.1393764019012451,
-0.9299445152282715,
-0.8536918759346008,
1.56565523147583,
-0.20641542971134186,
-1.3232091665267944,
0.1775520145893097,
-1.209576964378357,
1.7437632083892822,
-0.9298161864280701,
0.3524990677833557,
-1.5843348503112793,
0.16504746675491333,
-0.6726346015930176,
... |
https://github.com/huggingface/datasets/issues/5838 | Streaming support for `load_from_disk` | Cool ! I'm glad it worked out :)
Sure feel free to close the issue, since the original question about streaming with load_from_disk has been answered anyway | ### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ... | 152 | 27 | Streaming support for `load_from_disk`
### Feature request
Support for streaming datasets stored in object stores in `load_from_disk`.
### Motivation
The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very... | [
-1.1667375564575195,
-0.9901631474494934,
-0.866985559463501,
1.560174584388733,
-0.20836511254310608,
-1.3153737783432007,
0.2408045083284378,
-1.1217237710952759,
1.729688286781311,
-0.9517273902893066,
0.30514436960220337,
-1.655367136001587,
0.15841801464557648,
-0.6709691286087036,
... |
https://github.com/huggingface/datasets/issues/5837 | Use DeepSpeed load myself " .csv " dataset. | Hi ! Doing `load_dataset("path/to/data.csv")` is not supported yet, but you can do
```python
ds = load_dataset("csv", data_files=["path/to/data.csv"])
``` | ### Describe the bug
When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", line 1767, in load_dataset
builder_instan... | 153 | 18 | Use DeepSpeed load myself " .csv " dataset.
### Describe the bug
When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", ... | [
-1.1078572273254395,
-0.9771057963371277,
-0.6886123418807983,
1.417290449142456,
-0.07183212786912918,
-1.2041537761688232,
0.13848580420017242,
-1.0561885833740234,
1.5372542142868042,
-0.702926516532898,
0.16869564354419708,
-1.7251721620559692,
-0.2262987643480301,
-0.4913672208786011,... |
https://github.com/huggingface/datasets/issues/5837 | Use DeepSpeed load myself " .csv " dataset. | The other question:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", line 1767, in load_dataset
builder_instance = load_dataset_builder(
File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/da... | ### Describe the bug
When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", line 1767, in load_dataset
builder_instan... | 153 | 229 | Use DeepSpeed load myself " .csv " dataset.
### Describe the bug
When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", ... | [
-1.1713345050811768,
-0.9852048754692078,
-0.6814395189285278,
1.3524361848831177,
-0.07349710166454315,
-1.2071709632873535,
0.13374069333076477,
-1.0489364862442017,
1.5145426988601685,
-0.6892326474189758,
0.1229124665260315,
-1.6937755346298218,
-0.23447763919830322,
-0.462883621454238... |
https://github.com/huggingface/datasets/issues/5834 | Is uint8 supported? | Hi ! The numpy formatting detaults to int64 and float32 - but you can use uint8 using
```python
ds = ds.with_format("numpy", dtype=np.uint8)
``` | ### Describe the bug
I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead.
While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well.
Is there a way ... | 154 | 23 | Is uint8 supported?
### Describe the bug
I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead.
While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as ... | [
-1.2034748792648315,
-0.887102484703064,
-0.7158054113388062,
1.5004172325134277,
-0.15594178438186646,
-1.2199872732162476,
0.1520499736070633,
-1.0931395292282104,
1.7031718492507935,
-0.7893821597099304,
0.375015527009964,
-1.6874333620071411,
-0.007075731176882982,
-0.5761370062828064,... |
https://github.com/huggingface/datasets/issues/5834 | Is uint8 supported? | Thank you!
By setting `ds.with_format("numpy", dtype=np.uint8)`, the dataset returns the data in `uint8`.
However, `with_format` and `set_format` seem to cast the data on-the-fly.
I want to reduce the dataset size by using `uint8` instead of `int64` and I observe no difference between using `int64` and `uint8` for... | ### Describe the bug
I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead.
While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well.
Is there a way ... | 154 | 74 | Is uint8 supported?
### Describe the bug
I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead.
While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as ... | [
-1.213528037071228,
-0.8775548338890076,
-0.7121641635894775,
1.4739090204238892,
-0.16309478878974915,
-1.2679214477539062,
0.1714174598455429,
-1.1260467767715454,
1.7571793794631958,
-0.8333533406257629,
0.3388484716415405,
-1.6772592067718506,
0.005824077874422073,
-0.6078845262527466,... |
https://github.com/huggingface/datasets/issues/5834 | Is uint8 supported? | If the feature type is `Value("uint8")` then it's written an uint8 on disk using the uint8 Arrow dtype.
e.g.
```python
ds = Dataset.from_dict({"a": range(10)}, features=Features({"a": Value("uint8")}))
ds.data.nbytes
# 10
``` | ### Describe the bug
I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead.
While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well.
Is there a way ... | 154 | 30 | Is uint8 supported?
### Describe the bug
I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead.
While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as ... | [
-1.1952255964279175,
-0.9149463176727295,
-0.7227889895439148,
1.5222022533416748,
-0.17281107604503632,
-1.211870551109314,
0.14803460240364075,
-1.0799599885940552,
1.7145726680755615,
-0.7580557465553284,
0.39515259861946106,
-1.6825087070465088,
-0.00803280621767044,
-0.594983041286468... |
https://github.com/huggingface/datasets/issues/5834 | Is uint8 supported? | Oh, I understand now.
The data was stored in `uint8` from the beginning (when the dataset returns `int64`).
Thank you for your time!
My question is fully resolved. | ### Describe the bug
I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead.
While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well.
Is there a way ... | 154 | 28 | Is uint8 supported?
### Describe the bug
I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead.
While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as ... | [
-1.2012821435928345,
-0.8838594555854797,
-0.7114223837852478,
1.4705657958984375,
-0.13130874931812286,
-1.212508201599121,
0.12983840703964233,
-1.0950819253921509,
1.7106516361236572,
-0.7925498485565186,
0.3421580195426941,
-1.699402928352356,
-0.021122673526406288,
-0.5970733165740967... |
https://github.com/huggingface/datasets/issues/5833 | Unable to push dataset - `create_pr` problem | Thanks for reporting, @agombert.
In this case, I think the root issue is authentication: before pushing to Hub, you should authenticate. See our docs: https://huggingface.co/docs/datasets/upload_dataset#upload-with-python
> 2. To upload a dataset on the Hub in Python, you need to log in to your Hugging Face account... | ### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
```python
dataset.push_to_hub("agomberto/Fr... | 155 | 50 | Unable to push dataset - `create_pr` problem
### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
... | [
-1.1540454626083374,
-0.7597602605819702,
-0.6611561179161072,
1.4835621118545532,
-0.0010568797588348389,
-1.429667592048645,
0.05472240969538689,
-0.9580106735229492,
1.5048177242279053,
-0.7912386655807495,
0.3948341906070709,
-1.7450616359710693,
0.05644043907523155,
-0.588965237140655... |
https://github.com/huggingface/datasets/issues/5833 | Unable to push dataset - `create_pr` problem | Hey @albertvillanova well I actually did :D
<img width="1079" alt="Capture d’écran 2023-05-09 à 18 02 58" src="https://github.com/huggingface/datasets/assets/17645711/e091aa20-06b1-4dd3-bfdb-35e832c66f8d">
| ### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
```python
dataset.push_to_hub("agomberto/Fr... | 155 | 17 | Unable to push dataset - `create_pr` problem
### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
... | [
-1.1540454626083374,
-0.7597602605819702,
-0.6611561179161072,
1.4835621118545532,
-0.0010568797588348389,
-1.429667592048645,
0.05472240969538689,
-0.9580106735229492,
1.5048177242279053,
-0.7912386655807495,
0.3948341906070709,
-1.7450616359710693,
0.05644043907523155,
-0.588965237140655... |
https://github.com/huggingface/datasets/issues/5833 | Unable to push dataset - `create_pr` problem | That is weird that you get a Forbidden error if you are properly authenticated...
Today we had a big outage issue affecting the Hugging Face Hub. Could you please retry to push_to_hub your dataset? Maybe that was the cause... | ### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
```python
dataset.push_to_hub("agomberto/Fr... | 155 | 39 | Unable to push dataset - `create_pr` problem
### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
... | [
-1.1540454626083374,
-0.7597602605819702,
-0.6611561179161072,
1.4835621118545532,
-0.0010568797588348389,
-1.429667592048645,
0.05472240969538689,
-0.9580106735229492,
1.5048177242279053,
-0.7912386655807495,
0.3948341906070709,
-1.7450616359710693,
0.05644043907523155,
-0.588965237140655... |
https://github.com/huggingface/datasets/issues/5833 | Unable to push dataset - `create_pr` problem | Login successful but also got this error "Forbidden: pass `create_pr=1` as a query parameter to create a Pull Request" | ### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
```python
dataset.push_to_hub("agomberto/Fr... | 155 | 19 | Unable to push dataset - `create_pr` problem
### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
... | [
-1.1540454626083374,
-0.7597602605819702,
-0.6611561179161072,
1.4835621118545532,
-0.0010568797588348389,
-1.429667592048645,
0.05472240969538689,
-0.9580106735229492,
1.5048177242279053,
-0.7912386655807495,
0.3948341906070709,
-1.7450616359710693,
0.05644043907523155,
-0.588965237140655... |
https://github.com/huggingface/datasets/issues/5833 | Unable to push dataset - `create_pr` problem | Make sure your API token has a `write` role. I had the same issue as you with the `read` token. Creating a `write` token and using that solved the issue. | ### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
```python
dataset.push_to_hub("agomberto/Fr... | 155 | 30 | Unable to push dataset - `create_pr` problem
### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
... | [
-1.1540454626083374,
-0.7597602605819702,
-0.6611561179161072,
1.4835621118545532,
-0.0010568797588348389,
-1.429667592048645,
0.05472240969538689,
-0.9580106735229492,
1.5048177242279053,
-0.7912386655807495,
0.3948341906070709,
-1.7450616359710693,
0.05644043907523155,
-0.588965237140655... |
https://github.com/huggingface/datasets/issues/5833 | Unable to push dataset - `create_pr` problem | > Make sure your API token has a `write` role. I had the same issue as you with the `read` token. Creating a `write` token and using that solved the issue.
I generate a token with write role. It works! thank you so much. | ### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
```python
dataset.push_to_hub("agomberto/Fr... | 155 | 44 | Unable to push dataset - `create_pr` problem
### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
... | [
-1.1540454626083374,
-0.7597602605819702,
-0.6611561179161072,
1.4835621118545532,
-0.0010568797588348389,
-1.429667592048645,
0.05472240969538689,
-0.9580106735229492,
1.5048177242279053,
-0.7912386655807495,
0.3948341906070709,
-1.7450616359710693,
0.05644043907523155,
-0.588965237140655... |
https://github.com/huggingface/datasets/issues/5833 | Unable to push dataset - `create_pr` problem | @dmitrijsk amazing thanks so much !
The error should be clearer when the token is read-only – I wasted a lot of time there.. | ### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
```python
dataset.push_to_hub("agomberto/Fr... | 155 | 24 | Unable to push dataset - `create_pr` problem
### Describe the bug
I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible.
### Steps to reproduce the bug
here what I have:
... | [
-1.1540454626083374,
-0.7597602605819702,
-0.6611561179161072,
1.4835621118545532,
-0.0010568797588348389,
-1.429667592048645,
0.05472240969538689,
-0.9580106735229492,
1.5048177242279053,
-0.7912386655807495,
0.3948341906070709,
-1.7450616359710693,
0.05644043907523155,
-0.588965237140655... |
https://github.com/huggingface/datasets/issues/5831 | [Bug]504 Server Error when loading dataset which was already cached | I am experiencing the same problem with the following environment:
* `datasets` version: 2.11.0
* Platform: `Linux 5.19.0-41-generic x86_64 GNU/Linux`
* Python version: `3.8.5`
* Huggingface_hub version: 0.13.3
* PyArrow version: `11.0.0`
* Pandas version: `1.5.3`
Trying to get some diagnostics, I got the fo... | ### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it again using the same machine, I got this error:
```
Traceback (most rece... | 157 | 120 | [Bug]504 Server Error when loading dataset which was already cached
### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it agai... | [
-1.1827178001403809,
-0.8930608630180359,
-0.7200658321380615,
1.3524177074432373,
-0.07334086298942566,
-1.2466628551483154,
0.04284411296248436,
-1.0322849750518799,
1.5563263893127441,
-0.6889245510101318,
0.17516101896762848,
-1.7303858995437622,
-0.2097388356924057,
-0.497710257768631... |
https://github.com/huggingface/datasets/issues/5831 | [Bug]504 Server Error when loading dataset which was already cached | same problem with a private dataset repo, seems the huggingface hub server got some connection problem? | ### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it again using the same machine, I got this error:
```
Traceback (most rece... | 157 | 16 | [Bug]504 Server Error when loading dataset which was already cached
### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it agai... | [
-1.1827178001403809,
-0.8930608630180359,
-0.7200658321380615,
1.3524177074432373,
-0.07334086298942566,
-1.2466628551483154,
0.04284411296248436,
-1.0322849750518799,
1.5563263893127441,
-0.6889245510101318,
0.17516101896762848,
-1.7303858995437622,
-0.2097388356924057,
-0.497710257768631... |
https://github.com/huggingface/datasets/issues/5831 | [Bug]504 Server Error when loading dataset which was already cached | @SingL3 You can avoid this error by setting the [`HF_DATASETS_OFFLINE`](https://huggingface.co/docs/datasets/v2.12.0/en/loading#offline) env variable to 1. By default, if an internet connection is available, we check whether the cache of a cached dataset is up-to-date.
@lucidBrot `datasets`' cache is still not alig... | ### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it again using the same machine, I got this error:
```
Traceback (most rece... | 157 | 48 | [Bug]504 Server Error when loading dataset which was already cached
### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it agai... | [
-1.1827178001403809,
-0.8930608630180359,
-0.7200658321380615,
1.3524177074432373,
-0.07334086298942566,
-1.2466628551483154,
0.04284411296248436,
-1.0322849750518799,
1.5563263893127441,
-0.6889245510101318,
0.17516101896762848,
-1.7303858995437622,
-0.2097388356924057,
-0.497710257768631... |
https://github.com/huggingface/datasets/issues/5831 | [Bug]504 Server Error when loading dataset which was already cached | Today we had a big issue affecting the Hugging Face Hub, thus all the `504 Server Error: Gateway Time-out` errors.
It is fixed now and loading your datasets should work as expected. | ### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it again using the same machine, I got this error:
```
Traceback (most rece... | 157 | 32 | [Bug]504 Server Error when loading dataset which was already cached
### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it agai... | [
-1.1827178001403809,
-0.8930608630180359,
-0.7200658321380615,
1.3524177074432373,
-0.07334086298942566,
-1.2466628551483154,
0.04284411296248436,
-1.0322849750518799,
1.5563263893127441,
-0.6889245510101318,
0.17516101896762848,
-1.7303858995437622,
-0.2097388356924057,
-0.497710257768631... |
https://github.com/huggingface/datasets/issues/5831 | [Bug]504 Server Error when loading dataset which was already cached | Hi, @albertvillanova.
If there is a locally cached version of datasets or something cache using huggingface_hub, when a network problem(either client or server) occurs, is it a better way to fallback to use the current cached version rather than raise a exception and exit? | ### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it again using the same machine, I got this error:
```
Traceback (most rece... | 157 | 44 | [Bug]504 Server Error when loading dataset which was already cached
### Describe the bug
I have already cached the dataset using:
```
dataset = load_dataset("databricks/databricks-dolly-15k",
cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k")
```
After that, I tried to load it agai... | [
-1.1827178001403809,
-0.8930608630180359,
-0.7200658321380615,
1.3524177074432373,
-0.07334086298942566,
-1.2466628551483154,
0.04284411296248436,
-1.0322849750518799,
1.5563263893127441,
-0.6889245510101318,
0.17516101896762848,
-1.7303858995437622,
-0.2097388356924057,
-0.497710257768631... |
https://github.com/huggingface/datasets/issues/5829 | (mach-o file, but is an incompatible architecture (have 'arm64', need 'x86_64')) | That is weird. I can't reproduce it again after reboot.
```python
In [2]: import platform
In [3]: platform.platform()
Out[3]: 'macOS-13.2-arm64-arm-64bit'
In [4]: from datasets import load_dataset
...:
...: jazzy = load_dataset("nomic-ai/gpt4all-j-prompt-generations", revision='v1.2-jazzy')
Found cach... | ### Describe the bug
M2 MBP can't run
```python
from datasets import load_dataset
jazzy = load_dataset("nomic-ai/gpt4all-j-prompt-generations", revision='v1.2-jazzy')
```
### Steps to reproduce the bug
1. Use M2 MBP
2. Python 3.10.10 from pyenv
3. Run
```
from datasets import load_dataset
jazzy = load_... | 158 | 42 | (mach-o file, but is an incompatible architecture (have 'arm64', need 'x86_64'))
### Describe the bug
M2 MBP can't run
```python
from datasets import load_dataset
jazzy = load_dataset("nomic-ai/gpt4all-j-prompt-generations", revision='v1.2-jazzy')
```
### Steps to reproduce the bug
1. Use M2 MBP
2. Python 3... | [
-1.2452958822250366,
-0.8774450421333313,
-0.686089277267456,
1.5009102821350098,
-0.1302803009748459,
-1.2884726524353027,
0.1107284277677536,
-1.034258246421814,
1.5793386697769165,
-0.8080062866210938,
0.3421730101108551,
-1.63355553150177,
-0.0004932684823870659,
-0.599759578704834,
... |
https://github.com/huggingface/datasets/issues/5828 | Stream data concatenation issue | Hi!
You can call `map` as follows to avoid the error:
```python
augmented_dataset_cln = dataset_cln['train'].map(augment_dataset, features=dataset_cln['train'].features)
``` | ### Describe the bug
I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset.
ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string',
id=None), 'audio': {'array': Sequence(feature=Value(dtype='float32', id=... | 159 | 17 | Stream data concatenation issue
### Describe the bug
I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset.
ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string',
id=None), 'audio': {'array': Sequence(... | [
-1.3739922046661377,
-1.1050032377243042,
-0.708509087562561,
1.662628173828125,
-0.27772530913352966,
-0.9798835515975952,
0.11023248732089996,
-1.0560520887374878,
1.5177412033081055,
-0.753578245639801,
0.24591761827468872,
-1.5724157094955444,
-0.014820529147982597,
-0.7283430099487305... |
https://github.com/huggingface/datasets/issues/5828 | Stream data concatenation issue | Hi!
I have run into the same problem with you. Could you please let me know how you solve it? Thanks! | ### Describe the bug
I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset.
ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string',
id=None), 'audio': {'array': Sequence(feature=Value(dtype='float32', id=... | 159 | 21 | Stream data concatenation issue
### Describe the bug
I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset.
ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string',
id=None), 'audio': {'array': Sequence(... | [
-1.3739922046661377,
-1.1050032377243042,
-0.708509087562561,
1.662628173828125,
-0.27772530913352966,
-0.9798835515975952,
0.11023248732089996,
-1.0560520887374878,
1.5177412033081055,
-0.753578245639801,
0.24591761827468872,
-1.5724157094955444,
-0.014820529147982597,
-0.7283430099487305... |
https://github.com/huggingface/datasets/issues/5827 | load json dataset interrupt when dtype cast problem occured | Indeed the JSON dataset builder raises an error when it encounters an unexpected type.
There's an old PR open to add away to ignore such elements though, if it can help: https://github.com/huggingface/datasets/pull/2838 | ### Describe the bug
i have a json like this:
[
{"id": 1, "name": 1},
{"id": 2, "name": "Nan"},
{"id": 3, "name": 3},
....
]
,which have several problematic rows data like row 2, then i load it with datasets.load_dataset('json', data_files=['xx.json'], split='train'), it will report like this:
... | 160 | 32 | load json dataset interrupt when dtype cast problem occured
### Describe the bug
i have a json like this:
[
{"id": 1, "name": 1},
{"id": 2, "name": "Nan"},
{"id": 3, "name": 3},
....
]
,which have several problematic rows data like row 2, then i load it with datasets.load_dataset('json', data... | [
-1.2560925483703613,
-0.9595218896865845,
-0.7310112714767456,
1.4875977039337158,
-0.19069238007068634,
-1.1312564611434937,
0.042799677699804306,
-1.043371558189392,
1.718193531036377,
-0.6977907419204712,
0.19077137112617493,
-1.6873589754104614,
0.020541850477457047,
-0.606010735034942... |
https://github.com/huggingface/datasets/issues/5825 | FileNotFound even though exists | Hi!
This would only work if `bigscience/xP3` was a no-code dataset, but it isn't (it has a Python builder script).
But this should work:
```python
load_dataset("json", data_files="https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.... | ### Describe the bug
I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong?
```
Downloading builder script: 100%
2.82k/2.8... | 161 | 28 | FileNotFound even though exists
### Describe the bug
I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong?
```
Downloadi... | [
-1.2246572971343994,
-0.7424915432929993,
-0.5879114270210266,
1.5263761281967163,
0.044311460107564926,
-1.3627488613128662,
0.0702437236905098,
-0.8897339105606079,
1.5228021144866943,
-0.792717456817627,
0.4206331670284271,
-1.6202031373977661,
0.06020117178559303,
-0.5875002145767212,
... |
https://github.com/huggingface/datasets/issues/5825 | FileNotFound even though exists | > I see, it's not compatible w/ regex right? e.g. `load_dataset("json", data_files="https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/*")`
It should work for patterns that "reference" the local filesystem, but to make this work with the Hub, we must implement https://github.com/huggingface/datasets/issu... | ### Describe the bug
I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong?
```
Downloading builder script: 100%
2.82k/2.8... | 161 | 106 | FileNotFound even though exists
### Describe the bug
I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong?
```
Downloadi... | [
-1.2246572971343994,
-0.7424915432929993,
-0.5879114270210266,
1.5263761281967163,
0.044311460107564926,
-1.3627488613128662,
0.0702437236905098,
-0.8897339105606079,
1.5228021144866943,
-0.792717456817627,
0.4206331670284271,
-1.6202031373977661,
0.06020117178559303,
-0.5875002145767212,
... |
https://github.com/huggingface/datasets/issues/5820 | Incomplete docstring for `BuilderConfig` | Thanks for reporting! You are more than welcome to improve `BuilderConfig`'s docstring.
This class serves an identical purpose as `tensorflow_datasets`'s `BuilderConfig`, and its docstring is [here](https://github.com/tensorflow/datasets/blob/a95e38b5bb018312c3d3720619c2a8ef83ebf57f/tensorflow_datasets/core/dataset_... | Hi guys !
I stumbled upon this docstring while working on a project.
Some of the attributes have missing descriptions.
https://github.com/huggingface/datasets/blob/bc5fef5b6d91f009e4101684adcb374df2c170f6/src/datasets/builder.py#L104-L117 | 164 | 34 | Incomplete docstring for `BuilderConfig`
Hi guys !
I stumbled upon this docstring while working on a project.
Some of the attributes have missing descriptions.
https://github.com/huggingface/datasets/blob/bc5fef5b6d91f009e4101684adcb374df2c170f6/src/datasets/builder.py#L104-L117
Thanks for reporting! You are m... | [
-1.3261981010437012,
-0.8913604617118835,
-0.7317317128181458,
1.399688959121704,
-0.09366767853498459,
-1.2431700229644775,
0.09195101261138916,
-1.0481846332550049,
1.6627223491668701,
-0.6732012033462524,
0.27303406596183777,
-1.6899288892745972,
-0.025011997669935226,
-0.53314298391342... |
https://github.com/huggingface/datasets/issues/5819 | Cannot pickle error in Dataset.from_generator() | Hi! It should work if you put `model = torch.compile(model)` inside the `generate_data` function. If a referenced object is outside, it needs to be pickable, and that's not the case for the compiled models (or functions). | ### Describe the bug
I'm trying to use Dataset.from_generator() to generate a large dataset.
### Steps to reproduce the bug
Code to reproduce:
```
from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig
import torch
from tqdm import tqdm
from datasets import load_dataset
tokenizer... | 165 | 36 | Cannot pickle error in Dataset.from_generator()
### Describe the bug
I'm trying to use Dataset.from_generator() to generate a large dataset.
### Steps to reproduce the bug
Code to reproduce:
```
from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig
import torch
from tqdm import tq... | [
-1.312134861946106,
-1.0129865407943726,
-0.5742440819740295,
1.5412966012954712,
-0.1450708657503128,
-1.1299595832824707,
0.22083961963653564,
-1.0590128898620605,
1.5314139127731323,
-0.8891088962554932,
0.3064503073692322,
-1.626840353012085,
0.027890335768461227,
-0.5555327534675598,
... |
https://github.com/huggingface/datasets/issues/5819 | Cannot pickle error in Dataset.from_generator() | > Hi! It should work if you put `model = torch.compile(model)` inside the `generate_data` function. If a referenced object is outside, it needs to be pickable, and that's not the case for the compiled models (or functions).
Hi! Thank you for your reply! Everything works perfectly with your suggestion!
Closing the... | ### Describe the bug
I'm trying to use Dataset.from_generator() to generate a large dataset.
### Steps to reproduce the bug
Code to reproduce:
```
from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig
import torch
from tqdm import tqdm
from datasets import load_dataset
tokenizer... | 165 | 52 | Cannot pickle error in Dataset.from_generator()
### Describe the bug
I'm trying to use Dataset.from_generator() to generate a large dataset.
### Steps to reproduce the bug
Code to reproduce:
```
from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig
import torch
from tqdm import tq... | [
-1.312134861946106,
-1.0129865407943726,
-0.5742440819740295,
1.5412966012954712,
-0.1450708657503128,
-1.1299595832824707,
0.22083961963653564,
-1.0590128898620605,
1.5314139127731323,
-0.8891088962554932,
0.3064503073692322,
-1.626840353012085,
0.027890335768461227,
-0.5555327534675598,
... |
https://github.com/huggingface/datasets/issues/5818 | Ability to update a dataset | In this case, I think we can avoid the `PermissionError` by unpacking the underlying `ConcatenationTable` and saving only the newly added data blocks (in new files). | ### Feature request
The ability to load a dataset, add or change something, and save it back to disk.
Maybe it's possible, but I can't work out how to do it, e.g. this fails:
```py
import datasets
dataset = datasets.load_from_disk("data/test1")
dataset = dataset.add_item({"text": "A new item"})
dataset.sav... | 166 | 26 | Ability to update a dataset
### Feature request
The ability to load a dataset, add or change something, and save it back to disk.
Maybe it's possible, but I can't work out how to do it, e.g. this fails:
```py
import datasets
dataset = datasets.load_from_disk("data/test1")
dataset = dataset.add_item({"text... | [
-1.1833677291870117,
-1.0026443004608154,
-0.8135682940483093,
1.444919228553772,
-0.15244489908218384,
-1.2608505487442017,
0.15001557767391205,
-1.05531644821167,
1.690877079963684,
-0.8565082550048828,
0.32282742857933044,
-1.6625043153762817,
0.006127838045358658,
-0.6394720077514648,
... |
https://github.com/huggingface/datasets/issues/5818 | Ability to update a dataset | Thanks @stevhliu and @mariosasko , so saving to individual files then loading them later, concatenating again and saving again is the recommended way. Good to know.
Question that I hope doesn't sound rude: is this sort of thing (processing a dataset that doesn't fit in memory) outside of `datasets`'s core area of fo... | ### Feature request
The ability to load a dataset, add or change something, and save it back to disk.
Maybe it's possible, but I can't work out how to do it, e.g. this fails:
```py
import datasets
dataset = datasets.load_from_disk("data/test1")
dataset = dataset.add_item({"text": "A new item"})
dataset.sav... | 166 | 89 | Ability to update a dataset
### Feature request
The ability to load a dataset, add or change something, and save it back to disk.
Maybe it's possible, but I can't work out how to do it, e.g. this fails:
```py
import datasets
dataset = datasets.load_from_disk("data/test1")
dataset = dataset.add_item({"text... | [
-1.1918529272079468,
-1.032199740409851,
-0.7797137498855591,
1.3972488641738892,
-0.16604754328727722,
-1.2583508491516113,
0.12762945890426636,
-1.0819259881973267,
1.7094367742538452,
-0.8580541610717773,
0.27730515599250793,
-1.664820909500122,
-0.0029700882732868195,
-0.64612269401550... |
https://github.com/huggingface/datasets/issues/5817 | Setting `num_proc` errors when `.map` returns additional items. | Hi ! Unfortunately I couldn't reproduce on my side locally and with datasets 2.11 and python 3.10.11 on colab.
What version of `multiprocess` are you using ? | ### Describe the bug
I'm using a map function that returns more rows than are passed in.
If I try to use `num_proc` I get:
```
File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper
out: Union["Dataset", "DatasetDict"] = func(self, *args, **kw... | 167 | 27 | Setting `num_proc` errors when `.map` returns additional items.
### Describe the bug
I'm using a map function that returns more rows than are passed in.
If I try to use `num_proc` I get:
```
File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper
... | [
-1.1727386713027954,
-0.9139394760131836,
-0.6932677626609802,
1.4026706218719482,
-0.1811823844909668,
-1.2461044788360596,
0.11010365188121796,
-1.0843526124954224,
1.4787546396255493,
-0.6602264046669006,
0.26320338249206543,
-1.6500051021575928,
-0.14681124687194824,
-0.462516814470291... |
https://github.com/huggingface/datasets/issues/5817 | Setting `num_proc` errors when `.map` returns additional items. | I've got `multiprocess` version `0.70.14`.
I've done some more testing and the error only occurs in PyCharm's Python Console. It seems to be [this PyCharm bug](https://youtrack.jetbrains.com/issue/PY-51922/Multiprocessing-bug.-Can-only-run-in-debugger.), I'll close this. | ### Describe the bug
I'm using a map function that returns more rows than are passed in.
If I try to use `num_proc` I get:
```
File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper
out: Union["Dataset", "DatasetDict"] = func(self, *args, **kw... | 167 | 29 | Setting `num_proc` errors when `.map` returns additional items.
### Describe the bug
I'm using a map function that returns more rows than are passed in.
If I try to use `num_proc` I get:
```
File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper
... | [
-1.1727386713027954,
-0.9139394760131836,
-0.6932677626609802,
1.4026706218719482,
-0.1811823844909668,
-1.2461044788360596,
0.11010365188121796,
-1.0843526124954224,
1.4787546396255493,
-0.6602264046669006,
0.26320338249206543,
-1.6500051021575928,
-0.14681124687194824,
-0.462516814470291... |
https://github.com/huggingface/datasets/issues/5817 | Setting `num_proc` errors when `.map` returns additional items. | For other users facing this, my workaround is to conditionally set `num_proc` so I can work interactively in the PyCharm Python Console while developing, then when I'm ready to run on the whole dataset, run it as a script and use multiprocessing.
```py
mapped_ds = ds.map(
my_map_function,
batched=True,
... | ### Describe the bug
I'm using a map function that returns more rows than are passed in.
If I try to use `num_proc` I get:
```
File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper
out: Union["Dataset", "DatasetDict"] = func(self, *args, **kw... | 167 | 58 | Setting `num_proc` errors when `.map` returns additional items.
### Describe the bug
I'm using a map function that returns more rows than are passed in.
If I try to use `num_proc` I get:
```
File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper
... | [
-1.1727386713027954,
-0.9139394760131836,
-0.6932677626609802,
1.4026706218719482,
-0.1811823844909668,
-1.2461044788360596,
0.11010365188121796,
-1.0843526124954224,
1.4787546396255493,
-0.6602264046669006,
0.26320338249206543,
-1.6500051021575928,
-0.14681124687194824,
-0.462516814470291... |
https://github.com/huggingface/datasets/issues/5815 | Easy way to create a Kaggle dataset from a Huggingface dataset? | Hi @hrbigelow , I'm no expert for such a question so I'll ping @lhoestq from the `datasets` library (also this issue could be moved there if someone with permission can do it :) ) | I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset.
While Kaggle does provide the option to create a dataset from a URI, that URI must point to a single file. For example:
 | I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset.
While Kaggle does provide the option to create a dataset from a URI, that URI must point to a single file. For example:

Also, we should improve the error to make it more obvious what the pr... | ### Describe the bug
I've adapted Databrick's [train_dolly.py](/databrickslabs/dolly/blob/master/train_dolly.py) to train using a local dataset, which has been working. Upon changing the filenames of the `.json` & `.py` files in my local dataset directory, `dataset = load_dataset(path_or_dataset)["train"]` throws th... | 169 | 45 | load_dataset: TypeError: 'NoneType' object is not callable, on local dataset filename changes
### Describe the bug
I've adapted Databrick's [train_dolly.py](/databrickslabs/dolly/blob/master/train_dolly.py) to train using a local dataset, which has been working. Upon changing the filenames of the `.json` & `.py` f... | [
-1.1963320970535278,
-0.8724532127380371,
-0.6726217269897461,
1.4031168222427368,
-0.09186489135026932,
-1.2873766422271729,
0.10977230966091156,
-1.0881808996200562,
1.6891809701919556,
-0.8217951059341431,
0.24431470036506653,
-1.6554174423217773,
-0.0670352652668953,
-0.615050435066223... |
https://github.com/huggingface/datasets/issues/5809 | wiki_dpr details for Open Domain Question Answering tasks | Hi ! I don't remember exactly how it was done, but maybe you have to embed `f"{title}<sep>{text}"` ?
Using a HF tokenizer it corresponds to doing
```python
tokenized = tokenizer(titles, texts)
``` | Hey guys!
Thanks for creating the wiki_dpr dataset!
I am currently trying to combine wiki_dpr and my own datasets. but I don't know how to make the embedding value the same way as wiki_dpr.
As an experiment, I embeds the text of id="7" of wiki_dpr, but this result was very different from wiki_dpr. | 170 | 32 | wiki_dpr details for Open Domain Question Answering tasks
Hey guys!
Thanks for creating the wiki_dpr dataset!
I am currently trying to combine wiki_dpr and my own datasets. but I don't know how to make the embedding value the same way as wiki_dpr.
As an experiment, I embeds the text of id="7" of wiki_dpr, bu... | [
-1.2218024730682373,
-0.8924668431282043,
-0.7989919185638428,
1.584542989730835,
-0.21870309114456177,
-1.2377710342407227,
0.11410284042358398,
-1.123988389968872,
1.6763991117477417,
-1.0001040697097778,
0.3618963360786438,
-1.6350091695785522,
0.03576774522662163,
-0.541400671005249,
... |
https://github.com/huggingface/datasets/issues/5806 | Return the name of the currently loaded file in the load_dataset function. | Implementing this makes sense (e.g., `tensorflow_datasets`' imagefolder returns image filenames). Also, in Datasets 3.0, we plan only to store the bytes of an image/audio, not its path, so this feature would be useful when the path info is still needed. | ### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Motivation
When training large language models, machine problems may interrupt... | 171 | 40 | Return the name of the currently loaded file in the load_dataset function.
### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Mo... | [
-1.1434634923934937,
-0.9244081377983093,
-0.8602347373962402,
1.4389992952346802,
-0.1318759173154831,
-1.3050371408462524,
0.09337573498487473,
-1.0374008417129517,
1.6407697200775146,
-0.8867209553718567,
0.23797853291034698,
-1.6980918645858765,
-0.031181447207927704,
-0.59579062461853... |
https://github.com/huggingface/datasets/issues/5806 | Return the name of the currently loaded file in the load_dataset function. | Hey @mariosasko, Can I work on this issue, this one seems interesting to implement. I have contributed to jupyterlab recently, and would love to contribute here as well. | ### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Motivation
When training large language models, machine problems may interrupt... | 171 | 28 | Return the name of the currently loaded file in the load_dataset function.
### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Mo... | [
-1.1340118646621704,
-0.9239808320999146,
-0.9057530760765076,
1.4740118980407715,
-0.15070407092571259,
-1.250565767288208,
0.07254789769649506,
-1.060838222503662,
1.5978522300720215,
-0.8674106597900391,
0.2628788948059082,
-1.6827445030212402,
0.0029409388080239296,
-0.6156407594680786... |
https://github.com/huggingface/datasets/issues/5806 | Return the name of the currently loaded file in the load_dataset function. | @tsabbir96 if you are planning to start working on this, you can take on this issue by writing a comment with only the keyword: #self-assign | ### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Motivation
When training large language models, machine problems may interrupt... | 171 | 25 | Return the name of the currently loaded file in the load_dataset function.
### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Mo... | [
-1.1221041679382324,
-0.9181694984436035,
-0.92864990234375,
1.4660353660583496,
-0.12687699496746063,
-1.285989761352539,
0.08314664661884308,
-1.0642856359481812,
1.6114219427108765,
-0.855737566947937,
0.26281461119651794,
-1.6873915195465088,
0.000348774716258049,
-0.6051350831985474,
... |
https://github.com/huggingface/datasets/issues/5806 | Return the name of the currently loaded file in the load_dataset function. | @albertvillanova thank you for letting me contribute here.
@albertvillanova @mariosasko As I am totally new to this repo, could you tell me something more about this issue or perhaps give me some idea on how I can proceed with it? Thanks! | ### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Motivation
When training large language models, machine problems may interrupt... | 171 | 41 | Return the name of the currently loaded file in the load_dataset function.
### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Mo... | [
-1.1838386058807373,
-0.9403164386749268,
-0.8811663389205933,
1.478714108467102,
-0.12408968061208725,
-1.216644048690796,
0.10859201848506927,
-1.043196439743042,
1.6114376783370972,
-0.8631000518798828,
0.23932495713233948,
-1.675797462463379,
-0.01909806951880455,
-0.5746715068817139,
... |
https://github.com/huggingface/datasets/issues/5806 | Return the name of the currently loaded file in the load_dataset function. | Hello there, is this issue resolved? @tsabbir96 are you still working on it? Otherwise I would love to give it a try | ### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Motivation
When training large language models, machine problems may interrupt... | 171 | 22 | Return the name of the currently loaded file in the load_dataset function.
### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Mo... | [
-1.1220901012420654,
-0.9188336133956909,
-0.9206020832061768,
1.4837565422058105,
-0.12952572107315063,
-1.2754708528518677,
0.07025479525327682,
-1.0292587280273438,
1.6465485095977783,
-0.8408465385437012,
0.25097814202308655,
-1.6787211894989014,
-0.008177501149475574,
-0.5914601683616... |
https://github.com/huggingface/datasets/issues/5806 | Return the name of the currently loaded file in the load_dataset function. | Hey @mariosasko, I've taken the time to take a look at how we load the datasets usually. My main question now is about the final solution.
So the idea is that whenever we load the datasets we also add a new column in the _generate_tables() method from the builders called filename (or file_name) that should be relate... | ### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Motivation
When training large language models, machine problems may interrupt... | 171 | 74 | Return the name of the currently loaded file in the load_dataset function.
### Feature request
Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output.
### Mo... | [
-1.1459382772445679,
-0.9234645366668701,
-0.870688796043396,
1.4424853324890137,
-0.07883188873529434,
-1.279450535774231,
0.08283881843090057,
-1.0544235706329346,
1.643027663230896,
-0.8828725814819336,
0.26749086380004883,
-1.7021139860153198,
-0.008590882644057274,
-0.5644901990890503... |
https://github.com/huggingface/datasets/issues/5805 | Improve `Create a dataset` tutorial | I can work on this. The link to the tutorial seems to be broken though @polinaeterna. | Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading.
1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created from directory with data of required f... | 172 | 16 | Improve `Create a dataset` tutorial
Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading.
1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created... | [
-1.1967639923095703,
-0.8706130385398865,
-0.7348169684410095,
1.4267008304595947,
-0.12611046433448792,
-1.2761508226394653,
0.038396887481212616,
-1.041788101196289,
1.7662724256515503,
-0.8197718262672424,
0.30907607078552246,
-1.7420363426208496,
0.11263249814510345,
-0.602516174316406... |
https://github.com/huggingface/datasets/issues/5805 | Improve `Create a dataset` tutorial | @isunitha98selvan would be great, thank you! which link are you talking about? I think it should work: https://huggingface.co/docs/datasets/create_dataset | Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading.
1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created from directory with data of required f... | 172 | 18 | Improve `Create a dataset` tutorial
Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading.
1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created... | [
-1.1815582513809204,
-0.8871431350708008,
-0.7414290904998779,
1.395148754119873,
-0.0902702659368515,
-1.2827537059783936,
0.024044565856456757,
-1.037650465965271,
1.7516496181488037,
-0.793410062789917,
0.2662605345249176,
-1.7461603879928589,
0.03457608446478844,
-0.5815868377685547,
... |
https://github.com/huggingface/datasets/issues/5798 | Support parallelized downloading and processing in load_dataset with Spark | Hi ! We're using process pools for parallelism right now. I was wondering if there's a package that implements the same API as a process pool but runs with Spark under the hood ? That or something similar would be cool because users could use whatever distributed framework they want this way.
Feel free to ping us wh... | ### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```python
load_dataset(..., use_spark=True)
```
### Motivation
Further speed up ... | 173 | 81 | Support parallelized downloading and processing in load_dataset with Spark
### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```pyth... | [
-1.2188078165054321,
-0.9379724264144897,
-0.8021289110183716,
1.559159517288208,
-0.1595870405435562,
-1.3118427991867065,
0.0809711217880249,
-1.091163158416748,
1.7647982835769653,
-0.8780925869941711,
0.34803804755210876,
-1.636702537536621,
0.13220249116420746,
-0.6404291391372681,
... |
https://github.com/huggingface/datasets/issues/5798 | Support parallelized downloading and processing in load_dataset with Spark | Hi, thanks for taking a look and providing your input! I don't know of such packages, and even it exists, I don't think with the process pool API it's possible to run Spark as backend properly; otherwise I understand a unified API would be preferable.
The process pool API requires splitting the workload to a fixed n... | ### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```python
load_dataset(..., use_spark=True)
```
### Motivation
Further speed up ... | 173 | 132 | Support parallelized downloading and processing in load_dataset with Spark
### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```pyth... | [
-1.1745531558990479,
-0.9043828248977661,
-0.7972707152366638,
1.4532036781311035,
-0.1395183950662613,
-1.327866554260254,
0.08323068916797638,
-1.0543800592422485,
1.7299998998641968,
-0.8975285887718201,
0.33123713731765747,
-1.5892170667648315,
0.07251695543527603,
-0.6023445129394531,... |
https://github.com/huggingface/datasets/issues/5798 | Support parallelized downloading and processing in load_dataset with Spark | I see ! I think we only need an equivalent of `pool.map`. We use it to run download and conversion of data files on disk. That would require less changes in the internal code - and therefore less tests to write ;)
We also use `pool.apply_async` in some places with a `Queue` to get progress updates of the running job... | ### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```python
load_dataset(..., use_spark=True)
```
### Motivation
Further speed up ... | 173 | 84 | Support parallelized downloading and processing in load_dataset with Spark
### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```pyth... | [
-1.1881126165390015,
-0.9759367108345032,
-0.7956238389015198,
1.563268780708313,
-0.14781729876995087,
-1.4093343019485474,
0.17102780938148499,
-1.147376298904419,
1.7974903583526611,
-0.9399683475494385,
0.39250436425209045,
-1.6497114896774292,
0.12126322835683823,
-0.6536846160888672,... |
https://github.com/huggingface/datasets/issues/5798 | Support parallelized downloading and processing in load_dataset with Spark | For Spark, `rdd.map` (where `rdd` can be created by `sparkContext.parallelize`) is the most similar as `pool.map`, but it requires creating a Spark RDD first that is used for distributing the `iterable` and the actual parallelization is managed by the Spark framework; `pool.map` takes the splits of `iterable` that are ... | ### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```python
load_dataset(..., use_spark=True)
```
### Motivation
Further speed up ... | 173 | 127 | Support parallelized downloading and processing in load_dataset with Spark
### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```pyth... | [
-1.1539030075073242,
-0.8316649198532104,
-0.8357244729995728,
1.500771164894104,
-0.1406068503856659,
-1.3223670721054077,
0.1813368946313858,
-1.1577095985412598,
1.7817007303237915,
-0.8220474123954773,
0.34623152017593384,
-1.6578197479248047,
0.11396794021129608,
-0.6205869317054749,
... |
https://github.com/huggingface/datasets/issues/5798 | Support parallelized downloading and processing in load_dataset with Spark | Indeed I think the current use of multiprocessing.Pool in `map_nested` can be rewritten to work like `sparkContext.parallelize` - without splitting the iterable.
Maybe from the user's perspective it's ok to let multiprocessing.Pool or spark distribute the load on their own, as long as it takes a list and runs jobs i... | ### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```python
load_dataset(..., use_spark=True)
```
### Motivation
Further speed up ... | 173 | 56 | Support parallelized downloading and processing in load_dataset with Spark
### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```pyth... | [
-1.1622463464736938,
-0.911980926990509,
-0.8511636853218079,
1.5933550596237183,
-0.15241794288158417,
-1.3350327014923096,
0.18569372594356537,
-1.1170432567596436,
1.7492403984069824,
-0.880490779876709,
0.3449287414550781,
-1.6263542175292969,
0.08924947679042816,
-0.6167879700660706,
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.