url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.11B
node_id
stringlengths
18
32
number
int64
1
3.59k
title
stringlengths
1
276
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
sequence
created_at
int64
1,587B
1,643B
updated_at
int64
1,587B
1,643B
closed_at
int64
1,587B
1,643B
βŒ€
author_association
stringclasses
3 values
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
0
228k
βŒ€
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/1660
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1660/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1660/comments
https://api.github.com/repos/huggingface/datasets/issues/1660/events
https://github.com/huggingface/datasets/pull/1660
775,831,423
MDExOlB1bGxSZXF1ZXN0NTQ2NDM2MDg1
1,660
add dataset info
{ "login": "harshalmittal4", "id": 24206326, "node_id": "MDQ6VXNlcjI0MjA2MzI2", "avatar_url": "https://avatars.githubusercontent.com/u/24206326?v=4", "gravatar_id": "", "url": "https://api.github.com/users/harshalmittal4", "html_url": "https://github.com/harshalmittal4", "followers_url": "https://api.github.com/users/harshalmittal4/followers", "following_url": "https://api.github.com/users/harshalmittal4/following{/other_user}", "gists_url": "https://api.github.com/users/harshalmittal4/gists{/gist_id}", "starred_url": "https://api.github.com/users/harshalmittal4/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/harshalmittal4/subscriptions", "organizations_url": "https://api.github.com/users/harshalmittal4/orgs", "repos_url": "https://api.github.com/users/harshalmittal4/repos", "events_url": "https://api.github.com/users/harshalmittal4/events{/privacy}", "received_events_url": "https://api.github.com/users/harshalmittal4/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,239,499,000
1,609,347,870,000
1,609,347,870,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1660", "html_url": "https://github.com/huggingface/datasets/pull/1660", "diff_url": "https://github.com/huggingface/datasets/pull/1660.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1660.patch", "merged_at": 1609347870000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1660/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1660/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1659
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1659/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1659/comments
https://api.github.com/repos/huggingface/datasets/issues/1659/events
https://github.com/huggingface/datasets/pull/1659
775,831,288
MDExOlB1bGxSZXF1ZXN0NTQ2NDM1OTcy
1,659
update dataset info
{ "login": "harshalmittal4", "id": 24206326, "node_id": "MDQ6VXNlcjI0MjA2MzI2", "avatar_url": "https://avatars.githubusercontent.com/u/24206326?v=4", "gravatar_id": "", "url": "https://api.github.com/users/harshalmittal4", "html_url": "https://github.com/harshalmittal4", "followers_url": "https://api.github.com/users/harshalmittal4/followers", "following_url": "https://api.github.com/users/harshalmittal4/following{/other_user}", "gists_url": "https://api.github.com/users/harshalmittal4/gists{/gist_id}", "starred_url": "https://api.github.com/users/harshalmittal4/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/harshalmittal4/subscriptions", "organizations_url": "https://api.github.com/users/harshalmittal4/orgs", "repos_url": "https://api.github.com/users/harshalmittal4/repos", "events_url": "https://api.github.com/users/harshalmittal4/events{/privacy}", "received_events_url": "https://api.github.com/users/harshalmittal4/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,239,481,000
1,609,347,307,000
1,609,347,307,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1659", "html_url": "https://github.com/huggingface/datasets/pull/1659", "diff_url": "https://github.com/huggingface/datasets/pull/1659.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1659.patch", "merged_at": 1609347307000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1659/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1659/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1658
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1658/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1658/comments
https://api.github.com/repos/huggingface/datasets/issues/1658/events
https://github.com/huggingface/datasets/pull/1658
775,651,085
MDExOlB1bGxSZXF1ZXN0NTQ2Mjg4Njg4
1,658
brwac dataset: add instances and data splits info
{ "login": "jonatasgrosman", "id": 5097052, "node_id": "MDQ6VXNlcjUwOTcwNTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5097052?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonatasgrosman", "html_url": "https://github.com/jonatasgrosman", "followers_url": "https://api.github.com/users/jonatasgrosman/followers", "following_url": "https://api.github.com/users/jonatasgrosman/following{/other_user}", "gists_url": "https://api.github.com/users/jonatasgrosman/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonatasgrosman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonatasgrosman/subscriptions", "organizations_url": "https://api.github.com/users/jonatasgrosman/orgs", "repos_url": "https://api.github.com/users/jonatasgrosman/repos", "events_url": "https://api.github.com/users/jonatasgrosman/events{/privacy}", "received_events_url": "https://api.github.com/users/jonatasgrosman/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,205,085,000
1,609,347,266,000
1,609,347,266,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1658", "html_url": "https://github.com/huggingface/datasets/pull/1658", "diff_url": "https://github.com/huggingface/datasets/pull/1658.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1658.patch", "merged_at": 1609347266000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1658/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1658/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1657
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1657/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1657/comments
https://api.github.com/repos/huggingface/datasets/issues/1657/events
https://github.com/huggingface/datasets/pull/1657
775,647,000
MDExOlB1bGxSZXF1ZXN0NTQ2Mjg1NjU2
1,657
mac_morpho dataset: add data splits info
{ "login": "jonatasgrosman", "id": 5097052, "node_id": "MDQ6VXNlcjUwOTcwNTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5097052?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonatasgrosman", "html_url": "https://github.com/jonatasgrosman", "followers_url": "https://api.github.com/users/jonatasgrosman/followers", "following_url": "https://api.github.com/users/jonatasgrosman/following{/other_user}", "gists_url": "https://api.github.com/users/jonatasgrosman/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonatasgrosman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonatasgrosman/subscriptions", "organizations_url": "https://api.github.com/users/jonatasgrosman/orgs", "repos_url": "https://api.github.com/users/jonatasgrosman/repos", "events_url": "https://api.github.com/users/jonatasgrosman/events{/privacy}", "received_events_url": "https://api.github.com/users/jonatasgrosman/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,203,921,000
1,609,347,084,000
1,609,347,084,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1657", "html_url": "https://github.com/huggingface/datasets/pull/1657", "diff_url": "https://github.com/huggingface/datasets/pull/1657.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1657.patch", "merged_at": 1609347084000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1657/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1657/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1656
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1656/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1656/comments
https://api.github.com/repos/huggingface/datasets/issues/1656/events
https://github.com/huggingface/datasets/pull/1656
775,645,356
MDExOlB1bGxSZXF1ZXN0NTQ2Mjg0NDI3
1,656
assin 2 dataset: add instances and data splits info
{ "login": "jonatasgrosman", "id": 5097052, "node_id": "MDQ6VXNlcjUwOTcwNTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5097052?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonatasgrosman", "html_url": "https://github.com/jonatasgrosman", "followers_url": "https://api.github.com/users/jonatasgrosman/followers", "following_url": "https://api.github.com/users/jonatasgrosman/following{/other_user}", "gists_url": "https://api.github.com/users/jonatasgrosman/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonatasgrosman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonatasgrosman/subscriptions", "organizations_url": "https://api.github.com/users/jonatasgrosman/orgs", "repos_url": "https://api.github.com/users/jonatasgrosman/repos", "events_url": "https://api.github.com/users/jonatasgrosman/events{/privacy}", "received_events_url": "https://api.github.com/users/jonatasgrosman/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,203,471,000
1,609,347,056,000
1,609,347,056,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1656", "html_url": "https://github.com/huggingface/datasets/pull/1656", "diff_url": "https://github.com/huggingface/datasets/pull/1656.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1656.patch", "merged_at": 1609347056000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1656/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1656/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1655
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1655/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1655/comments
https://api.github.com/repos/huggingface/datasets/issues/1655/events
https://github.com/huggingface/datasets/pull/1655
775,643,418
MDExOlB1bGxSZXF1ZXN0NTQ2MjgyOTM4
1,655
assin dataset: add instances and data splits info
{ "login": "jonatasgrosman", "id": 5097052, "node_id": "MDQ6VXNlcjUwOTcwNTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5097052?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonatasgrosman", "html_url": "https://github.com/jonatasgrosman", "followers_url": "https://api.github.com/users/jonatasgrosman/followers", "following_url": "https://api.github.com/users/jonatasgrosman/following{/other_user}", "gists_url": "https://api.github.com/users/jonatasgrosman/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonatasgrosman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonatasgrosman/subscriptions", "organizations_url": "https://api.github.com/users/jonatasgrosman/orgs", "repos_url": "https://api.github.com/users/jonatasgrosman/repos", "events_url": "https://api.github.com/users/jonatasgrosman/events{/privacy}", "received_events_url": "https://api.github.com/users/jonatasgrosman/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,202,876,000
1,609,347,023,000
1,609,347,023,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1655", "html_url": "https://github.com/huggingface/datasets/pull/1655", "diff_url": "https://github.com/huggingface/datasets/pull/1655.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1655.patch", "merged_at": 1609347022000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1655/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1655/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1654
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1654/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1654/comments
https://api.github.com/repos/huggingface/datasets/issues/1654/events
https://github.com/huggingface/datasets/pull/1654
775,640,729
MDExOlB1bGxSZXF1ZXN0NTQ2MjgwODIy
1,654
lener_br dataset: add instances and data splits info
{ "login": "jonatasgrosman", "id": 5097052, "node_id": "MDQ6VXNlcjUwOTcwNTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5097052?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonatasgrosman", "html_url": "https://github.com/jonatasgrosman", "followers_url": "https://api.github.com/users/jonatasgrosman/followers", "following_url": "https://api.github.com/users/jonatasgrosman/following{/other_user}", "gists_url": "https://api.github.com/users/jonatasgrosman/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonatasgrosman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonatasgrosman/subscriptions", "organizations_url": "https://api.github.com/users/jonatasgrosman/orgs", "repos_url": "https://api.github.com/users/jonatasgrosman/repos", "events_url": "https://api.github.com/users/jonatasgrosman/events{/privacy}", "received_events_url": "https://api.github.com/users/jonatasgrosman/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,202,112,000
1,609,346,972,000
1,609,346,972,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1654", "html_url": "https://github.com/huggingface/datasets/pull/1654", "diff_url": "https://github.com/huggingface/datasets/pull/1654.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1654.patch", "merged_at": 1609346972000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1654/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1654/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1653
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1653/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1653/comments
https://api.github.com/repos/huggingface/datasets/issues/1653/events
https://github.com/huggingface/datasets/pull/1653
775,632,945
MDExOlB1bGxSZXF1ZXN0NTQ2Mjc0Njc0
1,653
harem dataset: add data splits info
{ "login": "jonatasgrosman", "id": 5097052, "node_id": "MDQ6VXNlcjUwOTcwNTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5097052?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonatasgrosman", "html_url": "https://github.com/jonatasgrosman", "followers_url": "https://api.github.com/users/jonatasgrosman/followers", "following_url": "https://api.github.com/users/jonatasgrosman/following{/other_user}", "gists_url": "https://api.github.com/users/jonatasgrosman/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonatasgrosman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonatasgrosman/subscriptions", "organizations_url": "https://api.github.com/users/jonatasgrosman/orgs", "repos_url": "https://api.github.com/users/jonatasgrosman/repos", "events_url": "https://api.github.com/users/jonatasgrosman/events{/privacy}", "received_events_url": "https://api.github.com/users/jonatasgrosman/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,199,900,000
1,609,346,943,000
1,609,346,943,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1653", "html_url": "https://github.com/huggingface/datasets/pull/1653", "diff_url": "https://github.com/huggingface/datasets/pull/1653.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1653.patch", "merged_at": 1609346943000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1653/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1653/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1652
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1652/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1652/comments
https://api.github.com/repos/huggingface/datasets/issues/1652/events
https://github.com/huggingface/datasets/pull/1652
775,571,813
MDExOlB1bGxSZXF1ZXN0NTQ2MjI1NTM1
1,652
Update dataset cards from previous sprint
{ "login": "j-chim", "id": 22435209, "node_id": "MDQ6VXNlcjIyNDM1MjA5", "avatar_url": "https://avatars.githubusercontent.com/u/22435209?v=4", "gravatar_id": "", "url": "https://api.github.com/users/j-chim", "html_url": "https://github.com/j-chim", "followers_url": "https://api.github.com/users/j-chim/followers", "following_url": "https://api.github.com/users/j-chim/following{/other_user}", "gists_url": "https://api.github.com/users/j-chim/gists{/gist_id}", "starred_url": "https://api.github.com/users/j-chim/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/j-chim/subscriptions", "organizations_url": "https://api.github.com/users/j-chim/orgs", "repos_url": "https://api.github.com/users/j-chim/repos", "events_url": "https://api.github.com/users/j-chim/events{/privacy}", "received_events_url": "https://api.github.com/users/j-chim/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,186,847,000
1,609,346,884,000
1,609,346,884,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1652", "html_url": "https://github.com/huggingface/datasets/pull/1652", "diff_url": "https://github.com/huggingface/datasets/pull/1652.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1652.patch", "merged_at": 1609346884000 }
This PR updates the dataset cards/readmes for the 4 approved PRs I submitted in the previous sprint.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1652/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1652/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1651
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1651/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1651/comments
https://api.github.com/repos/huggingface/datasets/issues/1651/events
https://github.com/huggingface/datasets/pull/1651
775,554,319
MDExOlB1bGxSZXF1ZXN0NTQ2MjExMjQw
1,651
Add twi wordsim353
{ "login": "dadelani", "id": 23586676, "node_id": "MDQ6VXNlcjIzNTg2Njc2", "avatar_url": "https://avatars.githubusercontent.com/u/23586676?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dadelani", "html_url": "https://github.com/dadelani", "followers_url": "https://api.github.com/users/dadelani/followers", "following_url": "https://api.github.com/users/dadelani/following{/other_user}", "gists_url": "https://api.github.com/users/dadelani/gists{/gist_id}", "starred_url": "https://api.github.com/users/dadelani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dadelani/subscriptions", "organizations_url": "https://api.github.com/users/dadelani/orgs", "repos_url": "https://api.github.com/users/dadelani/repos", "events_url": "https://api.github.com/users/dadelani/events{/privacy}", "received_events_url": "https://api.github.com/users/dadelani/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Well actually it looks like it was already added in #1428 \r\n\r\nMaybe we can close this one ? Or you wanted to make changes to this dataset ?", "Thank you, it's just a modification of Readme. I added the missing citation.", "Indeed thanks" ]
1,609,183,915,000
1,609,753,179,000
1,609,753,178,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1651", "html_url": "https://github.com/huggingface/datasets/pull/1651", "diff_url": "https://github.com/huggingface/datasets/pull/1651.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1651.patch", "merged_at": 1609753178000 }
Added the citation information to the README file
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1651/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1651/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1650
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1650/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1650/comments
https://api.github.com/repos/huggingface/datasets/issues/1650/events
https://github.com/huggingface/datasets/pull/1650
775,545,912
MDExOlB1bGxSZXF1ZXN0NTQ2MjA0MzYy
1,650
Update README.md
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,182,545,000
1,609,238,594,000
1,609,238,594,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1650", "html_url": "https://github.com/huggingface/datasets/pull/1650", "diff_url": "https://github.com/huggingface/datasets/pull/1650.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1650.patch", "merged_at": 1609238594000 }
added dataset summary
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1650/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1650/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1649
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1649/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1649/comments
https://api.github.com/repos/huggingface/datasets/issues/1649/events
https://github.com/huggingface/datasets/pull/1649
775,544,487
MDExOlB1bGxSZXF1ZXN0NTQ2MjAzMjE1
1,649
Update README.md
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,182,300,000
1,609,239,058,000
1,609,238,583,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1649", "html_url": "https://github.com/huggingface/datasets/pull/1649", "diff_url": "https://github.com/huggingface/datasets/pull/1649.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1649.patch", "merged_at": 1609238583000 }
Added information in the dataset card
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1649/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1649/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1648
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1648/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1648/comments
https://api.github.com/repos/huggingface/datasets/issues/1648/events
https://github.com/huggingface/datasets/pull/1648
775,542,360
MDExOlB1bGxSZXF1ZXN0NTQ2MjAxNTQ0
1,648
Update README.md
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,181,946,000
1,609,238,354,000
1,609,238,354,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1648", "html_url": "https://github.com/huggingface/datasets/pull/1648", "diff_url": "https://github.com/huggingface/datasets/pull/1648.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1648.patch", "merged_at": 1609238354000 }
added dataset summary
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1648/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1648/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1647
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1647/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1647/comments
https://api.github.com/repos/huggingface/datasets/issues/1647/events
https://github.com/huggingface/datasets/issues/1647
775,525,799
MDU6SXNzdWU3NzU1MjU3OTk=
1,647
NarrativeQA fails to load with `load_dataset`
{ "login": "eric-mitchell", "id": 56408839, "node_id": "MDQ6VXNlcjU2NDA4ODM5", "avatar_url": "https://avatars.githubusercontent.com/u/56408839?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eric-mitchell", "html_url": "https://github.com/eric-mitchell", "followers_url": "https://api.github.com/users/eric-mitchell/followers", "following_url": "https://api.github.com/users/eric-mitchell/following{/other_user}", "gists_url": "https://api.github.com/users/eric-mitchell/gists{/gist_id}", "starred_url": "https://api.github.com/users/eric-mitchell/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eric-mitchell/subscriptions", "organizations_url": "https://api.github.com/users/eric-mitchell/orgs", "repos_url": "https://api.github.com/users/eric-mitchell/repos", "events_url": "https://api.github.com/users/eric-mitchell/events{/privacy}", "received_events_url": "https://api.github.com/users/eric-mitchell/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi @eric-mitchell,\r\nI think the issue might be that this dataset was added during the community sprint and has not been released yet. It will be available with the v2 of `datasets`.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of `datasets` using pip:\r\n`pip install git+https://github.com/huggingface/datasets.git@master`", "@bhavitvyamalik Great, thanks for this! Confirmed that the problem is resolved on master at [cbbda53](https://github.com/huggingface/datasets/commit/cbbda53ac1520b01f0f67ed6017003936c41ec59).", "Update: HuggingFace did an intermediate release yesterday just before the v2.0.\r\n\r\nTo load it you can just update `datasets`\r\n\r\n`pip install --upgrade datasets`" ]
1,609,179,369,000
1,609,848,308,000
1,609,696,685,000
NONE
null
null
null
When loading the NarrativeQA dataset with `load_dataset('narrativeqa')` as given in the documentation [here](https://huggingface.co/datasets/narrativeqa), I receive a cascade of exceptions, ending with FileNotFoundError: Couldn't find file locally at narrativeqa/narrativeqa.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/narrativeqa/narrativeqa.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/narrativeqa/narrativeqa.py Workaround: manually copy the `narrativeqa.py` builder into my local directory with curl https://raw.githubusercontent.com/huggingface/datasets/master/datasets/narrativeqa/narrativeqa.py -o narrativeqa.py and load the dataset as `load_dataset('narrativeqa.py')` everything works fine. I'm on datasets v1.1.3 using Python 3.6.10.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1647/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1647/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1646
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1646/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1646/comments
https://api.github.com/repos/huggingface/datasets/issues/1646/events
https://github.com/huggingface/datasets/pull/1646
775,499,344
MDExOlB1bGxSZXF1ZXN0NTQ2MTY4MTk3
1,646
Add missing homepage in some dataset cards
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,175,388,000
1,609,769,337,000
1,609,769,336,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1646", "html_url": "https://github.com/huggingface/datasets/pull/1646", "diff_url": "https://github.com/huggingface/datasets/pull/1646.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1646.patch", "merged_at": 1609769336000 }
In some dataset cards the homepage field in the `Dataset Description` section was missing/empty
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1646/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1646/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1645
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1645/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1645/comments
https://api.github.com/repos/huggingface/datasets/issues/1645/events
https://github.com/huggingface/datasets/pull/1645
775,473,106
MDExOlB1bGxSZXF1ZXN0NTQ2MTQ4OTUx
1,645
Rename "part-of-speech-tagging" tag in some dataset cards
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,171,749,000
1,610,014,094,000
1,610,014,093,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1645", "html_url": "https://github.com/huggingface/datasets/pull/1645", "diff_url": "https://github.com/huggingface/datasets/pull/1645.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1645.patch", "merged_at": 1610014093000 }
`part-of-speech-tagging` was not part of the tagging taxonomy under `structure-prediction`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1645/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1645/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1644
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1644/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1644/comments
https://api.github.com/repos/huggingface/datasets/issues/1644/events
https://github.com/huggingface/datasets/issues/1644
775,375,880
MDU6SXNzdWU3NzUzNzU4ODA=
1,644
HoVeR dataset fails to load
{ "login": "urikz", "id": 1473778, "node_id": "MDQ6VXNlcjE0NzM3Nzg=", "avatar_url": "https://avatars.githubusercontent.com/u/1473778?v=4", "gravatar_id": "", "url": "https://api.github.com/users/urikz", "html_url": "https://github.com/urikz", "followers_url": "https://api.github.com/users/urikz/followers", "following_url": "https://api.github.com/users/urikz/following{/other_user}", "gists_url": "https://api.github.com/users/urikz/gists{/gist_id}", "starred_url": "https://api.github.com/users/urikz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/urikz/subscriptions", "organizations_url": "https://api.github.com/users/urikz/orgs", "repos_url": "https://api.github.com/users/urikz/repos", "events_url": "https://api.github.com/users/urikz/events{/privacy}", "received_events_url": "https://api.github.com/users/urikz/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Hover was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `hover` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"hover\")\r\n```" ]
1,609,158,427,000
1,609,785,991,000
null
NONE
null
null
null
Hi! I'm getting an error when trying to load **HoVeR** dataset. Another one (**SQuAD**) does work for me. I'm using the latest (1.1.3) version of the library. Steps to reproduce the error: ```python >>> from datasets import load_dataset >>> dataset = load_dataset("hover") Traceback (most recent call last): File "/Users/urikz/anaconda/envs/mentionmemory/lib/python3.7/site-packages/datasets/load.py", line 267, in prepare_module local_path = cached_path(file_path, download_config=download_config) File "/Users/urikz/anaconda/envs/mentionmemory/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 308, in cached_path use_etag=download_config.use_etag, File "/Users/urikz/anaconda/envs/mentionmemory/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 486, in get_from_cache raise FileNotFoundError("Couldn't find file at {}".format(url)) FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/hover/hover.py During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/Users/urikz/anaconda/envs/mentionmemory/lib/python3.7/site-packages/datasets/load.py", line 278, in prepare_module local_path = cached_path(file_path, download_config=download_config) File "/Users/urikz/anaconda/envs/mentionmemory/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 308, in cached_path use_etag=download_config.use_etag, File "/Users/urikz/anaconda/envs/mentionmemory/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 486, in get_from_cache raise FileNotFoundError("Couldn't find file at {}".format(url)) FileNotFoundError: Couldn't find file at https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/hover/hover.py During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/urikz/anaconda/envs/mentionmemory/lib/python3.7/site-packages/datasets/load.py", line 589, in load_dataset path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True File "/Users/urikz/anaconda/envs/mentionmemory/lib/python3.7/site-packages/datasets/load.py", line 282, in prepare_module combined_path, github_file_path, file_path FileNotFoundError: Couldn't find file locally at hover/hover.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/hover/hover.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/hover/hover.py ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1644/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1644/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1643
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1643/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1643/comments
https://api.github.com/repos/huggingface/datasets/issues/1643/events
https://github.com/huggingface/datasets/issues/1643
775,280,046
MDU6SXNzdWU3NzUyODAwNDY=
1,643
Dataset social_bias_frames 404
{ "login": "atemate", "id": 7501517, "node_id": "MDQ6VXNlcjc1MDE1MTc=", "avatar_url": "https://avatars.githubusercontent.com/u/7501517?v=4", "gravatar_id": "", "url": "https://api.github.com/users/atemate", "html_url": "https://github.com/atemate", "followers_url": "https://api.github.com/users/atemate/followers", "following_url": "https://api.github.com/users/atemate/following{/other_user}", "gists_url": "https://api.github.com/users/atemate/gists{/gist_id}", "starred_url": "https://api.github.com/users/atemate/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/atemate/subscriptions", "organizations_url": "https://api.github.com/users/atemate/orgs", "repos_url": "https://api.github.com/users/atemate/repos", "events_url": "https://api.github.com/users/atemate/events{/privacy}", "received_events_url": "https://api.github.com/users/atemate/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I see, master is already fixed in https://github.com/huggingface/datasets/commit/9e058f098a0919efd03a136b9b9c3dec5076f626" ]
1,609,144,534,000
1,609,144,687,000
1,609,144,687,000
NONE
null
null
null
``` >>> from datasets import load_dataset >>> dataset = load_dataset("social_bias_frames") ... Downloading and preparing dataset social_bias_frames/default ... ~/.pyenv/versions/3.7.6/lib/python3.7/site-packages/datasets/utils/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag) 484 ) 485 elif response is not None and response.status_code == 404: --> 486 raise FileNotFoundError("Couldn't find file at {}".format(url)) 487 raise ConnectionError("Couldn't reach {}".format(url)) 488 FileNotFoundError: Couldn't find file at https://homes.cs.washington.edu/~msap/social-bias-frames/SocialBiasFrames_v2.tgz ``` [Here](https://homes.cs.washington.edu/~msap/social-bias-frames/) we find button `Download data` with the correct URL for the data: https://homes.cs.washington.edu/~msap/social-bias-frames/SBIC.v2.tgz
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1643/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1643/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1642
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1642/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1642/comments
https://api.github.com/repos/huggingface/datasets/issues/1642/events
https://github.com/huggingface/datasets/pull/1642
775,159,568
MDExOlB1bGxSZXF1ZXN0NTQ1ODk1MzY1
1,642
Ollie dataset
{ "login": "ontocord", "id": 8900094, "node_id": "MDQ6VXNlcjg5MDAwOTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8900094?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ontocord", "html_url": "https://github.com/ontocord", "followers_url": "https://api.github.com/users/ontocord/followers", "following_url": "https://api.github.com/users/ontocord/following{/other_user}", "gists_url": "https://api.github.com/users/ontocord/gists{/gist_id}", "starred_url": "https://api.github.com/users/ontocord/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ontocord/subscriptions", "organizations_url": "https://api.github.com/users/ontocord/orgs", "repos_url": "https://api.github.com/users/ontocord/repos", "events_url": "https://api.github.com/users/ontocord/events{/privacy}", "received_events_url": "https://api.github.com/users/ontocord/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,123,417,000
1,609,767,325,000
1,609,767,324,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1642", "html_url": "https://github.com/huggingface/datasets/pull/1642", "diff_url": "https://github.com/huggingface/datasets/pull/1642.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1642.patch", "merged_at": 1609767324000 }
This is the dataset used to train the Ollie open information extraction algorithm. It has over 21M sentences. See http://knowitall.github.io/ollie/ for more details.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1642/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1642/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1641
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1641/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1641/comments
https://api.github.com/repos/huggingface/datasets/issues/1641/events
https://github.com/huggingface/datasets/issues/1641
775,110,872
MDU6SXNzdWU3NzUxMTA4NzI=
1,641
muchocine dataset cannot be dowloaded
{ "login": "mrm8488", "id": 3653789, "node_id": "MDQ6VXNlcjM2NTM3ODk=", "avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mrm8488", "html_url": "https://github.com/mrm8488", "followers_url": "https://api.github.com/users/mrm8488/followers", "following_url": "https://api.github.com/users/mrm8488/following{/other_user}", "gists_url": "https://api.github.com/users/mrm8488/gists{/gist_id}", "starred_url": "https://api.github.com/users/mrm8488/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mrm8488/subscriptions", "organizations_url": "https://api.github.com/users/mrm8488/orgs", "repos_url": "https://api.github.com/users/mrm8488/repos", "events_url": "https://api.github.com/users/mrm8488/events{/privacy}", "received_events_url": "https://api.github.com/users/mrm8488/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892913, "node_id": "MDU6TGFiZWwxOTM1ODkyOTEz", "url": "https://api.github.com/repos/huggingface/datasets/labels/wontfix", "name": "wontfix", "color": "ffffff", "default": true, "description": "This will not be worked on" }, { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
null
[]
null
[ "I have encountered the same error with `v1.0.1` and `v1.0.2` on both Windows and Linux environments. However, cloning the repo and using the path to the dataset's root directory worked for me. Even after having the dataset cached - passing the path is the only way (for now) to load the dataset.\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"squad\") # Works\r\ndataset = load_dataset(\"code_search_net\", \"python\") # Error\r\ndataset = load_dataset(\"covid_qa_deepset\") # Error\r\n\r\npath = \"/huggingface/datasets/datasets/{}/\"\r\ndataset = load_dataset(path.format(\"code_search_net\"), \"python\") # Works\r\ndataset = load_dataset(path.format(\"covid_qa_deepset\")) # Works\r\n```\r\n\r\n", "Hi @mrm8488 and @amoux!\r\n The datasets you are trying to load have been added to the library during the community sprint for v2 last month. They will be available with the v2 release!\r\nFor now, there are still a couple of solutions to load the datasets:\r\n1. As suggested by @amoux, you can clone the git repo and pass the local path to the script\r\n2. You can also install the latest (master) version of `datasets` using pip: `pip install git+https://github.com/huggingface/datasets.git@master`", "If you don't want to clone entire `datasets` repo, just download the `muchocine` directory and pass the local path to the directory. Cheers!", "Muchocine was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `muchocine` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"muchocine\", split=\"train\")\r\n```", "Thanks @lhoestq " ]
1,609,104,388,000
1,627,967,249,000
1,627,967,249,000
NONE
null
null
null
```python --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) /usr/local/lib/python3.6/dist-packages/datasets/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs) 267 try: --> 268 local_path = cached_path(file_path, download_config=download_config) 269 except FileNotFoundError: 7 frames FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/1.0.2/datasets/muchocine/muchocine.py During handling of the above exception, another exception occurred: FileNotFoundError Traceback (most recent call last) FileNotFoundError: Couldn't find file at https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/muchocine/muchocine.py During handling of the above exception, another exception occurred: FileNotFoundError Traceback (most recent call last) /usr/local/lib/python3.6/dist-packages/datasets/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs) 281 raise FileNotFoundError( 282 "Couldn't find file locally at {}, or remotely at {} or {}".format( --> 283 combined_path, github_file_path, file_path 284 ) 285 ) FileNotFoundError: Couldn't find file locally at muchocine/muchocine.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.0.2/datasets/muchocine/muchocine.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/muchocine/muchocine.py ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1641/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1641/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1640
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1640/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1640/comments
https://api.github.com/repos/huggingface/datasets/issues/1640/events
https://github.com/huggingface/datasets/pull/1640
774,921,836
MDExOlB1bGxSZXF1ZXN0NTQ1NzI2NzY4
1,640
Fix "'BertTokenizerFast' object has no attribute 'max_len'"
{ "login": "mflis", "id": 15031715, "node_id": "MDQ6VXNlcjE1MDMxNzE1", "avatar_url": "https://avatars.githubusercontent.com/u/15031715?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mflis", "html_url": "https://github.com/mflis", "followers_url": "https://api.github.com/users/mflis/followers", "following_url": "https://api.github.com/users/mflis/following{/other_user}", "gists_url": "https://api.github.com/users/mflis/gists{/gist_id}", "starred_url": "https://api.github.com/users/mflis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mflis/subscriptions", "organizations_url": "https://api.github.com/users/mflis/orgs", "repos_url": "https://api.github.com/users/mflis/repos", "events_url": "https://api.github.com/users/mflis/events{/privacy}", "received_events_url": "https://api.github.com/users/mflis/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,609,010,741,000
1,609,176,395,000
1,609,176,395,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1640", "html_url": "https://github.com/huggingface/datasets/pull/1640", "diff_url": "https://github.com/huggingface/datasets/pull/1640.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1640.patch", "merged_at": 1609176395000 }
Tensorflow 2.3.0 gives: FutureWarning: The `max_len` attribute has been deprecated and will be removed in a future version, use `model_max_length` instead. Tensorflow 2.4.0 gives: AttributeError 'BertTokenizerFast' object has no attribute 'max_len'
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1640/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1640/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1639
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1639/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1639/comments
https://api.github.com/repos/huggingface/datasets/issues/1639/events
https://github.com/huggingface/datasets/issues/1639
774,903,472
MDU6SXNzdWU3NzQ5MDM0NzI=
1,639
bug with sst2 in glue
{ "login": "ghost", "id": 10137, "node_id": "MDQ6VXNlcjEwMTM3", "avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghost", "html_url": "https://github.com/ghost", "followers_url": "https://api.github.com/users/ghost/followers", "following_url": "https://api.github.com/users/ghost/following{/other_user}", "gists_url": "https://api.github.com/users/ghost/gists{/gist_id}", "starred_url": "https://api.github.com/users/ghost/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ghost/subscriptions", "organizations_url": "https://api.github.com/users/ghost/orgs", "repos_url": "https://api.github.com/users/ghost/repos", "events_url": "https://api.github.com/users/ghost/events{/privacy}", "received_events_url": "https://api.github.com/users/ghost/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Maybe you can use nltk's treebank detokenizer ?\r\n```python\r\nfrom nltk.tokenize.treebank import TreebankWordDetokenizer\r\n\r\nTreebankWordDetokenizer().detokenize(\"it 's a charming and often affecting journey . \".split())\r\n# \"it's a charming and often affecting journey.\"\r\n```", "I am looking for alternative file URL here instead of adding extra processing code: https://github.com/huggingface/datasets/blob/171f2bba9dd8b92006b13cf076a5bf31d67d3e69/datasets/glue/glue.py#L174", "I don't know if there exists a detokenized version somewhere. Even the version on kaggle is tokenized" ]
1,609,001,843,000
1,630,076,603,000
null
NONE
null
null
null
Hi I am getting very low accuracy on SST2 I investigate this and observe that for this dataset sentences are tokenized, while this is correct for the other datasets in GLUE, please see below. Is there any alternatives I could get untokenized sentences? I am unfortunately under time pressure to report some results on this dataset. thank you for your help. @lhoestq ``` >>> a = datasets.load_dataset('glue', 'sst2', split="validation", script_version="master") Reusing dataset glue (/julia/datasets/glue/sst2/1.0.0/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4) >>> a[:10] {'idx': [0, 1, 2, 3, 4, 5, 6, 7, 8, 9], 'label': [1, 0, 1, 1, 0, 1, 0, 0, 1, 0], 'sentence': ["it 's a charming and often affecting journey . ", 'unflinchingly bleak and desperate ', 'allows us to hope that nolan is poised to embark a major career as a commercial yet inventive filmmaker . ', "the acting , costumes , music , cinematography and sound are all astounding given the production 's austere locales . ", "it 's slow -- very , very slow . ", 'although laced with humor and a few fanciful touches , the film is a refreshingly serious look at young women . ', 'a sometimes tedious film . ', "or doing last year 's taxes with your ex-wife . ", "you do n't have to know about music to appreciate the film 's easygoing blend of comedy and romance . ", "in exactly 89 minutes , most of which passed as slowly as if i 'd been sitting naked on an igloo , formula 51 sank from quirky to jerky to utter turkey . "]} ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1639/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1639/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1638
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1638/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1638/comments
https://api.github.com/repos/huggingface/datasets/issues/1638/events
https://github.com/huggingface/datasets/pull/1638
774,869,184
MDExOlB1bGxSZXF1ZXN0NTQ1Njg5ODQ5
1,638
Add id_puisi dataset
{ "login": "ilhamfp", "id": 31740013, "node_id": "MDQ6VXNlcjMxNzQwMDEz", "avatar_url": "https://avatars.githubusercontent.com/u/31740013?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ilhamfp", "html_url": "https://github.com/ilhamfp", "followers_url": "https://api.github.com/users/ilhamfp/followers", "following_url": "https://api.github.com/users/ilhamfp/following{/other_user}", "gists_url": "https://api.github.com/users/ilhamfp/gists{/gist_id}", "starred_url": "https://api.github.com/users/ilhamfp/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ilhamfp/subscriptions", "organizations_url": "https://api.github.com/users/ilhamfp/orgs", "repos_url": "https://api.github.com/users/ilhamfp/repos", "events_url": "https://api.github.com/users/ilhamfp/events{/privacy}", "received_events_url": "https://api.github.com/users/ilhamfp/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,986,515,000
1,609,346,057,000
1,609,346,057,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1638", "html_url": "https://github.com/huggingface/datasets/pull/1638", "diff_url": "https://github.com/huggingface/datasets/pull/1638.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1638.patch", "merged_at": 1609346057000 }
Puisi (poem) is an Indonesian poetic form. The dataset contains 7223 Indonesian puisi with its title and author. :)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1638/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1638/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1637
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1637/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1637/comments
https://api.github.com/repos/huggingface/datasets/issues/1637/events
https://github.com/huggingface/datasets/pull/1637
774,710,014
MDExOlB1bGxSZXF1ZXN0NTQ1NTc1NTMw
1,637
Added `pn_summary` dataset
{ "login": "m3hrdadfi", "id": 2601833, "node_id": "MDQ6VXNlcjI2MDE4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/2601833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/m3hrdadfi", "html_url": "https://github.com/m3hrdadfi", "followers_url": "https://api.github.com/users/m3hrdadfi/followers", "following_url": "https://api.github.com/users/m3hrdadfi/following{/other_user}", "gists_url": "https://api.github.com/users/m3hrdadfi/gists{/gist_id}", "starred_url": "https://api.github.com/users/m3hrdadfi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/m3hrdadfi/subscriptions", "organizations_url": "https://api.github.com/users/m3hrdadfi/orgs", "repos_url": "https://api.github.com/users/m3hrdadfi/repos", "events_url": "https://api.github.com/users/m3hrdadfi/events{/privacy}", "received_events_url": "https://api.github.com/users/m3hrdadfi/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "As always, I got stuck in the correct order of imports πŸ˜…\r\n@lhoestq, It's finished!", "@lhoestq, It's done! Is there anything else that needs changes?" ]
1,608,894,084,000
1,609,767,799,000
1,609,767,799,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1637", "html_url": "https://github.com/huggingface/datasets/pull/1637", "diff_url": "https://github.com/huggingface/datasets/pull/1637.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1637.patch", "merged_at": 1609767799000 }
#1635 You did a great job with the fluent procedure regarding adding a dataset. I took the chance to add the dataset on my own. Thank you for your awesome job, and I hope this dataset found the researchers happy, specifically those interested in Persian Language (Farsi)!
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1637/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1637/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1636
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1636/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1636/comments
https://api.github.com/repos/huggingface/datasets/issues/1636/events
https://github.com/huggingface/datasets/issues/1636
774,574,378
MDU6SXNzdWU3NzQ1NzQzNzg=
1,636
winogrande cannot be dowloaded
{ "login": "ghost", "id": 10137, "node_id": "MDQ6VXNlcjEwMTM3", "avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghost", "html_url": "https://github.com/ghost", "followers_url": "https://api.github.com/users/ghost/followers", "following_url": "https://api.github.com/users/ghost/following{/other_user}", "gists_url": "https://api.github.com/users/ghost/gists{/gist_id}", "starred_url": "https://api.github.com/users/ghost/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ghost/subscriptions", "organizations_url": "https://api.github.com/users/ghost/orgs", "repos_url": "https://api.github.com/users/ghost/repos", "events_url": "https://api.github.com/users/ghost/events{/privacy}", "received_events_url": "https://api.github.com/users/ghost/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "I have same issue for other datasets (`myanmar_news` in my case).\r\n\r\nA version of `datasets` runs correctly on my local machine (**without GPU**) which looking for the dataset at \r\n```\r\nhttps://raw.githubusercontent.com/huggingface/datasets/master/datasets/myanmar_news/myanmar_news.py\r\n```\r\n\r\nMeanwhile, other version runs on Colab (**with GPU**) failed to download the dataset. It try to find the dataset at `1.1.3` instead of `master` . If I disable GPU on my Colab, the code can load the dataset without any problem.\r\n\r\nMaybe there is some version missmatch with the GPU and CPU version of code for these datasets?", "It looks like they're two different issues\r\n\r\n----------\r\n\r\nFirst for `myanmar_news`: \r\n\r\nIt must come from the way you installed `datasets`.\r\nIf you install `datasets` from source, then the `myanmar_news` script will be loaded from `master`.\r\nHowever if you install from `pip` it will get it using the version of the lib (here `1.1.3`) and `myanmar_news` is not available in `1.1.3`.\r\n\r\nThe difference between your GPU and CPU executions must be the environment, one seems to have installed `datasets` from source and not the other.\r\n\r\n----------\r\n\r\nThen for `winogrande`:\r\n\r\nThe errors says that the url https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/winogrande/winogrande.py is not reachable.\r\nHowever it works fine on my side.\r\n\r\nDoes your machine have an internet connection ? Are connections to github blocked by some sort of proxy ?\r\nCan you also try again in case github had issues when you tried the first time ?\r\n" ]
1,608,848,902,000
1,609,163,629,000
null
NONE
null
null
null
Hi, I am getting this error when trying to run the codes on the cloud. Thank you for any suggestion and help on this @lhoestq ``` File "./finetune_trainer.py", line 318, in <module> main() File "./finetune_trainer.py", line 148, in main for task in data_args.tasks] File "./finetune_trainer.py", line 148, in <listcomp> for task in data_args.tasks] File "/workdir/seq2seq/data/tasks.py", line 65, in get_dataset dataset = self.load_dataset(split=split) File "/workdir/seq2seq/data/tasks.py", line 466, in load_dataset return datasets.load_dataset('winogrande', 'winogrande_l', split=split) File "/usr/local/lib/python3.6/dist-packages/datasets/load.py", line 589, in load_dataset path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True File "/usr/local/lib/python3.6/dist-packages/datasets/load.py", line 267, in prepare_module local_path = cached_path(file_path, download_config=download_config) File "/usr/local/lib/python3.6/dist-packages/datasets/utils/file_utils.py", line 308, in cached_path use_etag=download_config.use_etag, File "/usr/local/lib/python3.6/dist-packages/datasets/utils/file_utils.py", line 487, in get_from_cache raise ConnectionError("Couldn't reach {}".format(url)) ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/winogrande/winogrande.py yo/0 I1224 14:17:46.419031 31226 main shadow.py:122 > Traceback (most recent call last): File "/usr/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/usr/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/usr/local/lib/python3.6/dist-packages/torch/distributed/launch.py", line 260, in <module> main() File "/usr/local/lib/python3.6/dist-packages/torch/distributed/launch.py", line 256, in main cmd=cmd) ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1636/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1636/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1635
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1635/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1635/comments
https://api.github.com/repos/huggingface/datasets/issues/1635/events
https://github.com/huggingface/datasets/issues/1635
774,524,492
MDU6SXNzdWU3NzQ1MjQ0OTI=
1,635
Persian Abstractive/Extractive Text Summarization
{ "login": "m3hrdadfi", "id": 2601833, "node_id": "MDQ6VXNlcjI2MDE4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/2601833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/m3hrdadfi", "html_url": "https://github.com/m3hrdadfi", "followers_url": "https://api.github.com/users/m3hrdadfi/followers", "following_url": "https://api.github.com/users/m3hrdadfi/following{/other_user}", "gists_url": "https://api.github.com/users/m3hrdadfi/gists{/gist_id}", "starred_url": "https://api.github.com/users/m3hrdadfi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/m3hrdadfi/subscriptions", "organizations_url": "https://api.github.com/users/m3hrdadfi/orgs", "repos_url": "https://api.github.com/users/m3hrdadfi/repos", "events_url": "https://api.github.com/users/m3hrdadfi/events{/privacy}", "received_events_url": "https://api.github.com/users/m3hrdadfi/received_events", "type": "User", "site_admin": false }
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,608,832,032,000
1,609,773,064,000
1,609,773,064,000
CONTRIBUTOR
null
null
null
Assembling datasets tailored to different tasks and languages is a precious target. This would be great to have this dataset included. ## Adding a Dataset - **Name:** *pn-summary* - **Description:** *A well-structured summarization dataset for the Persian language consists of 93,207 records. It is prepared for Abstractive/Extractive tasks (like cnn_dailymail for English). It can also be used in other scopes like Text Generation, Title Generation, and News Category Classification.* - **Paper:** *https://arxiv.org/abs/2012.11204* - **Data:** *https://github.com/hooshvare/pn-summary/#download* - **Motivation:** *It is the first Persian abstractive/extractive Text summarization dataset (like cnn_dailymail for English)!* Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1635/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1635/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1634
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1634/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1634/comments
https://api.github.com/repos/huggingface/datasets/issues/1634/events
https://github.com/huggingface/datasets/issues/1634
774,487,934
MDU6SXNzdWU3NzQ0ODc5MzQ=
1,634
Inspecting datasets per category
{ "login": "ghost", "id": 10137, "node_id": "MDQ6VXNlcjEwMTM3", "avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghost", "html_url": "https://github.com/ghost", "followers_url": "https://api.github.com/users/ghost/followers", "following_url": "https://api.github.com/users/ghost/following{/other_user}", "gists_url": "https://api.github.com/users/ghost/gists{/gist_id}", "starred_url": "https://api.github.com/users/ghost/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ghost/subscriptions", "organizations_url": "https://api.github.com/users/ghost/orgs", "repos_url": "https://api.github.com/users/ghost/repos", "events_url": "https://api.github.com/users/ghost/events{/privacy}", "received_events_url": "https://api.github.com/users/ghost/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "That's interesting, can you tell me what you think would be useful to access to inspect a dataset?\r\n\r\nYou can filter them in the hub with the search by the way: https://huggingface.co/datasets have you seen it?", "Hi @thomwolf \r\nthank you, I was not aware of this, I was looking into the data viewer linked into readme page. \r\n\r\nThis is exactly what I was looking for, but this does not work currently, please see the attached \r\nI am selecting to see all nli datasets in english and it retrieves none. thanks\r\n\r\n![5tarDHn9CP6ngeM](https://user-images.githubusercontent.com/53898419/103107612-1509aa80-4638-11eb-85b5-0c995a189969.png)\r\n\r\n\r\n\r\n", "I see 4 results for NLI in English but indeed some are not tagged yet and missing (GLUE), we will focus on that in January (cc @yjernite): https://huggingface.co/datasets?filter=task_ids:natural-language-inference,languages:en" ]
1,608,823,594,000
1,610,098,084,000
null
NONE
null
null
null
Hi Is there a way I could get all NLI datasets/all QA datasets to get some understanding of available datasets per category? this is hard for me to inspect the datasets one by one in the webpage, thanks for the suggestions @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1634/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1634/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1633
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1633/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1633/comments
https://api.github.com/repos/huggingface/datasets/issues/1633/events
https://github.com/huggingface/datasets/issues/1633
774,422,603
MDU6SXNzdWU3NzQ0MjI2MDM=
1,633
social_i_qa wrong format of labels
{ "login": "ghost", "id": 10137, "node_id": "MDQ6VXNlcjEwMTM3", "avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghost", "html_url": "https://github.com/ghost", "followers_url": "https://api.github.com/users/ghost/followers", "following_url": "https://api.github.com/users/ghost/following{/other_user}", "gists_url": "https://api.github.com/users/ghost/gists{/gist_id}", "starred_url": "https://api.github.com/users/ghost/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ghost/subscriptions", "organizations_url": "https://api.github.com/users/ghost/orgs", "repos_url": "https://api.github.com/users/ghost/repos", "events_url": "https://api.github.com/users/ghost/events{/privacy}", "received_events_url": "https://api.github.com/users/ghost/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "@lhoestq, should I raise a PR for this? Just a minor change while reading labels text file", "Sure feel free to open a PR thanks !" ]
1,608,815,514,000
1,609,348,729,000
1,609,348,729,000
NONE
null
null
null
Hi, there is extra "\n" in labels of social_i_qa datasets, no big deal, but I was wondering if you could remove it to make it consistent. so label is 'label': '1\n', not '1' thanks ``` >>> import datasets >>> from datasets import load_dataset >>> dataset = load_dataset( ... 'social_i_qa') cahce dir /julia/cache/datasets Downloading: 4.72kB [00:00, 3.52MB/s] cahce dir /julia/cache/datasets Downloading: 2.19kB [00:00, 1.81MB/s] Using custom data configuration default Reusing dataset social_i_qa (/julia/datasets/social_i_qa/default/0.1.0/4a4190cc2d2482d43416c2167c0c5dccdd769d4482e84893614bd069e5c3ba06) >>> dataset['train'][0] {'answerA': 'like attending', 'answerB': 'like staying home', 'answerC': 'a good friend to have', 'context': 'Cameron decided to have a barbecue and gathered her friends together.', 'label': '1\n', 'question': 'How would Others feel as a result?'} ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1633/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1633/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1632
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1632/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1632/comments
https://api.github.com/repos/huggingface/datasets/issues/1632/events
https://github.com/huggingface/datasets/issues/1632
774,388,625
MDU6SXNzdWU3NzQzODg2MjU=
1,632
SICK dataset
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/followers", "following_url": "https://api.github.com/users/rabeehk/following{/other_user}", "gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}", "starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions", "organizations_url": "https://api.github.com/users/rabeehk/orgs", "repos_url": "https://api.github.com/users/rabeehk/repos", "events_url": "https://api.github.com/users/rabeehk/events{/privacy}", "received_events_url": "https://api.github.com/users/rabeehk/received_events", "type": "User", "site_admin": false }
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,608,813,614,000
1,612,540,165,000
1,612,540,165,000
CONTRIBUTOR
null
null
null
Hi, this would be great to have this dataset included. I might be missing something, but I could not find it in the list of already included datasets. Thank you. ## Adding a Dataset - **Name:** SICK - **Description:** SICK consists of about 10,000 English sentence pairs that include many examples of the lexical, syntactic, and semantic phenomena. - **Paper:** https://www.aclweb.org/anthology/L14-1314/ - **Data:** http://marcobaroni.org/composes/sick.html - **Motivation:** This dataset is well-known in the NLP community used for recognizing entailment between sentences. Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1632/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1632/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1631
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1631/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1631/comments
https://api.github.com/repos/huggingface/datasets/issues/1631/events
https://github.com/huggingface/datasets/pull/1631
774,349,222
MDExOlB1bGxSZXF1ZXN0NTQ1Mjc5MTE2
1,631
Update README.md
{ "login": "savasy", "id": 6584825, "node_id": "MDQ6VXNlcjY1ODQ4MjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6584825?v=4", "gravatar_id": "", "url": "https://api.github.com/users/savasy", "html_url": "https://github.com/savasy", "followers_url": "https://api.github.com/users/savasy/followers", "following_url": "https://api.github.com/users/savasy/following{/other_user}", "gists_url": "https://api.github.com/users/savasy/gists{/gist_id}", "starred_url": "https://api.github.com/users/savasy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/savasy/subscriptions", "organizations_url": "https://api.github.com/users/savasy/orgs", "repos_url": "https://api.github.com/users/savasy/repos", "events_url": "https://api.github.com/users/savasy/events{/privacy}", "received_events_url": "https://api.github.com/users/savasy/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,810,352,000
1,609,176,941,000
1,609,175,764,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1631", "html_url": "https://github.com/huggingface/datasets/pull/1631", "diff_url": "https://github.com/huggingface/datasets/pull/1631.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1631.patch", "merged_at": 1609175764000 }
I made small change for citation
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1631/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1631/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1630
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1630/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1630/comments
https://api.github.com/repos/huggingface/datasets/issues/1630/events
https://github.com/huggingface/datasets/issues/1630
774,332,129
MDU6SXNzdWU3NzQzMzIxMjk=
1,630
Adding UKP Argument Aspect Similarity Corpus
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/followers", "following_url": "https://api.github.com/users/rabeehk/following{/other_user}", "gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}", "starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions", "organizations_url": "https://api.github.com/users/rabeehk/orgs", "repos_url": "https://api.github.com/users/rabeehk/repos", "events_url": "https://api.github.com/users/rabeehk/events{/privacy}", "received_events_url": "https://api.github.com/users/rabeehk/received_events", "type": "User", "site_admin": false }
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "Adding a link to the guide on adding a dataset if someone want to give it a try: https://github.com/huggingface/datasets#add-a-new-dataset-to-the-hub\r\n\r\nwe should add this guide to the issue template @lhoestq ", "thanks @thomwolf , this is added now. The template is correct, sorry my mistake not to include it. " ]
1,608,807,691,000
1,608,809,418,000
null
CONTRIBUTOR
null
null
null
Hi, this would be great to have this dataset included. ## Adding a Dataset - **Name:** UKP Argument Aspect Similarity Corpus - **Description:** The UKP Argument Aspect Similarity Corpus (UKP ASPECT) includes 3,595 sentence pairs over 28 controversial topics. Each sentence pair was annotated via crowdsourcing as either β€œhigh similarity”, β€œsome similarity”, β€œno similarity” or β€œnot related” with respect to the topic. - **Paper:** https://www.aclweb.org/anthology/P19-1054/ - **Data:** https://tudatalib.ulb.tu-darmstadt.de/handle/tudatalib/1998 - **Motivation:** this is one of the datasets currently used frequently in recent adapter papers like https://arxiv.org/pdf/2005.00247.pdf Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md). Thank you
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1630/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1630/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1629
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1629/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1629/comments
https://api.github.com/repos/huggingface/datasets/issues/1629/events
https://github.com/huggingface/datasets/pull/1629
774,255,716
MDExOlB1bGxSZXF1ZXN0NTQ1MjAwNTQ3
1,629
add wongnai_reviews test set labels
{ "login": "cstorm125", "id": 15519308, "node_id": "MDQ6VXNlcjE1NTE5MzA4", "avatar_url": "https://avatars.githubusercontent.com/u/15519308?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cstorm125", "html_url": "https://github.com/cstorm125", "followers_url": "https://api.github.com/users/cstorm125/followers", "following_url": "https://api.github.com/users/cstorm125/following{/other_user}", "gists_url": "https://api.github.com/users/cstorm125/gists{/gist_id}", "starred_url": "https://api.github.com/users/cstorm125/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cstorm125/subscriptions", "organizations_url": "https://api.github.com/users/cstorm125/orgs", "repos_url": "https://api.github.com/users/cstorm125/repos", "events_url": "https://api.github.com/users/cstorm125/events{/privacy}", "received_events_url": "https://api.github.com/users/cstorm125/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,796,951,000
1,609,176,219,000
1,609,176,219,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1629", "html_url": "https://github.com/huggingface/datasets/pull/1629", "diff_url": "https://github.com/huggingface/datasets/pull/1629.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1629.patch", "merged_at": 1609176219000 }
- add test set labels provided by @ekapolc - refactor `star_rating` to a `datasets.features.ClassLabel` field
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1629/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1629/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1628
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1628/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1628/comments
https://api.github.com/repos/huggingface/datasets/issues/1628/events
https://github.com/huggingface/datasets/pull/1628
774,091,411
MDExOlB1bGxSZXF1ZXN0NTQ1MDY5NTAy
1,628
made suggested changes to hate-speech-and-offensive-language
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,765,932,000
1,609,150,280,000
1,609,150,280,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1628", "html_url": "https://github.com/huggingface/datasets/pull/1628", "diff_url": "https://github.com/huggingface/datasets/pull/1628.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1628.patch", "merged_at": 1609150280000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1628/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1628/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1627
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1627/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1627/comments
https://api.github.com/repos/huggingface/datasets/issues/1627/events
https://github.com/huggingface/datasets/issues/1627
773,960,255
MDU6SXNzdWU3NzM5NjAyNTU=
1,627
`Dataset.map` disable progress bar
{ "login": "Nickil21", "id": 8767964, "node_id": "MDQ6VXNlcjg3Njc5NjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8767964?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Nickil21", "html_url": "https://github.com/Nickil21", "followers_url": "https://api.github.com/users/Nickil21/followers", "following_url": "https://api.github.com/users/Nickil21/following{/other_user}", "gists_url": "https://api.github.com/users/Nickil21/gists{/gist_id}", "starred_url": "https://api.github.com/users/Nickil21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nickil21/subscriptions", "organizations_url": "https://api.github.com/users/Nickil21/orgs", "repos_url": "https://api.github.com/users/Nickil21/repos", "events_url": "https://api.github.com/users/Nickil21/events{/privacy}", "received_events_url": "https://api.github.com/users/Nickil21/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Progress bar can be disabled like this:\r\n```python\r\nfrom datasets.utils.logging import set_verbosity_error\r\nset_verbosity_error()\r\n```\r\n\r\nThere is this line in `Dataset.map`:\r\n```python\r\nnot_verbose = bool(logger.getEffectiveLevel() > WARNING)\r\n```\r\n\r\nSo any logging level higher than `WARNING` turns off the progress bar." ]
1,608,746,022,000
1,609,012,656,000
1,609,012,637,000
NONE
null
null
null
I can't find anything to turn off the `tqdm` progress bars while running a preprocessing function using `Dataset.map`. I want to do akin to `disable_tqdm=True` in the case of `transformers`. Is there something like that?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1627/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1627/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1626
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1626/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1626/comments
https://api.github.com/repos/huggingface/datasets/issues/1626/events
https://github.com/huggingface/datasets/pull/1626
773,840,368
MDExOlB1bGxSZXF1ZXN0NTQ0ODYxMDE4
1,626
Fix dataset_dict.shuffle with single seed
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,734,016,000
1,609,754,404,000
1,609,754,403,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1626", "html_url": "https://github.com/huggingface/datasets/pull/1626", "diff_url": "https://github.com/huggingface/datasets/pull/1626.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1626.patch", "merged_at": 1609754403000 }
Fix #1610 I added support for single integer used in `DatasetDict.shuffle`. Previously only a dictionary of seed was allowed. Moreover I added the missing `seed` parameter. Previously only `seeds` was allowed.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1626/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1626/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1625
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1625/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1625/comments
https://api.github.com/repos/huggingface/datasets/issues/1625/events
https://github.com/huggingface/datasets/pull/1625
773,771,596
MDExOlB1bGxSZXF1ZXN0NTQ0Nzk4MDM1
1,625
Fixed bug in the shape property
{ "login": "noaonoszko", "id": 47183162, "node_id": "MDQ6VXNlcjQ3MTgzMTYy", "avatar_url": "https://avatars.githubusercontent.com/u/47183162?v=4", "gravatar_id": "", "url": "https://api.github.com/users/noaonoszko", "html_url": "https://github.com/noaonoszko", "followers_url": "https://api.github.com/users/noaonoszko/followers", "following_url": "https://api.github.com/users/noaonoszko/following{/other_user}", "gists_url": "https://api.github.com/users/noaonoszko/gists{/gist_id}", "starred_url": "https://api.github.com/users/noaonoszko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/noaonoszko/subscriptions", "organizations_url": "https://api.github.com/users/noaonoszko/orgs", "repos_url": "https://api.github.com/users/noaonoszko/repos", "events_url": "https://api.github.com/users/noaonoszko/events{/privacy}", "received_events_url": "https://api.github.com/users/noaonoszko/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,730,401,000
1,609,629,772,000
1,608,732,793,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1625", "html_url": "https://github.com/huggingface/datasets/pull/1625", "diff_url": "https://github.com/huggingface/datasets/pull/1625.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1625.patch", "merged_at": 1608732793000 }
Fix to the bug reported in issue #1622. Just replaced `return tuple(self._indices.num_rows, self._data.num_columns)` by `return (self._indices.num_rows, self._data.num_columns)`.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1625/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1625/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1624
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1624/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1624/comments
https://api.github.com/repos/huggingface/datasets/issues/1624/events
https://github.com/huggingface/datasets/issues/1624
773,669,700
MDU6SXNzdWU3NzM2Njk3MDA=
1,624
Cannot download ade_corpus_v2
{ "login": "him1411", "id": 20259310, "node_id": "MDQ6VXNlcjIwMjU5MzEw", "avatar_url": "https://avatars.githubusercontent.com/u/20259310?v=4", "gravatar_id": "", "url": "https://api.github.com/users/him1411", "html_url": "https://github.com/him1411", "followers_url": "https://api.github.com/users/him1411/followers", "following_url": "https://api.github.com/users/him1411/following{/other_user}", "gists_url": "https://api.github.com/users/him1411/gists{/gist_id}", "starred_url": "https://api.github.com/users/him1411/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/him1411/subscriptions", "organizations_url": "https://api.github.com/users/him1411/orgs", "repos_url": "https://api.github.com/users/him1411/repos", "events_url": "https://api.github.com/users/him1411/events{/privacy}", "received_events_url": "https://api.github.com/users/him1411/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi @him1411, the dataset you are trying to load has been added during the community sprint and has not been released yet. It will be available with the v2 of `datasets`.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of `datasets` using pip:\r\n`pip install git+https://github.com/huggingface/datasets.git@master`", "`ade_corpus_v2` was added recently, that's why it wasn't available yet.\r\n\r\nTo load it you can just update `datasets`\r\n```\r\npip install --upgrade datasets\r\n```\r\n\r\nand then you can load `ade_corpus_v2` with\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"ade_corpus_v2\", \"Ade_corpos_v2_drug_ade_relation\")\r\n```\r\n\r\n(looks like there is a typo in the configuration name, we'll fix it for the v2.0 release of `datasets` soon)" ]
1,608,721,094,000
1,627,967,334,000
1,627,967,334,000
NONE
null
null
null
I tried this to get the dataset following this url : https://huggingface.co/datasets/ade_corpus_v2 but received this error : `Traceback (most recent call last): File "/opt/anaconda3/lib/python3.7/site-packages/datasets/load.py", line 267, in prepare_module local_path = cached_path(file_path, download_config=download_config) File "/opt/anaconda3/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 308, in cached_path use_etag=download_config.use_etag, File "/opt/anaconda3/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 486, in get_from_cache raise FileNotFoundError("Couldn't find file at {}".format(url)) FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/ade_corpus_v2/ade_corpus_v2.py During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/opt/anaconda3/lib/python3.7/site-packages/datasets/load.py", line 278, in prepare_module local_path = cached_path(file_path, download_config=download_config) File "/opt/anaconda3/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 308, in cached_path use_etag=download_config.use_etag, File "/opt/anaconda3/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 486, in get_from_cache raise FileNotFoundError("Couldn't find file at {}".format(url)) FileNotFoundError: Couldn't find file at https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/ade_corpus_v2/ade_corpus_v2.py During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/opt/anaconda3/lib/python3.7/site-packages/datasets/load.py", line 589, in load_dataset path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True File "/opt/anaconda3/lib/python3.7/site-packages/datasets/load.py", line 282, in prepare_module combined_path, github_file_path, file_path FileNotFoundError: Couldn't find file locally at ade_corpus_v2/ade_corpus_v2.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/ade_corpus_v2/ade_corpus_v2.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/ade_corpus_v2/ade_corpus_v2.py`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1624/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1624/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1623
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1623/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1623/comments
https://api.github.com/repos/huggingface/datasets/issues/1623/events
https://github.com/huggingface/datasets/pull/1623
772,950,710
MDExOlB1bGxSZXF1ZXN0NTQ0MTI2ODQ4
1,623
Add CLIMATE-FEVER dataset
{ "login": "tdiggelm", "id": 1658969, "node_id": "MDQ6VXNlcjE2NTg5Njk=", "avatar_url": "https://avatars.githubusercontent.com/u/1658969?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tdiggelm", "html_url": "https://github.com/tdiggelm", "followers_url": "https://api.github.com/users/tdiggelm/followers", "following_url": "https://api.github.com/users/tdiggelm/following{/other_user}", "gists_url": "https://api.github.com/users/tdiggelm/gists{/gist_id}", "starred_url": "https://api.github.com/users/tdiggelm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tdiggelm/subscriptions", "organizations_url": "https://api.github.com/users/tdiggelm/orgs", "repos_url": "https://api.github.com/users/tdiggelm/repos", "events_url": "https://api.github.com/users/tdiggelm/events{/privacy}", "received_events_url": "https://api.github.com/users/tdiggelm/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Thank you @lhoestq for your comments! πŸ˜„ I added your suggested changes, ran the tests and regenerated `dataset_infos.json` and `dummy_data`." ]
1,608,644,045,000
1,608,659,633,000
1,608,659,633,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1623", "html_url": "https://github.com/huggingface/datasets/pull/1623", "diff_url": "https://github.com/huggingface/datasets/pull/1623.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1623.patch", "merged_at": 1608659633000 }
As suggested by @SBrandeis , fresh PR that adds CLIMATE-FEVER. Replaces PR #1579. --- A dataset adopting the FEVER methodology that consists of 1,535 real-world claims regarding climate-change collected on the internet. Each claim is accompanied by five manually annotated evidence sentences retrieved from the English Wikipedia that support, refute or do not give enough information to validate the claim totalling in 7,675 claim-evidence pairs. The dataset features challenging claims that relate multiple facets and disputed cases of claims where both supporting and refuting evidence are present. More information can be found at: * Homepage: http://climatefever.ai * Paper: https://arxiv.org/abs/2012.00614
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1623/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1623/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1622
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1622/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1622/comments
https://api.github.com/repos/huggingface/datasets/issues/1622/events
https://github.com/huggingface/datasets/issues/1622
772,940,768
MDU6SXNzdWU3NzI5NDA3Njg=
1,622
Can't call shape on the output of select()
{ "login": "noaonoszko", "id": 47183162, "node_id": "MDQ6VXNlcjQ3MTgzMTYy", "avatar_url": "https://avatars.githubusercontent.com/u/47183162?v=4", "gravatar_id": "", "url": "https://api.github.com/users/noaonoszko", "html_url": "https://github.com/noaonoszko", "followers_url": "https://api.github.com/users/noaonoszko/followers", "following_url": "https://api.github.com/users/noaonoszko/following{/other_user}", "gists_url": "https://api.github.com/users/noaonoszko/gists{/gist_id}", "starred_url": "https://api.github.com/users/noaonoszko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/noaonoszko/subscriptions", "organizations_url": "https://api.github.com/users/noaonoszko/orgs", "repos_url": "https://api.github.com/users/noaonoszko/repos", "events_url": "https://api.github.com/users/noaonoszko/events{/privacy}", "received_events_url": "https://api.github.com/users/noaonoszko/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Indeed that's a typo, do you want to open a PR to fix it?", "Yes, created a PR" ]
1,608,643,120,000
1,608,730,633,000
1,608,730,632,000
CONTRIBUTOR
null
null
null
I get the error `TypeError: tuple expected at most 1 argument, got 2` when calling `shape` on the output of `select()`. It's line 531 in shape in arrow_dataset.py that causes the problem: ``return tuple(self._indices.num_rows, self._data.num_columns)`` This makes sense, since `tuple(num1, num2)` is not a valid call. Full code to reproduce: ```python dataset = load_dataset("cnn_dailymail", "3.0.0") train_set = dataset["train"] t = train_set.select(range(10)) print(t.shape)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1622/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1622/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1621
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1621/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1621/comments
https://api.github.com/repos/huggingface/datasets/issues/1621/events
https://github.com/huggingface/datasets/pull/1621
772,940,417
MDExOlB1bGxSZXF1ZXN0NTQ0MTE4MTAz
1,621
updated dutch_social.py for loading jsonl (lines instead of list) files
{ "login": "skyprince999", "id": 9033954, "node_id": "MDQ6VXNlcjkwMzM5NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/9033954?v=4", "gravatar_id": "", "url": "https://api.github.com/users/skyprince999", "html_url": "https://github.com/skyprince999", "followers_url": "https://api.github.com/users/skyprince999/followers", "following_url": "https://api.github.com/users/skyprince999/following{/other_user}", "gists_url": "https://api.github.com/users/skyprince999/gists{/gist_id}", "starred_url": "https://api.github.com/users/skyprince999/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/skyprince999/subscriptions", "organizations_url": "https://api.github.com/users/skyprince999/orgs", "repos_url": "https://api.github.com/users/skyprince999/repos", "events_url": "https://api.github.com/users/skyprince999/events{/privacy}", "received_events_url": "https://api.github.com/users/skyprince999/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,643,091,000
1,608,724,311,000
1,608,724,311,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1621", "html_url": "https://github.com/huggingface/datasets/pull/1621", "diff_url": "https://github.com/huggingface/datasets/pull/1621.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1621.patch", "merged_at": 1608724311000 }
the data_loader is modified to load files on the fly. Earlier it was reading the entire file and then processing the records Pls refer to previous PR #1321
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1621/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1621/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1620
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1620/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1620/comments
https://api.github.com/repos/huggingface/datasets/issues/1620/events
https://github.com/huggingface/datasets/pull/1620
772,620,056
MDExOlB1bGxSZXF1ZXN0NTQzODUxNTY3
1,620
Adding myPOS2017 dataset
{ "login": "hungluumfc", "id": 69781878, "node_id": "MDQ6VXNlcjY5NzgxODc4", "avatar_url": "https://avatars.githubusercontent.com/u/69781878?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hungluumfc", "html_url": "https://github.com/hungluumfc", "followers_url": "https://api.github.com/users/hungluumfc/followers", "following_url": "https://api.github.com/users/hungluumfc/following{/other_user}", "gists_url": "https://api.github.com/users/hungluumfc/gists{/gist_id}", "starred_url": "https://api.github.com/users/hungluumfc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hungluumfc/subscriptions", "organizations_url": "https://api.github.com/users/hungluumfc/orgs", "repos_url": "https://api.github.com/users/hungluumfc/repos", "events_url": "https://api.github.com/users/hungluumfc/events{/privacy}", "received_events_url": "https://api.github.com/users/hungluumfc/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "I've updated the code and Readme to reflect your comments.\r\nThank you very much,", "looks like this PR includes changes about many other files than the ones for myPOS2017\r\n\r\nCould you open another branch and another PR please ?\r\n(or fix this branch)", "Hi @hungluumfc ! Have you had a chance to fix this PR so that it only includes the changes for `mypos` ? \r\n\r\nFeel free to ping me if you have questions or if I can help :) " ]
1,608,609,895,000
1,611,915,817,000
null
NONE
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1620", "html_url": "https://github.com/huggingface/datasets/pull/1620", "diff_url": "https://github.com/huggingface/datasets/pull/1620.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1620.patch", "merged_at": null }
myPOS Corpus (Myanmar Part-of-Speech Corpus) for Myanmar language NLP Research and Developments
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1620/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1620/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1619
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1619/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1619/comments
https://api.github.com/repos/huggingface/datasets/issues/1619/events
https://github.com/huggingface/datasets/pull/1619
772,508,558
MDExOlB1bGxSZXF1ZXN0NTQzNzYyMTUw
1,619
data loader for reading comprehension task
{ "login": "songfeng", "id": 2062185, "node_id": "MDQ6VXNlcjIwNjIxODU=", "avatar_url": "https://avatars.githubusercontent.com/u/2062185?v=4", "gravatar_id": "", "url": "https://api.github.com/users/songfeng", "html_url": "https://github.com/songfeng", "followers_url": "https://api.github.com/users/songfeng/followers", "following_url": "https://api.github.com/users/songfeng/following{/other_user}", "gists_url": "https://api.github.com/users/songfeng/gists{/gist_id}", "starred_url": "https://api.github.com/users/songfeng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/songfeng/subscriptions", "organizations_url": "https://api.github.com/users/songfeng/orgs", "repos_url": "https://api.github.com/users/songfeng/repos", "events_url": "https://api.github.com/users/songfeng/events{/privacy}", "received_events_url": "https://api.github.com/users/songfeng/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Thank you for all the feedback! I have updated the dummy data with a zip under 30KB, which needs to include at least one data instance from both document domain and dialog domain. Please let me know if it is still too big. Thanks!", "Thank you again for the feedback! I am not too sure what the preferable style for data instance in readme, but still added my edits. Thanks!" ]
1,608,590,434,000
1,609,151,573,000
1,609,151,573,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1619", "html_url": "https://github.com/huggingface/datasets/pull/1619", "diff_url": "https://github.com/huggingface/datasets/pull/1619.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1619.patch", "merged_at": 1609151573000 }
added doc2dial data loader and dummy data for reading comprehension task.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1619/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1619/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1618
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1618/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1618/comments
https://api.github.com/repos/huggingface/datasets/issues/1618/events
https://github.com/huggingface/datasets/issues/1618
772,248,730
MDU6SXNzdWU3NzIyNDg3MzA=
1,618
Can't filter language:EN on https://huggingface.co/datasets
{ "login": "davidefiocco", "id": 4547987, "node_id": "MDQ6VXNlcjQ1NDc5ODc=", "avatar_url": "https://avatars.githubusercontent.com/u/4547987?v=4", "gravatar_id": "", "url": "https://api.github.com/users/davidefiocco", "html_url": "https://github.com/davidefiocco", "followers_url": "https://api.github.com/users/davidefiocco/followers", "following_url": "https://api.github.com/users/davidefiocco/following{/other_user}", "gists_url": "https://api.github.com/users/davidefiocco/gists{/gist_id}", "starred_url": "https://api.github.com/users/davidefiocco/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davidefiocco/subscriptions", "organizations_url": "https://api.github.com/users/davidefiocco/orgs", "repos_url": "https://api.github.com/users/davidefiocco/repos", "events_url": "https://api.github.com/users/davidefiocco/events{/privacy}", "received_events_url": "https://api.github.com/users/davidefiocco/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "cc'ing @mapmeld ", "Full language list is now deployed to https://huggingface.co/datasets ! Recommend close", "Cool @mapmeld ! My 2 cents (for a next iteration), it would be cool to have a small search widget in the filter dropdown as you have a ton of languages now here! Closing this in the meantime." ]
1,608,564,203,000
1,608,657,420,000
1,608,657,369,000
NONE
null
null
null
When visiting https://huggingface.co/datasets, I don't see an obvious way to filter only English datasets. This is unexpected for me, am I missing something? I'd expect English to be selectable in the language widget. This problem reproduced on Mozilla Firefox and MS Edge: ![screenshot](https://user-images.githubusercontent.com/4547987/102792244-892e1f00-43a8-11eb-9e89-4826ca201a87.png)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1618/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1618/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1617
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1617/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1617/comments
https://api.github.com/repos/huggingface/datasets/issues/1617/events
https://github.com/huggingface/datasets/pull/1617
772,084,764
MDExOlB1bGxSZXF1ZXN0NTQzNDE5MTM5
1,617
cifar10 initial commit
{ "login": "czabo", "id": 75574105, "node_id": "MDQ6VXNlcjc1NTc0MTA1", "avatar_url": "https://avatars.githubusercontent.com/u/75574105?v=4", "gravatar_id": "", "url": "https://api.github.com/users/czabo", "html_url": "https://github.com/czabo", "followers_url": "https://api.github.com/users/czabo/followers", "following_url": "https://api.github.com/users/czabo/following{/other_user}", "gists_url": "https://api.github.com/users/czabo/gists{/gist_id}", "starred_url": "https://api.github.com/users/czabo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/czabo/subscriptions", "organizations_url": "https://api.github.com/users/czabo/orgs", "repos_url": "https://api.github.com/users/czabo/repos", "events_url": "https://api.github.com/users/czabo/events{/privacy}", "received_events_url": "https://api.github.com/users/czabo/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Yee a Computer Vision dataset!", "Yep, the first one ! Thank @czabo " ]
1,608,549,530,000
1,608,632,285,000
1,608,631,888,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1617", "html_url": "https://github.com/huggingface/datasets/pull/1617", "diff_url": "https://github.com/huggingface/datasets/pull/1617.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1617.patch", "merged_at": 1608631888000 }
CIFAR-10 dataset. Didn't add the tagging since there are no vision related tags.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1617/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1617/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1616
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1616/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1616/comments
https://api.github.com/repos/huggingface/datasets/issues/1616/events
https://github.com/huggingface/datasets/pull/1616
772,074,229
MDExOlB1bGxSZXF1ZXN0NTQzNDEwNDc1
1,616
added TurkishMovieSentiment dataset
{ "login": "yavuzKomecoglu", "id": 5150963, "node_id": "MDQ6VXNlcjUxNTA5NjM=", "avatar_url": "https://avatars.githubusercontent.com/u/5150963?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yavuzKomecoglu", "html_url": "https://github.com/yavuzKomecoglu", "followers_url": "https://api.github.com/users/yavuzKomecoglu/followers", "following_url": "https://api.github.com/users/yavuzKomecoglu/following{/other_user}", "gists_url": "https://api.github.com/users/yavuzKomecoglu/gists{/gist_id}", "starred_url": "https://api.github.com/users/yavuzKomecoglu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yavuzKomecoglu/subscriptions", "organizations_url": "https://api.github.com/users/yavuzKomecoglu/orgs", "repos_url": "https://api.github.com/users/yavuzKomecoglu/repos", "events_url": "https://api.github.com/users/yavuzKomecoglu/events{/privacy}", "received_events_url": "https://api.github.com/users/yavuzKomecoglu/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "> I just generated the dataset_infos.json file\r\n> \r\n> Thanks for adding this one !\r\n\r\nThank you very much for your support." ]
1,608,548,596,000
1,608,793,721,000
1,608,742,206,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1616", "html_url": "https://github.com/huggingface/datasets/pull/1616", "diff_url": "https://github.com/huggingface/datasets/pull/1616.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1616.patch", "merged_at": 1608742206000 }
This PR adds the **TurkishMovieSentiment: This dataset contains turkish movie reviews.** - **Homepage:** [https://www.kaggle.com/mustfkeskin/turkish-movie-sentiment-analysis-dataset/tasks](https://www.kaggle.com/mustfkeskin/turkish-movie-sentiment-analysis-dataset/tasks) - **Point of Contact:** [Mustafa Keskin](https://www.linkedin.com/in/mustfkeskin/)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1616/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1616/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1615
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1615/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1615/comments
https://api.github.com/repos/huggingface/datasets/issues/1615/events
https://github.com/huggingface/datasets/issues/1615
771,641,088
MDU6SXNzdWU3NzE2NDEwODg=
1,615
Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`
{ "login": "SapirWeissbuch", "id": 44585792, "node_id": "MDQ6VXNlcjQ0NTg1Nzky", "avatar_url": "https://avatars.githubusercontent.com/u/44585792?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SapirWeissbuch", "html_url": "https://github.com/SapirWeissbuch", "followers_url": "https://api.github.com/users/SapirWeissbuch/followers", "following_url": "https://api.github.com/users/SapirWeissbuch/following{/other_user}", "gists_url": "https://api.github.com/users/SapirWeissbuch/gists{/gist_id}", "starred_url": "https://api.github.com/users/SapirWeissbuch/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SapirWeissbuch/subscriptions", "organizations_url": "https://api.github.com/users/SapirWeissbuch/orgs", "repos_url": "https://api.github.com/users/SapirWeissbuch/repos", "events_url": "https://api.github.com/users/SapirWeissbuch/events{/privacy}", "received_events_url": "https://api.github.com/users/SapirWeissbuch/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Hi @SapirWeissbuch,\r\nWhen you are saying it freezes, at that time it is unzipping the file from the zip file it downloaded. Since it's a very heavy file it'll take some time. It was taking ~11GB after unzipping when it started reading examples for me. Hope that helps!\r\n![Screenshot 2020-12-21 at 23 40 52](https://user-images.githubusercontent.com/19718818/102808355-3b380c00-43e6-11eb-81ab-c31019ae6322.png)\r\n", "Hi @bhavitvyamalik \r\nThanks for the reply!\r\nActually I let it run for 30 minutes before I killed the process. In this time, 30GB were extracted (much more than 11GB), I checked the size of the destination directory.\r\n\r\nWhat version of Datasets are you using?\r\n", "I'm using datasets version: 1.1.3. I think you should drop `cache_dir` and use only\r\n`dataset = datasets.load_dataset(\"trivia_qa\", \"rc\")`\r\n\r\nTried that on colab and it's working there too\r\n![image](https://user-images.githubusercontent.com/19718818/102814269-4db74300-43f0-11eb-8f26-ecfcf4632002.png)\r\n", "Train, Validation, and Test splits contain 138384, 18669, and 17210 samples respectively. It takes some time to read the samples. Even in your colab notebook it was reading the samples before you killed the process. Let me know if it works now!", "Hi, it works on colab but it still doesn't work on my computer, same problem as before - overly large and long extraction process.\r\nI have to use a custom 'cache_dir' because I don't have any space left in my home directory where it is defaulted, maybe this could be the issue?", "I tried running this again - More details of the problem:\r\nCode:\r\n```\r\ndatasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir=\"/path/to/cache\")\r\n```\r\n\r\nThe output:\r\n```\r\nDownloading and preparing dataset trivia_qa/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to path/to/cache/trivia_qa/rc/1.1.0/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... \r\nDownloading: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2.67G/2.67G [03:38<00:00, 12.2MB/s]\r\n\r\n```\r\nThe process continues (no progress bar is visible).\r\nI tried `du -sh .` in `path/to/cache`, and the size keeps increasing, reached 35G before I killed the process.\r\n\r\nGoogle Colab with custom `cache_dir` has same issue.\r\nhttps://colab.research.google.com/drive/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing#scrollTo=2G2O0AeNIXan", "1) You can clear the huggingface folder in your `.cache` directory to use default directory for datasets. Speed of extraction and loading of samples depends a lot on your machine's configurations too.\r\n\r\n2) I tried on colab `dataset = datasets.load_dataset(\"trivia_qa\", \"rc\", cache_dir = \"./datasets\")`. After memory usage reached around 42GB (starting from 32GB used already), the dataset was loaded in the memory. Even Your colab notebook shows \r\n![image](https://user-images.githubusercontent.com/19718818/102852229-c7c4e780-4443-11eb-91d6-bf21024358a3.png)\r\nwhich means it's loaded now.", "Facing the same issue.\r\nI am able to download datasets without `cache_dir`, however, when I specify the `cache_dir`, the process hangs indefinitely after partial download. \r\nTried for `data = load_dataset(\"cnn_dailymail\", \"3.0.0\")`", "Hi @ashutoshml,\r\nI tried this and it worked for me:\r\n`data = load_dataset(\"cnn_dailymail\", \"3.0.0\", cache_dir=\"./dummy\")`\r\n\r\nI'm using datasets==1.8.0. It took around 3-4 mins for dataset to unpack and start loading examples.", "Ok. I waited for 20-30 mins, and it still is stuck.\r\nI am using datasets==1.8.0.\r\n\r\nIs there anyway to check what is happening? like a` --verbose` flag?\r\n\r\n![Screenshot 2021-06-25 at 6 37 43 PM](https://user-images.githubusercontent.com/2375919/123429653-cdfb7280-d5e4-11eb-9fa7-ff295800cc86.png)\r\n" ]
1,608,485,258,000
1,624,626,693,000
null
NONE
null
null
null
Hello, I'm having issue downloading TriviaQA dataset with `load_dataset`. ## Environment info - `datasets` version: 1.1.3 - Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1 - Python version: 3.7.3 ## The code I'm running: ```python import datasets dataset = datasets.load_dataset("trivia_qa", "rc", cache_dir = "./datasets") ``` ## The output: 1. Download begins: ``` Downloading and preparing dataset trivia_qa/rc (download: 2.48 GiB, generated: 14.92 GiB, post-processed: Unknown size, total: 17.40 GiB) to /cs/labs/gabis/sapirweissbuch/tr ivia_qa/rc/1.1.0/e734e28133f4d9a353af322aa52b9f266f6f27cbf2f072690a1694e577546b0d... Downloading: 17%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 446M/2.67G [00:37<04:45, 7.77MB/s] ``` 2. 100% is reached 3. It got stuck here for about an hour, and added additional 30G of data to "./datasets" directory. I killed the process eventually. A similar issue can be observed in Google Colab: https://colab.research.google.com/drive/1nn1Lw02GhfGFylzbS2j6yksGjPo7kkN-?usp=sharing ## Expected behaviour: The dataset "TriviaQA" should be successfully downloaded.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1615/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1615/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1613
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1613/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1613/comments
https://api.github.com/repos/huggingface/datasets/issues/1613/events
https://github.com/huggingface/datasets/pull/1613
771,577,050
MDExOlB1bGxSZXF1ZXN0NTQzMDYwNzEx
1,613
Add id_clickbait
{ "login": "cahya-wirawan", "id": 7669893, "node_id": "MDQ6VXNlcjc2Njk4OTM=", "avatar_url": "https://avatars.githubusercontent.com/u/7669893?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cahya-wirawan", "html_url": "https://github.com/cahya-wirawan", "followers_url": "https://api.github.com/users/cahya-wirawan/followers", "following_url": "https://api.github.com/users/cahya-wirawan/following{/other_user}", "gists_url": "https://api.github.com/users/cahya-wirawan/gists{/gist_id}", "starred_url": "https://api.github.com/users/cahya-wirawan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cahya-wirawan/subscriptions", "organizations_url": "https://api.github.com/users/cahya-wirawan/orgs", "repos_url": "https://api.github.com/users/cahya-wirawan/repos", "events_url": "https://api.github.com/users/cahya-wirawan/events{/privacy}", "received_events_url": "https://api.github.com/users/cahya-wirawan/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,467,089,000
1,608,659,127,000
1,608,659,127,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1613", "html_url": "https://github.com/huggingface/datasets/pull/1613", "diff_url": "https://github.com/huggingface/datasets/pull/1613.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1613.patch", "merged_at": 1608659127000 }
This is the CLICK-ID dataset, a collection of annotated clickbait Indonesian news headlines that was collected from 12 local online news
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1613/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1613/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1612
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1612/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1612/comments
https://api.github.com/repos/huggingface/datasets/issues/1612/events
https://github.com/huggingface/datasets/pull/1612
771,558,160
MDExOlB1bGxSZXF1ZXN0NTQzMDQ3NjQ1
1,612
Adding wiki asp dataset as new PR
{ "login": "katnoria", "id": 7674948, "node_id": "MDQ6VXNlcjc2NzQ5NDg=", "avatar_url": "https://avatars.githubusercontent.com/u/7674948?v=4", "gravatar_id": "", "url": "https://api.github.com/users/katnoria", "html_url": "https://github.com/katnoria", "followers_url": "https://api.github.com/users/katnoria/followers", "following_url": "https://api.github.com/users/katnoria/following{/other_user}", "gists_url": "https://api.github.com/users/katnoria/gists{/gist_id}", "starred_url": "https://api.github.com/users/katnoria/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/katnoria/subscriptions", "organizations_url": "https://api.github.com/users/katnoria/orgs", "repos_url": "https://api.github.com/users/katnoria/repos", "events_url": "https://api.github.com/users/katnoria/events{/privacy}", "received_events_url": "https://api.github.com/users/katnoria/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,459,908,000
1,608,560,013,000
1,608,560,013,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1612", "html_url": "https://github.com/huggingface/datasets/pull/1612", "diff_url": "https://github.com/huggingface/datasets/pull/1612.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1612.patch", "merged_at": 1608560013000 }
Hi @lhoestq, Adding wiki asp as new branch because #1539 has other commits. This version has dummy data for each domain <20/30KB.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1612/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1612/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1611
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1611/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1611/comments
https://api.github.com/repos/huggingface/datasets/issues/1611/events
https://github.com/huggingface/datasets/issues/1611
771,486,456
MDU6SXNzdWU3NzE0ODY0NTY=
1,611
shuffle with torch generator
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url": "https://api.github.com/users/rabeehkarimimahabadi/followers", "following_url": "https://api.github.com/users/rabeehkarimimahabadi/following{/other_user}", "gists_url": "https://api.github.com/users/rabeehkarimimahabadi/gists{/gist_id}", "starred_url": "https://api.github.com/users/rabeehkarimimahabadi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rabeehkarimimahabadi/subscriptions", "organizations_url": "https://api.github.com/users/rabeehkarimimahabadi/orgs", "repos_url": "https://api.github.com/users/rabeehkarimimahabadi/repos", "events_url": "https://api.github.com/users/rabeehkarimimahabadi/events{/privacy}", "received_events_url": "https://api.github.com/users/rabeehkarimimahabadi/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Is there a way one can convert the two generator? not sure overall what alternatives I could have to shuffle the datasets with a torch generator, thanks ", "@lhoestq let me please expalin in more details, maybe you could help me suggesting an alternative to solve the issue for now, I have multiple large datasets using huggingface library, then I need to define a distributed sampler on top of it, for this I need to shard the datasets and give each shard to each core, but before sharding I need to shuffle the dataset, if you are familiar with distributed sampler in pytorch, this needs to be done based on seed+epoch generator to make it consistent across the cores they do it through defining a torch generator, I was wondering if you could tell me how I can shuffle the data for now, I am unfortunately blocked by this and have a limited time left, and I greatly appreciate your help on this. thanks ", "@lhoestq Is there a way I could shuffle the datasets from this library with a custom defined shuffle function? thanks for your help on this. ", "Right now the shuffle method only accepts the `seed` (optional int) or `generator` (optional `np.random.Generator`) parameters.\r\n\r\nHere is a suggestion to shuffle the data using your own shuffle method using `select`.\r\n`select` can be used to re-order the dataset samples or simply pick a few ones if you want.\r\nIt's what is used under the hood when you call `dataset.shuffle`.\r\n\r\nTo use `select` you must have the list of re-ordered indices of your samples.\r\n\r\nLet's say you have a `shuffle` methods that you want to use. Then you can first build your shuffled list of indices:\r\n```python\r\nshuffled_indices = shuffle(range(len(dataset)))\r\n```\r\n\r\nThen you can shuffle your dataset using the shuffled indices with \r\n```python\r\nshuffled_dataset = dataset.select(shuffled_indices)\r\n```\r\n\r\nHope that helps", "thank you @lhoestq thank you very much for responding to my question, this greatly helped me and remove the blocking for continuing my work, thanks. ", "@lhoestq could you confirm the method proposed does not bring the whole data into memory? thanks ", "Yes the dataset is not loaded into memory", "great. thanks a lot." ]
1,608,425,834,000
1,608,574,339,000
null
NONE
null
null
null
Hi I need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I really need to make shuffle work with this generator and I was wondering what I can do about this issue, thanks for your help @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1611/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1611/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1610
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1610/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1610/comments
https://api.github.com/repos/huggingface/datasets/issues/1610/events
https://github.com/huggingface/datasets/issues/1610
771,453,599
MDU6SXNzdWU3NzE0NTM1OTk=
1,610
shuffle does not accept seed
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/followers", "following_url": "https://api.github.com/users/rabeehk/following{/other_user}", "gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}", "starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions", "organizations_url": "https://api.github.com/users/rabeehk/orgs", "repos_url": "https://api.github.com/users/rabeehk/repos", "events_url": "https://api.github.com/users/rabeehk/events{/privacy}", "received_events_url": "https://api.github.com/users/rabeehk/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi, did you check the doc on `shuffle`?\r\nhttps://huggingface.co/docs/datasets/package_reference/main_classes.html?datasets.Dataset.shuffle#datasets.Dataset.shuffle", "Hi Thomas\r\nthanks for reponse, yes, I did checked it, but this does not work for me please see \r\n\r\n```\r\n(internship) rkarimi@italix17:/idiap/user/rkarimi/dev$ python \r\nPython 3.7.9 (default, Aug 31 2020, 12:42:55) \r\n[GCC 7.3.0] :: Anaconda, Inc. on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> import datasets \r\n2020-12-20 01:48:50.766004: W tensorflow/stream_executor/platform/default/dso_loader.cc:60] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudart.so.11.0: cannot open shared object file: No such file or directory\r\n2020-12-20 01:48:50.766029: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\n>>> data = datasets.load_dataset(\"scitail\", \"snli_format\")\r\ncahce dir /idiap/temp/rkarimi/cache_home_1/datasets\r\ncahce dir /idiap/temp/rkarimi/cache_home_1/datasets\r\nReusing dataset scitail (/idiap/temp/rkarimi/cache_home_1/datasets/scitail/snli_format/1.1.0/fd8ccdfc3134ce86eb4ef10ba7f21ee2a125c946e26bb1dd3625fe74f48d3b90)\r\n>>> data.shuffle(seed=2)\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nTypeError: shuffle() got an unexpected keyword argument 'seed'\r\n\r\n```\r\n\r\ndatasets version\r\n`datasets 1.1.2 <pip>\r\n`\r\n", "Thanks for reporting ! \r\n\r\nIndeed it looks like an issue with `suffle` on `DatasetDict`. We're going to fix that.\r\nIn the meantime you can shuffle each split (train, validation, test) separately:\r\n```python\r\nshuffled_train_dataset = data[\"train\"].shuffle(seed=42)\r\n```\r\n" ]
1,608,411,579,000
1,609,754,403,000
1,609,754,403,000
CONTRIBUTOR
null
null
null
Hi I need to shuffle the dataset, but this needs to be based on epoch+seed to be consistent across the cores, when I pass seed to shuffle, this does not accept seed, could you assist me with this? thanks @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1610/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1610/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1609
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1609/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1609/comments
https://api.github.com/repos/huggingface/datasets/issues/1609/events
https://github.com/huggingface/datasets/issues/1609
771,421,881
MDU6SXNzdWU3NzE0MjE4ODE=
1,609
Not able to use 'jigsaw_toxicity_pred' dataset
{ "login": "jassimran", "id": 7424133, "node_id": "MDQ6VXNlcjc0MjQxMzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7424133?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jassimran", "html_url": "https://github.com/jassimran", "followers_url": "https://api.github.com/users/jassimran/followers", "following_url": "https://api.github.com/users/jassimran/following{/other_user}", "gists_url": "https://api.github.com/users/jassimran/gists{/gist_id}", "starred_url": "https://api.github.com/users/jassimran/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jassimran/subscriptions", "organizations_url": "https://api.github.com/users/jassimran/orgs", "repos_url": "https://api.github.com/users/jassimran/repos", "events_url": "https://api.github.com/users/jassimran/events{/privacy}", "received_events_url": "https://api.github.com/users/jassimran/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi @jassimran,\r\nThe `jigsaw_toxicity_pred` dataset has not been released yet, it will be available with version 2 of `datasets`, coming soon.\r\nYou can still access it by installing the master (unreleased) version of datasets directly :\r\n`pip install git+https://github.com/huggingface/datasets.git@master`\r\nPlease let me know if this helps", "Thanks.That works for now." ]
1,608,399,348,000
1,608,655,344,000
1,608,655,343,000
NONE
null
null
null
When trying to use jigsaw_toxicity_pred dataset, like this in a [colab](https://colab.research.google.com/drive/1LwO2A5M2X5dvhkAFYE4D2CUT3WUdWnkn?usp=sharing): ``` from datasets import list_datasets, list_metrics, load_dataset, load_metric ds = load_dataset("jigsaw_toxicity_pred") ``` I see below error: > FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/jigsaw_toxicity_pred/jigsaw_toxicity_pred.py During handling of the above exception, another exception occurred: FileNotFoundError Traceback (most recent call last) FileNotFoundError: Couldn't find file at https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/jigsaw_toxicity_pred/jigsaw_toxicity_pred.py During handling of the above exception, another exception occurred: FileNotFoundError Traceback (most recent call last) /usr/local/lib/python3.6/dist-packages/datasets/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, **download_kwargs) 280 raise FileNotFoundError( 281 "Couldn't find file locally at {}, or remotely at {} or {}".format( --> 282 combined_path, github_file_path, file_path 283 ) 284 ) FileNotFoundError: Couldn't find file locally at jigsaw_toxicity_pred/jigsaw_toxicity_pred.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/jigsaw_toxicity_pred/jigsaw_toxicity_pred.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/jigsaw_toxicity_pred/jigsaw_toxicity_pred.py
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1609/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1609/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1608
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1608/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1608/comments
https://api.github.com/repos/huggingface/datasets/issues/1608/events
https://github.com/huggingface/datasets/pull/1608
771,329,434
MDExOlB1bGxSZXF1ZXN0NTQyODkyMTQ4
1,608
adding ted_talks_iwslt
{ "login": "skyprince999", "id": 9033954, "node_id": "MDQ6VXNlcjkwMzM5NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/9033954?v=4", "gravatar_id": "", "url": "https://api.github.com/users/skyprince999", "html_url": "https://github.com/skyprince999", "followers_url": "https://api.github.com/users/skyprince999/followers", "following_url": "https://api.github.com/users/skyprince999/following{/other_user}", "gists_url": "https://api.github.com/users/skyprince999/gists{/gist_id}", "starred_url": "https://api.github.com/users/skyprince999/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/skyprince999/subscriptions", "organizations_url": "https://api.github.com/users/skyprince999/orgs", "repos_url": "https://api.github.com/users/skyprince999/repos", "events_url": "https://api.github.com/users/skyprince999/events{/privacy}", "received_events_url": "https://api.github.com/users/skyprince999/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Closing this with reference to the new approach #1676 " ]
1,608,363,401,000
1,609,602,252,000
1,609,602,251,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1608", "html_url": "https://github.com/huggingface/datasets/pull/1608", "diff_url": "https://github.com/huggingface/datasets/pull/1608.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1608.patch", "merged_at": null }
UPDATE2: (2nd Jan) Wrote a long writeup on the slack channel. I don't think this approach is correct. Basically this created language pairs (109*108) Running the `pytest `went for more than 40+ hours and it was still running! So working on a different approach, such that the number of configs = number of languages. Will make a new pull request with that. UPDATE: This requires manual download dataset This is a draft version
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1608/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1608/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1607
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1607/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1607/comments
https://api.github.com/repos/huggingface/datasets/issues/1607/events
https://github.com/huggingface/datasets/pull/1607
771,325,852
MDExOlB1bGxSZXF1ZXN0NTQyODg5OTky
1,607
modified tweets hate speech detection
{ "login": "darshan-gandhi", "id": 44197177, "node_id": "MDQ6VXNlcjQ0MTk3MTc3", "avatar_url": "https://avatars.githubusercontent.com/u/44197177?v=4", "gravatar_id": "", "url": "https://api.github.com/users/darshan-gandhi", "html_url": "https://github.com/darshan-gandhi", "followers_url": "https://api.github.com/users/darshan-gandhi/followers", "following_url": "https://api.github.com/users/darshan-gandhi/following{/other_user}", "gists_url": "https://api.github.com/users/darshan-gandhi/gists{/gist_id}", "starred_url": "https://api.github.com/users/darshan-gandhi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/darshan-gandhi/subscriptions", "organizations_url": "https://api.github.com/users/darshan-gandhi/orgs", "repos_url": "https://api.github.com/users/darshan-gandhi/repos", "events_url": "https://api.github.com/users/darshan-gandhi/events{/privacy}", "received_events_url": "https://api.github.com/users/darshan-gandhi/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,362,020,000
1,608,566,928,000
1,608,566,928,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1607", "html_url": "https://github.com/huggingface/datasets/pull/1607", "diff_url": "https://github.com/huggingface/datasets/pull/1607.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1607.patch", "merged_at": 1608566928000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1607/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1607/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1606
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1606/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1606/comments
https://api.github.com/repos/huggingface/datasets/issues/1606/events
https://github.com/huggingface/datasets/pull/1606
771,116,455
MDExOlB1bGxSZXF1ZXN0NTQyNzMwNTEw
1,606
added Semantic Scholar Open Research Corpus
{ "login": "bhavitvyamalik", "id": 19718818, "node_id": "MDQ6VXNlcjE5NzE4ODE4", "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bhavitvyamalik", "html_url": "https://github.com/bhavitvyamalik", "followers_url": "https://api.github.com/users/bhavitvyamalik/followers", "following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}", "gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}", "starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions", "organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs", "repos_url": "https://api.github.com/users/bhavitvyamalik/repos", "events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}", "received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I think we’ll need complete dataset_infos.json to create YAML tags. I ran the script again with 100 files after going through your comments and it was occupying ~16 GB space. So in total it should take ~960GB and I don’t have this much memory available with me. Also, I'll have to download the whole dataset for generating dummy data, right?" ]
1,608,319,284,000
1,612,344,659,000
1,612,344,659,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1606", "html_url": "https://github.com/huggingface/datasets/pull/1606", "diff_url": "https://github.com/huggingface/datasets/pull/1606.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1606.patch", "merged_at": 1612344659000 }
I picked up this dataset [Semantic Scholar Open Research Corpus](https://allenai.org/data/s2orc) but it contains 6000 files to be downloaded. I tried the current code with 100 files and it worked fine (took ~15GB space). For 6000 files it would occupy ~900GB space which I don’t have. Can someone from the HF team with that much of disk space help me with generate dataset_infos and dummy_data?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1606/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1606/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1605
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1605/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1605/comments
https://api.github.com/repos/huggingface/datasets/issues/1605/events
https://github.com/huggingface/datasets/issues/1605
770,979,620
MDU6SXNzdWU3NzA5Nzk2MjA=
1,605
Navigation version breaking
{ "login": "mttk", "id": 3007947, "node_id": "MDQ6VXNlcjMwMDc5NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3007947?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mttk", "html_url": "https://github.com/mttk", "followers_url": "https://api.github.com/users/mttk/followers", "following_url": "https://api.github.com/users/mttk/following{/other_user}", "gists_url": "https://api.github.com/users/mttk/gists{/gist_id}", "starred_url": "https://api.github.com/users/mttk/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mttk/subscriptions", "organizations_url": "https://api.github.com/users/mttk/orgs", "repos_url": "https://api.github.com/users/mttk/repos", "events_url": "https://api.github.com/users/mttk/events{/privacy}", "received_events_url": "https://api.github.com/users/mttk/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[]
1,608,305,784,000
1,608,306,112,000
null
NONE
null
null
null
Hi, when navigating docs (Chrome, Ubuntu) (e.g. on this page: https://huggingface.co/docs/datasets/loading_metrics.html#using-a-custom-metric-script) the version control dropdown has the wrong string displayed as the current version: ![image](https://user-images.githubusercontent.com/3007947/102632187-02cad080-414f-11eb-813b-28f3c8d80def.png) **Edit:** this actually happens _only_ if you open a link to a concrete subsection. IMO, the best way to fix this without getting too deep into the intricacies of retrieving version numbers from the URL would be to change [this](https://github.com/huggingface/datasets/blob/master/docs/source/_static/js/custom.js#L112) line to: ``` let label = (version in versionMapping) ? version : stableVersion ``` which delegates the check to the (already maintained) keys of the version mapping dictionary & should be more robust. There's a similar ternary expression [here](https://github.com/huggingface/datasets/blob/master/docs/source/_static/js/custom.js#L97) which should also fail in this case. I'd also suggest swapping this [block](https://github.com/huggingface/datasets/blob/master/docs/source/_static/js/custom.js#L80-L90) to `string.contains(version) for version in versionMapping` which might be more robust. I'd add a PR myself but I'm by no means competent in JS :) I also have a side question wrt. docs versioning: I'm trying to make docs for a project which are versioned alike to your dropdown versioning. I was wondering how do you handle storage of multiple doc versions on your server? Do you update what `https://huggingface.co/docs/datasets` points to for every stable release & manually create new folders for each released version? So far I'm building & publishing (scping) the docs to the server with a github action which works well for a single version, but would ideally need to reorder the public files triggered on a new release.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1605/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1605/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1604
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1604/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1604/comments
https://api.github.com/repos/huggingface/datasets/issues/1604/events
https://github.com/huggingface/datasets/issues/1604
770,862,112
MDU6SXNzdWU3NzA4NjIxMTI=
1,604
Add tests for the download functions ?
{ "login": "SBrandeis", "id": 33657802, "node_id": "MDQ6VXNlcjMzNjU3ODAy", "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SBrandeis", "html_url": "https://github.com/SBrandeis", "followers_url": "https://api.github.com/users/SBrandeis/followers", "following_url": "https://api.github.com/users/SBrandeis/following{/other_user}", "gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}", "starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions", "organizations_url": "https://api.github.com/users/SBrandeis/orgs", "repos_url": "https://api.github.com/users/SBrandeis/repos", "events_url": "https://api.github.com/users/SBrandeis/events{/privacy}", "received_events_url": "https://api.github.com/users/SBrandeis/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[]
1,608,295,765,000
1,608,295,765,000
null
CONTRIBUTOR
null
null
null
AFAIK the download functions in `DownloadManager` are not tested yet. It could be good to add some to ensure behavior is as expected.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1604/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1604/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1603
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1603/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1603/comments
https://api.github.com/repos/huggingface/datasets/issues/1603/events
https://github.com/huggingface/datasets/pull/1603
770,857,221
MDExOlB1bGxSZXF1ZXN0NTQyNTIwNDkx
1,603
Add retries to HTTP requests
{ "login": "SBrandeis", "id": 33657802, "node_id": "MDQ6VXNlcjMzNjU3ODAy", "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SBrandeis", "html_url": "https://github.com/SBrandeis", "followers_url": "https://api.github.com/users/SBrandeis/followers", "following_url": "https://api.github.com/users/SBrandeis/following{/other_user}", "gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}", "starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions", "organizations_url": "https://api.github.com/users/SBrandeis/orgs", "repos_url": "https://api.github.com/users/SBrandeis/repos", "events_url": "https://api.github.com/users/SBrandeis/events{/privacy}", "received_events_url": "https://api.github.com/users/SBrandeis/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
[ "merging this one then :) " ]
1,608,295,291,000
1,608,651,247,000
1,608,651,247,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1603", "html_url": "https://github.com/huggingface/datasets/pull/1603", "diff_url": "https://github.com/huggingface/datasets/pull/1603.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1603.patch", "merged_at": 1608651246000 }
## What does this PR do ? Adding retries to HTTP GET & HEAD requests, when they fail with a `ConnectTimeout` exception. The "canonical" way to do this is to use [urllib's Retry class](https://urllib3.readthedocs.io/en/latest/reference/urllib3.util.html#urllib3.util.Retry) and wrap it in a [HttpAdapter](https://requests.readthedocs.io/en/master/api/#requests.adapters.HTTPAdapter). Seems a bit overkill to me, plus it forces us to use the `requests.Session` object. I prefer this simpler implementation. I'm open to remarks and suggestions @lhoestq @yjernite Fixes #1102
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1603/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1603/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1602
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1602/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1602/comments
https://api.github.com/repos/huggingface/datasets/issues/1602/events
https://github.com/huggingface/datasets/pull/1602
770,841,810
MDExOlB1bGxSZXF1ZXN0NTQyNTA4NTM4
1,602
second update of id_newspapers_2018
{ "login": "cahya-wirawan", "id": 7669893, "node_id": "MDQ6VXNlcjc2Njk4OTM=", "avatar_url": "https://avatars.githubusercontent.com/u/7669893?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cahya-wirawan", "html_url": "https://github.com/cahya-wirawan", "followers_url": "https://api.github.com/users/cahya-wirawan/followers", "following_url": "https://api.github.com/users/cahya-wirawan/following{/other_user}", "gists_url": "https://api.github.com/users/cahya-wirawan/gists{/gist_id}", "starred_url": "https://api.github.com/users/cahya-wirawan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cahya-wirawan/subscriptions", "organizations_url": "https://api.github.com/users/cahya-wirawan/orgs", "repos_url": "https://api.github.com/users/cahya-wirawan/repos", "events_url": "https://api.github.com/users/cahya-wirawan/events{/privacy}", "received_events_url": "https://api.github.com/users/cahya-wirawan/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,293,797,000
1,608,633,675,000
1,608,633,674,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1602", "html_url": "https://github.com/huggingface/datasets/pull/1602", "diff_url": "https://github.com/huggingface/datasets/pull/1602.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1602.patch", "merged_at": 1608633674000 }
The feature "url" is currently set wrongly to data["date"], this PR fix it to data["url"]. I added also an additional POC.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1602/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1602/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1601
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1601/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1601/comments
https://api.github.com/repos/huggingface/datasets/issues/1601/events
https://github.com/huggingface/datasets/pull/1601
770,758,914
MDExOlB1bGxSZXF1ZXN0NTQyNDQzNDE3
1,601
second update of the id_newspapers_2018
{ "login": "cahya-wirawan", "id": 7669893, "node_id": "MDQ6VXNlcjc2Njk4OTM=", "avatar_url": "https://avatars.githubusercontent.com/u/7669893?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cahya-wirawan", "html_url": "https://github.com/cahya-wirawan", "followers_url": "https://api.github.com/users/cahya-wirawan/followers", "following_url": "https://api.github.com/users/cahya-wirawan/following{/other_user}", "gists_url": "https://api.github.com/users/cahya-wirawan/gists{/gist_id}", "starred_url": "https://api.github.com/users/cahya-wirawan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cahya-wirawan/subscriptions", "organizations_url": "https://api.github.com/users/cahya-wirawan/orgs", "repos_url": "https://api.github.com/users/cahya-wirawan/repos", "events_url": "https://api.github.com/users/cahya-wirawan/events{/privacy}", "received_events_url": "https://api.github.com/users/cahya-wirawan/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I close this PR, since it based on 1 week old repo. And I will create a new one" ]
1,608,286,220,000
1,608,293,731,000
1,608,293,731,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1601", "html_url": "https://github.com/huggingface/datasets/pull/1601", "diff_url": "https://github.com/huggingface/datasets/pull/1601.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1601.patch", "merged_at": null }
The feature "url" is currently set wrongly to data["date"], this PR fix it to data["url"]. I added also an additional POC.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1601/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1601/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1600
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1600/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1600/comments
https://api.github.com/repos/huggingface/datasets/issues/1600/events
https://github.com/huggingface/datasets/issues/1600
770,582,960
MDU6SXNzdWU3NzA1ODI5NjA=
1,600
AttributeError: 'DatasetDict' object has no attribute 'train_test_split'
{ "login": "david-waterworth", "id": 5028974, "node_id": "MDQ6VXNlcjUwMjg5NzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/5028974?v=4", "gravatar_id": "", "url": "https://api.github.com/users/david-waterworth", "html_url": "https://github.com/david-waterworth", "followers_url": "https://api.github.com/users/david-waterworth/followers", "following_url": "https://api.github.com/users/david-waterworth/following{/other_user}", "gists_url": "https://api.github.com/users/david-waterworth/gists{/gist_id}", "starred_url": "https://api.github.com/users/david-waterworth/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/david-waterworth/subscriptions", "organizations_url": "https://api.github.com/users/david-waterworth/orgs", "repos_url": "https://api.github.com/users/david-waterworth/repos", "events_url": "https://api.github.com/users/david-waterworth/events{/privacy}", "received_events_url": "https://api.github.com/users/david-waterworth/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892912, "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "Further information is requested" } ]
closed
false
null
[]
null
[ "Hi @david-waterworth!\r\n\r\nAs indicated in the error message, `load_dataset(\"csv\")` returns a `DatasetDict` object, which is mapping of `str` to `Dataset` objects. I believe in this case the behavior is to return a `train` split with all the data.\r\n`train_test_split` is a method of the `Dataset` object, so you will need to do something like this:\r\n```python\r\ndataset_dict = load_dataset(`'csv', data_files='data.txt')\r\ndataset = dataset_dict['split name, eg train']\r\ndataset.train_test_split(test_size=0.1)\r\n```\r\n\r\nPlease let me know if this helps. πŸ™‚ ", "Thanks, that's working - the same issue also tripped me up with training. \r\n\r\nI also agree https://github.com/huggingface/datasets/issues/767 would be a useful addition. ", "Closing this now", "> ```python\r\n> dataset_dict = load_dataset(`'csv', data_files='data.txt')\r\n> dataset = dataset_dict['split name, eg train']\r\n> dataset.train_test_split(test_size=0.1)\r\n> ```\r\n\r\nI am getting error like\r\nKeyError: 'split name, eg train'\r\nCould you please tell me how to solve this?", "dataset = load_dataset('csv', data_files=['files/datasets/dataset.csv'])\r\ndataset = dataset['train']\r\ndataset = dataset.train_test_split(test_size=0.1)" ]
1,608,269,830,000
1,623,756,346,000
1,608,536,338,000
NONE
null
null
null
The following code fails with "'DatasetDict' object has no attribute 'train_test_split'" - am I doing something wrong? ``` from datasets import load_dataset dataset = load_dataset('csv', data_files='data.txt') dataset = dataset.train_test_split(test_size=0.1) ``` > AttributeError: 'DatasetDict' object has no attribute 'train_test_split'
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1600/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1600/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1599
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1599/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1599/comments
https://api.github.com/repos/huggingface/datasets/issues/1599/events
https://github.com/huggingface/datasets/pull/1599
770,431,389
MDExOlB1bGxSZXF1ZXN0NTQyMTgwMzI4
1,599
add Korean Sarcasm Dataset
{ "login": "stevhliu", "id": 59462357, "node_id": "MDQ6VXNlcjU5NDYyMzU3", "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stevhliu", "html_url": "https://github.com/stevhliu", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "repos_url": "https://api.github.com/users/stevhliu/repos", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,245,396,000
1,631,897,672,000
1,608,744,359,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1599", "html_url": "https://github.com/huggingface/datasets/pull/1599", "diff_url": "https://github.com/huggingface/datasets/pull/1599.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1599.patch", "merged_at": 1608744359000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1599/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1599/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1598
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1598/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1598/comments
https://api.github.com/repos/huggingface/datasets/issues/1598/events
https://github.com/huggingface/datasets/pull/1598
770,332,440
MDExOlB1bGxSZXF1ZXN0NTQyMDk2NTM4
1,598
made suggested changes in fake-news-english
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,235,589,000
1,608,284,638,000
1,608,284,637,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1598", "html_url": "https://github.com/huggingface/datasets/pull/1598", "diff_url": "https://github.com/huggingface/datasets/pull/1598.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1598.patch", "merged_at": 1608284637000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1598/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1598/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1597
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1597/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1597/comments
https://api.github.com/repos/huggingface/datasets/issues/1597/events
https://github.com/huggingface/datasets/pull/1597
770,276,140
MDExOlB1bGxSZXF1ZXN0NTQyMDUwMTc5
1,597
adding hate-speech-and-offensive-language
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "made suggested changes and opened PR https://github.com/huggingface/datasets/pull/1628" ]
1,608,230,115,000
1,608,766,037,000
1,608,766,036,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1597", "html_url": "https://github.com/huggingface/datasets/pull/1597", "diff_url": "https://github.com/huggingface/datasets/pull/1597.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1597.patch", "merged_at": null }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1597/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1597/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1596
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1596/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1596/comments
https://api.github.com/repos/huggingface/datasets/issues/1596/events
https://github.com/huggingface/datasets/pull/1596
770,260,531
MDExOlB1bGxSZXF1ZXN0NTQyMDM3NTU0
1,596
made suggested changes to hate-speech-and-offensive-language
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,228,566,000
1,608,230,162,000
1,608,230,153,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1596", "html_url": "https://github.com/huggingface/datasets/pull/1596", "diff_url": "https://github.com/huggingface/datasets/pull/1596.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1596.patch", "merged_at": null }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1596/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1596/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1595
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1595/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1595/comments
https://api.github.com/repos/huggingface/datasets/issues/1595/events
https://github.com/huggingface/datasets/pull/1595
770,153,693
MDExOlB1bGxSZXF1ZXN0NTQxOTUwNDk4
1,595
Logiqa en
{ "login": "aclifton314", "id": 53267795, "node_id": "MDQ6VXNlcjUzMjY3Nzk1", "avatar_url": "https://avatars.githubusercontent.com/u/53267795?v=4", "gravatar_id": "", "url": "https://api.github.com/users/aclifton314", "html_url": "https://github.com/aclifton314", "followers_url": "https://api.github.com/users/aclifton314/followers", "following_url": "https://api.github.com/users/aclifton314/following{/other_user}", "gists_url": "https://api.github.com/users/aclifton314/gists{/gist_id}", "starred_url": "https://api.github.com/users/aclifton314/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aclifton314/subscriptions", "organizations_url": "https://api.github.com/users/aclifton314/orgs", "repos_url": "https://api.github.com/users/aclifton314/repos", "events_url": "https://api.github.com/users/aclifton314/events{/privacy}", "received_events_url": "https://api.github.com/users/aclifton314/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "I'm getting an error when I try to create the dummy data:\r\n```python\r\naclifton@pop-os:~/data/hf_datasets_sprint/datasets$ python datasets-cli dummy_data ./datasets/logiqa_en/ --auto_generate \r\n2021-01-07 10:50:12.024791: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcudart.so.10.1'; dlerror: libcudart.so.10.1: cannot open shared object file: No such file or directory\r\n2021-01-07 10:50:12.024814: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\nUsing custom data configuration default\r\nCouldn't generate dummy file 'datasets/dummy/1.1.0/dummy_data/master.zip/LogiQA-dataset-master/README.md'. Ignore that if this file is not useful for dummy data.\r\nDummy data generation done but dummy data test failed since splits ['train', 'test', 'validation'] have 0 examples for config 'default''.\r\nAutomatic dummy data generation failed for some configs of './datasets/logiqa_en/'\r\n```", "Hi ! Sorry for the delay\r\n\r\nTo fix your issue for the dummy data you must increase the number of lines that will be kept to generate the dummy files. By default it's 5, and as you need at least 8 lines here to yield one example you must increase this.\r\n\r\nYou can increase the number of lines to 32 for example by doing\r\n```\r\ndatasets-cli dummy_data ./datasets/logica_en --auto_generate --n_lines 32\r\n```\r\n\r\nAlso it looks like there are changes about other datasets in this PR (imppres). Can you fix that ? You may need to create another branch and another PR.", "To fix the branch issue, I went ahead and made a backup of the dataset then deleted my local copy of my fork of `datasets`. I then followed the [detailed guide](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md) from the beginning to reclone the fork and start a new branch. \r\n\r\nHowever, when it came time to create the dummy data I got the following error:\r\n```python\r\naclifton@pop-os:~/data/hf_datasets_sprint/datasets$ datasets-cli dummy_data ./datasets/logiqa_en --auto_generate --n_lines 32\r\n2021-02-03 11:23:23.145885: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcudart.so.10.1'; dlerror: libcudart.so.10.1: cannot open shared object file: No such file or directory\r\n2021-02-03 11:23:23.145914: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\nUsing custom data configuration default\r\nCouldn't generate dummy file 'datasets/logiqa_en/dummy/1.1.0/dummy_data/master.zip/LogiQA-dataset-master/README.md'. Ignore that if this file is not useful for dummy data.\r\nTraceback (most recent call last):\r\n File \"/home/aclifton/anaconda3/bin/datasets-cli\", line 36, in <module>\r\n service.run()\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/commands/dummy_data.py\", line 317, in run\r\n keep_uncompressed=self._keep_uncompressed,\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/commands/dummy_data.py\", line 355, in _autogenerate_dummy_data\r\n dataset_builder._prepare_split(split_generator)\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/builder.py\", line 905, in _prepare_split\r\n example = self.info.features.encode_example(record)\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/features.py\", line 799, in encode_example\r\n return encode_nested_example(self, example)\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/features.py\", line 710, in encode_nested_example\r\n (k, encode_nested_example(sub_schema, sub_obj)) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/features.py\", line 710, in <genexpr>\r\n (k, encode_nested_example(sub_schema, sub_obj)) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/features.py\", line 737, in encode_nested_example\r\n return schema.encode_example(obj)\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/features.py\", line 522, in encode_example\r\n example_data = self.str2int(example_data)\r\n File \"/home/aclifton/anaconda3/lib/python3.7/site-packages/datasets/features.py\", line 481, in str2int\r\n output.append(self._str2int[str(value)])\r\nKeyError: \"Some Cantonese don't like chili, so some southerners don't like chili.\"\r\n```", "Hi ! The error happens when the script is verifying that the generated dummy data work fine with the dataset script.\r\nApparently it fails because the text `\"Some Cantonese don't like chili, so some southerners don't like chili.\"` was given in a field that is a ClassLabel feature (probably the `answer` field), while it actually expects \"a\", \"b\", \"c\" or \"d\". Can you fix the script so that it returns the expected labels for this field instead of the text ?\r\n\r\nAlso it would be awesome to rename this field `answerKey` instead of `answer` to have the same column names as the other multiple-choice-QA datasets in the library :) ", "Ok getting closer! I got the dummy data to work. However I am now getting the following error:\r\n```python\r\naclifton@pop-os:~/data/hf_datasets_sprint/datasets$ RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_logiqa_en\r\n===================================================================== test session starts ======================================================================\r\nplatform linux -- Python 3.7.6, pytest-5.3.5, py-1.8.1, pluggy-0.13.1\r\nrootdir: /home/aclifton/data/hf_datasets_sprint/datasets\r\nplugins: astropy-header-0.1.2, xdist-2.1.0, doctestplus-0.5.0, forked-1.3.0, hypothesis-5.5.4, arraydiff-0.3, remotedata-0.3.2, openfiles-0.4.0\r\ncollected 0 items / 1 error \r\n\r\n============================================================================ ERRORS ============================================================================\r\n________________________________________________________ ERROR collecting tests/test_dataset_common.py _________________________________________________________\r\nImportError while importing test module '/home/aclifton/data/hf_datasets_sprint/datasets/tests/test_dataset_common.py'.\r\nHint: make sure your test modules/packages have valid Python names.\r\nTraceback:\r\ntests/test_dataset_common.py:42: in <module>\r\n from datasets.packaged_modules import _PACKAGED_DATASETS_MODULES\r\nE ModuleNotFoundError: No module named 'datasets.packaged_modules'\r\n----------------------------------------------------------------------- Captured stderr ------------------------------------------------------------------------\r\n2021-02-10 11:06:14.345510: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcudart.so.10.1'; dlerror: libcudart.so.10.1: cannot open shared object file: No such file or directory\r\n2021-02-10 11:06:14.345551: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\n======================================================================= warnings summary =======================================================================\r\n/home/aclifton/anaconda3/lib/python3.7/site-packages/tensorflow/python/autograph/utils/testing.py:21\r\n /home/aclifton/anaconda3/lib/python3.7/site-packages/tensorflow/python/autograph/utils/testing.py:21: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses\r\n import imp\r\n\r\n/home/aclifton/anaconda3/lib/python3.7/site-packages/apache_beam/typehints/typehints.py:693\r\n /home/aclifton/anaconda3/lib/python3.7/site-packages/apache_beam/typehints/typehints.py:693: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working\r\n if not isinstance(type_params, collections.Iterable):\r\n\r\n/home/aclifton/anaconda3/lib/python3.7/site-packages/apache_beam/typehints/typehints.py:532\r\n /home/aclifton/anaconda3/lib/python3.7/site-packages/apache_beam/typehints/typehints.py:532: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working\r\n if not isinstance(type_params, (collections.Sequence, set)):\r\n\r\n/home/aclifton/anaconda3/lib/python3.7/site-packages/elasticsearch/compat.py:38\r\n /home/aclifton/anaconda3/lib/python3.7/site-packages/elasticsearch/compat.py:38: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working\r\n from collections import Mapping\r\n\r\n-- Docs: https://docs.pytest.org/en/latest/warnings.html\r\n================================================================= 4 warnings, 1 error in 2.74s =================================================================\r\nERROR: not found: /home/aclifton/data/hf_datasets_sprint/datasets/tests/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_logiqa_en\r\n(no name '/home/aclifton/data/hf_datasets_sprint/datasets/tests/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_logiqa_en' in any of [<Module test_dataset_common.py>])\r\n\r\n```", "Hi ! It looks like the version of `datasets` that is installed in your environment doesn't match the version of `datasets` you're using for the tests. Can you try uninstalling datasets and reinstall it again ?\r\n```\r\npip uninstall datasets -y\r\npip install -e .\r\n```", "Closer still!\r\n```python\r\naclifton@pop-os:~/data/hf_datasets_sprint/datasets$ git commit\r\n[logiqa_en 2664fe7f] fixed several issues with logiqa_en.\r\n 4 files changed, 324 insertions(+)\r\n create mode 100644 datasets/logiqa_en/README.md\r\n create mode 100644 datasets/logiqa_en/dataset_infos.json\r\n create mode 100644 datasets/logiqa_en/dummy/1.1.0/dummy_data.zip\r\n create mode 100644 datasets/logiqa_en/logiqa_en.py\r\naclifton@pop-os:~/data/hf_datasets_sprint/datasets$ git fetch upstream\r\nremote: Enumerating objects: 1, done.\r\nremote: Counting objects: 100% (1/1), done.\r\nremote: Total 1 (delta 0), reused 0 (delta 0), pack-reused 0\r\nUnpacking objects: 100% (1/1), 590 bytes | 590.00 KiB/s, done.\r\nFrom https://github.com/huggingface/datasets\r\n 6e114a0c..318b09eb master -> upstream/master\r\naclifton@pop-os:~/data/hf_datasets_sprint/datasets$ git rebase upstream/master \r\nerror: cannot rebase: You have unstaged changes.\r\nerror: Please commit or stash them.\r\naclifton@pop-os:~/data/hf_datasets_sprint/datasets$ git push -u origin logiqa_en\r\nUsername for 'https://github.com': aclifton314\r\nPassword for 'https://aclifton314@github.com': \r\nTo https://github.com/aclifton314/datasets\r\n ! [rejected] logiqa_en -> logiqa_en (non-fast-forward)\r\nerror: failed to push some refs to 'https://github.com/aclifton314/datasets'\r\nhint: Updates were rejected because the tip of your current branch is behind\r\nhint: its remote counterpart. Integrate the remote changes (e.g.\r\nhint: 'git pull ...') before pushing again.\r\nhint: See the 'Note about fast-forwards' in 'git push --help' for details.\r\n```" ]
1,608,219,720,000
1,612,981,992,000
null
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1595", "html_url": "https://github.com/huggingface/datasets/pull/1595", "diff_url": "https://github.com/huggingface/datasets/pull/1595.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1595.patch", "merged_at": null }
logiqa in english.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1595/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1595/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1594
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1594/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1594/comments
https://api.github.com/repos/huggingface/datasets/issues/1594/events
https://github.com/huggingface/datasets/issues/1594
769,747,767
MDU6SXNzdWU3Njk3NDc3Njc=
1,594
connection error
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url": "https://api.github.com/users/rabeehkarimimahabadi/followers", "following_url": "https://api.github.com/users/rabeehkarimimahabadi/following{/other_user}", "gists_url": "https://api.github.com/users/rabeehkarimimahabadi/gists{/gist_id}", "starred_url": "https://api.github.com/users/rabeehkarimimahabadi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rabeehkarimimahabadi/subscriptions", "organizations_url": "https://api.github.com/users/rabeehkarimimahabadi/orgs", "repos_url": "https://api.github.com/users/rabeehkarimimahabadi/repos", "events_url": "https://api.github.com/users/rabeehkarimimahabadi/events{/privacy}", "received_events_url": "https://api.github.com/users/rabeehkarimimahabadi/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "This happen quite often when they are too many concurrent requests to github.\r\n\r\ni can understand it’s a bit cumbersome to handle on the user side. Maybe we should try a few times in the lib (eg with timeout) before failing, what do you think @lhoestq ?", "Yes currently there's no retry afaik. We should add retries", "Retries were added in #1603 :) \r\nIt will be available in the next release", "Hi @lhoestq thank you for the modification, I will use`script_version=\"master\"` for now :), to my experience, also setting timeout to a larger number like 3*60 which I normally use helps a lot on this.\r\n" ]
1,608,196,714,000
1,608,850,653,000
null
NONE
null
null
null
Hi I am hitting to this error, thanks ``` > Traceback (most recent call last): File "finetune_t5_trainer.py", line 379, in <module> main() File "finetune_t5_trainer.py", line 208, in main if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO File "finetune_t5_trainer.py", line 207, in <dictcomp> for task in data_args.eval_tasks} File "/workdir/seq2seq/data/tasks.py", line 70, in get_dataset dataset = self.load_dataset(split=split) File "/workdir/seq2seq/data/tasks.py", line 66, in load_dataset return datasets.load_dataset(self.task.name, split=split, script_version="master") File "/usr/local/lib/python3.6/dist-packages/datasets/load.py", line 589, in load_dataset path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True File "/usr/local/lib/python3.6/dist-packages/datasets/load.py", line 267, in prepare_module local_path = cached_path(file_path, download_config=download_config) File "/usr/local/lib/python3.6/dist-packages/datasets/utils/file_utils.py", line 308, in cached_path use_etag=download_config.use_etag, File "/usr/local/lib/python3.6/dist-packages/datasets/utils/file_utils.py", line 487, in get_from_cache raise ConnectionError("Couldn't reach {}".format(url)) ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/master/datasets/boolq/boolq.py el/0 I1217 01:11:33.898849 354161 main shadow.py:210 Current job status: FINISHED ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1594/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1594/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1593
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1593/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1593/comments
https://api.github.com/repos/huggingface/datasets/issues/1593/events
https://github.com/huggingface/datasets/issues/1593
769,611,386
MDU6SXNzdWU3Njk2MTEzODY=
1,593
Access to key in DatasetDict map
{ "login": "ZhaofengWu", "id": 11954789, "node_id": "MDQ6VXNlcjExOTU0Nzg5", "avatar_url": "https://avatars.githubusercontent.com/u/11954789?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ZhaofengWu", "html_url": "https://github.com/ZhaofengWu", "followers_url": "https://api.github.com/users/ZhaofengWu/followers", "following_url": "https://api.github.com/users/ZhaofengWu/following{/other_user}", "gists_url": "https://api.github.com/users/ZhaofengWu/gists{/gist_id}", "starred_url": "https://api.github.com/users/ZhaofengWu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZhaofengWu/subscriptions", "organizations_url": "https://api.github.com/users/ZhaofengWu/orgs", "repos_url": "https://api.github.com/users/ZhaofengWu/repos", "events_url": "https://api.github.com/users/ZhaofengWu/events{/privacy}", "received_events_url": "https://api.github.com/users/ZhaofengWu/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Indeed that would be cool\r\n\r\nAlso FYI right now the easiest way to do this is\r\n```python\r\ndataset_dict[\"train\"] = dataset_dict[\"train\"].map(my_transform_for_the_train_set)\r\ndataset_dict[\"test\"] = dataset_dict[\"test\"].map(my_transform_for_the_test_set)\r\n```" ]
1,608,188,540,000
1,610,534,283,000
null
NONE
null
null
null
It is possible that we want to do different things in the `map` function (and possibly other functions too) of a `DatasetDict`, depending on the key. I understand that `DatasetDict.map` is a really thin wrapper of `Dataset.map`, so it is easy to directly implement this functionality in the client code. Still, it'd be nice if there can be a flag, similar to `with_indices`, that allows the callable to know the key inside `DatasetDict`.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1593/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1593/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1592
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1592/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1592/comments
https://api.github.com/repos/huggingface/datasets/issues/1592/events
https://github.com/huggingface/datasets/issues/1592
769,529,421
MDU6SXNzdWU3Njk1Mjk0MjE=
1,592
Using datasets.Metric with Trainer()
{ "login": "YipingNUS", "id": 5652584, "node_id": "MDQ6VXNlcjU2NTI1ODQ=", "avatar_url": "https://avatars.githubusercontent.com/u/5652584?v=4", "gravatar_id": "", "url": "https://api.github.com/users/YipingNUS", "html_url": "https://github.com/YipingNUS", "followers_url": "https://api.github.com/users/YipingNUS/followers", "following_url": "https://api.github.com/users/YipingNUS/following{/other_user}", "gists_url": "https://api.github.com/users/YipingNUS/gists{/gist_id}", "starred_url": "https://api.github.com/users/YipingNUS/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/YipingNUS/subscriptions", "organizations_url": "https://api.github.com/users/YipingNUS/orgs", "repos_url": "https://api.github.com/users/YipingNUS/repos", "events_url": "https://api.github.com/users/YipingNUS/events{/privacy}", "received_events_url": "https://api.github.com/users/YipingNUS/received_events", "type": "User", "site_admin": false }
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "We are indeed working on the integration with `Trainer` :)" ]
1,608,182,224,000
1,608,205,744,000
null
NONE
null
null
null
## Using datasets.Metric with Trainer() Hi team, I was quite surprised in the [Metric documentation](https://huggingface.co/docs/datasets/using_metrics.html) I don't see how it can be used with `Trainer()`. That would be the most intuitive use case instead of having to iterate the batches and add predictions and references to the metric, then compute the metric manually. Ideally, any pre-built metrics can be added to `compute_metrics` argument of `Trainer()` and they will be calculated at an interval specified by `TrainingArguments.evaluation_strategy`. Is this option available but just not mentioned in the documentation or it's not possible at the moment? I notice in the [Transformer | Training and fine-tuning](https://huggingface.co/transformers/training.html) tutorial, you are using custom scripts to calculate the accuracy, P/R/F, which are already in the pre-built metrics.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1592/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1592/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1591
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1591/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1591/comments
https://api.github.com/repos/huggingface/datasets/issues/1591/events
https://github.com/huggingface/datasets/issues/1591
769,383,714
MDU6SXNzdWU3NjkzODM3MTQ=
1,591
IWSLT-17 Link Broken
{ "login": "ZhaofengWu", "id": 11954789, "node_id": "MDQ6VXNlcjExOTU0Nzg5", "avatar_url": "https://avatars.githubusercontent.com/u/11954789?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ZhaofengWu", "html_url": "https://github.com/ZhaofengWu", "followers_url": "https://api.github.com/users/ZhaofengWu/followers", "following_url": "https://api.github.com/users/ZhaofengWu/following{/other_user}", "gists_url": "https://api.github.com/users/ZhaofengWu/gists{/gist_id}", "starred_url": "https://api.github.com/users/ZhaofengWu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZhaofengWu/subscriptions", "organizations_url": "https://api.github.com/users/ZhaofengWu/orgs", "repos_url": "https://api.github.com/users/ZhaofengWu/repos", "events_url": "https://api.github.com/users/ZhaofengWu/events{/privacy}", "received_events_url": "https://api.github.com/users/ZhaofengWu/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892865, "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate", "name": "duplicate", "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists" }, { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
null
[]
null
[ "Sorry, this is a duplicate of #1287. Not sure why it didn't come up when I searched `iwslt` in the issues list.", "Closing this since its a duplicate" ]
1,608,166,002,000
1,608,278,796,000
1,608,278,728,000
NONE
null
null
null
``` FileNotFoundError: Couldn't find file at https://wit3.fbk.eu/archive/2017-01-trnmted//texts/DeEnItNlRo/DeEnItNlRo/DeEnItNlRo-DeEnItNlRo.tgz ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1591/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1591/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1590
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1590/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1590/comments
https://api.github.com/repos/huggingface/datasets/issues/1590/events
https://github.com/huggingface/datasets/issues/1590
769,242,858
MDU6SXNzdWU3NjkyNDI4NTg=
1,590
Add helper to resolve namespace collision
{ "login": "jramapuram", "id": 8204807, "node_id": "MDQ6VXNlcjgyMDQ4MDc=", "avatar_url": "https://avatars.githubusercontent.com/u/8204807?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jramapuram", "html_url": "https://github.com/jramapuram", "followers_url": "https://api.github.com/users/jramapuram/followers", "following_url": "https://api.github.com/users/jramapuram/following{/other_user}", "gists_url": "https://api.github.com/users/jramapuram/gists{/gist_id}", "starred_url": "https://api.github.com/users/jramapuram/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jramapuram/subscriptions", "organizations_url": "https://api.github.com/users/jramapuram/orgs", "repos_url": "https://api.github.com/users/jramapuram/repos", "events_url": "https://api.github.com/users/jramapuram/events{/privacy}", "received_events_url": "https://api.github.com/users/jramapuram/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Do you have an example?", "I was thinking about using something like [importlib](https://docs.python.org/3/library/importlib.html#importing-a-source-file-directly) to over-ride the collision. \r\n\r\n**Reason requested**: I use the [following template](https://github.com/jramapuram/ml_base/) repo where I house all my datasets as a submodule.", "Alternatively huggingface could consider some submodule type structure like:\r\n\r\n`import huggingface.datasets`\r\n`import huggingface.transformers`\r\n\r\n`datasets` is a very common module in ML and should be an end-user decision and not scope all of python Β―\\_(ツ)_/Β― \r\n", "That's a interesting option indeed. We'll think about it.", "It also wasn't initially obvious to me that the samples which contain `import datasets` were in fact importing a huggingface library (in fact all the huggingface imports are very generic - transformers, tokenizers, datasets...)" ]
1,608,149,844,000
1,608,349,238,000
null
NONE
null
null
null
Many projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1590/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1590/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1589
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1589/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1589/comments
https://api.github.com/repos/huggingface/datasets/issues/1589/events
https://github.com/huggingface/datasets/pull/1589
769,187,141
MDExOlB1bGxSZXF1ZXN0NTQxMzcwMTM0
1,589
Update doc2dial.py
{ "login": "songfeng", "id": 2062185, "node_id": "MDQ6VXNlcjIwNjIxODU=", "avatar_url": "https://avatars.githubusercontent.com/u/2062185?v=4", "gravatar_id": "", "url": "https://api.github.com/users/songfeng", "html_url": "https://github.com/songfeng", "followers_url": "https://api.github.com/users/songfeng/followers", "following_url": "https://api.github.com/users/songfeng/following{/other_user}", "gists_url": "https://api.github.com/users/songfeng/gists{/gist_id}", "starred_url": "https://api.github.com/users/songfeng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/songfeng/subscriptions", "organizations_url": "https://api.github.com/users/songfeng/orgs", "repos_url": "https://api.github.com/users/songfeng/repos", "events_url": "https://api.github.com/users/songfeng/events{/privacy}", "received_events_url": "https://api.github.com/users/songfeng/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Thanks for adding the `doc2dial_rc` config :) \r\n\r\nIt looks like you're missing the dummy data for this config though. Could you add them please ?\r\nAlso to fix the CI you'll need to format the code with `make style`" ]
1,608,144,656,000
1,608,571,111,000
null
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1589", "html_url": "https://github.com/huggingface/datasets/pull/1589", "diff_url": "https://github.com/huggingface/datasets/pull/1589.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1589.patch", "merged_at": null }
Added data loader for machine reading comprehension tasks proposed in the Doc2Dial EMNLP 2020 paper.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1589/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1589/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1588
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1588/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1588/comments
https://api.github.com/repos/huggingface/datasets/issues/1588/events
https://github.com/huggingface/datasets/pull/1588
769,068,227
MDExOlB1bGxSZXF1ZXN0NTQxMjg3OTcz
1,588
Modified hind encorp
{ "login": "rahul-art", "id": 56379013, "node_id": "MDQ6VXNlcjU2Mzc5MDEz", "avatar_url": "https://avatars.githubusercontent.com/u/56379013?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rahul-art", "html_url": "https://github.com/rahul-art", "followers_url": "https://api.github.com/users/rahul-art/followers", "following_url": "https://api.github.com/users/rahul-art/following{/other_user}", "gists_url": "https://api.github.com/users/rahul-art/gists{/gist_id}", "starred_url": "https://api.github.com/users/rahul-art/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rahul-art/subscriptions", "organizations_url": "https://api.github.com/users/rahul-art/orgs", "repos_url": "https://api.github.com/users/rahul-art/repos", "events_url": "https://api.github.com/users/rahul-art/events{/privacy}", "received_events_url": "https://api.github.com/users/rahul-art/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "welcome, awesome " ]
1,608,136,094,000
1,608,158,513,000
1,608,139,228,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1588", "html_url": "https://github.com/huggingface/datasets/pull/1588", "diff_url": "https://github.com/huggingface/datasets/pull/1588.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1588.patch", "merged_at": 1608139228000 }
description added, unnecessary comments removed from .py and readme.md reformated @lhoestq for #1584
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1588/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1588/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1587
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1587/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1587/comments
https://api.github.com/repos/huggingface/datasets/issues/1587/events
https://github.com/huggingface/datasets/pull/1587
768,929,877
MDExOlB1bGxSZXF1ZXN0NTQxMjAwMDk3
1,587
Add nq_open question answering dataset
{ "login": "Nilanshrajput", "id": 28673745, "node_id": "MDQ6VXNlcjI4NjczNzQ1", "avatar_url": "https://avatars.githubusercontent.com/u/28673745?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Nilanshrajput", "html_url": "https://github.com/Nilanshrajput", "followers_url": "https://api.github.com/users/Nilanshrajput/followers", "following_url": "https://api.github.com/users/Nilanshrajput/following{/other_user}", "gists_url": "https://api.github.com/users/Nilanshrajput/gists{/gist_id}", "starred_url": "https://api.github.com/users/Nilanshrajput/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nilanshrajput/subscriptions", "organizations_url": "https://api.github.com/users/Nilanshrajput/orgs", "repos_url": "https://api.github.com/users/Nilanshrajput/repos", "events_url": "https://api.github.com/users/Nilanshrajput/events{/privacy}", "received_events_url": "https://api.github.com/users/Nilanshrajput/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "@SBrandeis all checks passing" ]
1,608,128,528,000
1,608,221,230,000
1,608,221,230,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1587", "html_url": "https://github.com/huggingface/datasets/pull/1587", "diff_url": "https://github.com/huggingface/datasets/pull/1587.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1587.patch", "merged_at": 1608221230000 }
this is pr is a copy of #1506 due to messed up git history in that pr.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1587/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1587/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1586
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1586/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1586/comments
https://api.github.com/repos/huggingface/datasets/issues/1586/events
https://github.com/huggingface/datasets/pull/1586
768,864,502
MDExOlB1bGxSZXF1ZXN0NTQxMTY0MDc2
1,586
added irc disentangle dataset
{ "login": "dhruvjoshi1998", "id": 32560035, "node_id": "MDQ6VXNlcjMyNTYwMDM1", "avatar_url": "https://avatars.githubusercontent.com/u/32560035?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhruvjoshi1998", "html_url": "https://github.com/dhruvjoshi1998", "followers_url": "https://api.github.com/users/dhruvjoshi1998/followers", "following_url": "https://api.github.com/users/dhruvjoshi1998/following{/other_user}", "gists_url": "https://api.github.com/users/dhruvjoshi1998/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhruvjoshi1998/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhruvjoshi1998/subscriptions", "organizations_url": "https://api.github.com/users/dhruvjoshi1998/orgs", "repos_url": "https://api.github.com/users/dhruvjoshi1998/repos", "events_url": "https://api.github.com/users/dhruvjoshi1998/events{/privacy}", "received_events_url": "https://api.github.com/users/dhruvjoshi1998/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "@lhoestq sorry, this was the only way I was able to fix the pull request ", "@lhoestq Thank you for the feedback. I wondering whether I should be passing an 'id' field in the dictionary since the 'connections' reference the 'id' of the linked messages. This 'id' would just be the same as the id_ that is in the yielded tuple.", "Yes indeed it would be cool to have the ids in the dictionary. This way the dataset can be shuffled and all without losing information about the connections. Can you add it if you don't mind ?", "Thanks :) could you also add the ids in the dictionary since they're useful for the connection links ?", "Thanks !\r\nAlso it looks like the dummy_data.zip were regenerated and are now back to being too big (300KB each).\r\nCan you reduce their sizes ? You can actually just revert to the ones you had before the last commit" ]
1,608,125,158,000
1,611,916,133,000
1,611,916,133,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1586", "html_url": "https://github.com/huggingface/datasets/pull/1586", "diff_url": "https://github.com/huggingface/datasets/pull/1586.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1586.patch", "merged_at": 1611916133000 }
added irc disentanglement dataset
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1586/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1586/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1585
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1585/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1585/comments
https://api.github.com/repos/huggingface/datasets/issues/1585/events
https://github.com/huggingface/datasets/issues/1585
768,831,171
MDU6SXNzdWU3Njg4MzExNzE=
1,585
FileNotFoundError for `amazon_polarity`
{ "login": "phtephanx", "id": 24647404, "node_id": "MDQ6VXNlcjI0NjQ3NDA0", "avatar_url": "https://avatars.githubusercontent.com/u/24647404?v=4", "gravatar_id": "", "url": "https://api.github.com/users/phtephanx", "html_url": "https://github.com/phtephanx", "followers_url": "https://api.github.com/users/phtephanx/followers", "following_url": "https://api.github.com/users/phtephanx/following{/other_user}", "gists_url": "https://api.github.com/users/phtephanx/gists{/gist_id}", "starred_url": "https://api.github.com/users/phtephanx/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/phtephanx/subscriptions", "organizations_url": "https://api.github.com/users/phtephanx/orgs", "repos_url": "https://api.github.com/users/phtephanx/repos", "events_url": "https://api.github.com/users/phtephanx/events{/privacy}", "received_events_url": "https://api.github.com/users/phtephanx/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi @phtephanx , the `amazon_polarity` dataset has not been released yet. It will be available in the coming soon v2of `datasets` :) \r\n\r\nYou can still access it now if you want, but you will need to install datasets via the master branch:\r\n`pip install git+https://github.com/huggingface/datasets.git@master`" ]
1,608,123,065,000
1,608,134,576,000
1,608,134,576,000
NONE
null
null
null
Version: `datasets==v1.1.3` ### Reproduction ```python from datasets import load_dataset data = load_dataset("amazon_polarity") ``` crashes with ```bash FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/amazon_polarity/amazon_polarity.py ``` and ```bash FileNotFoundError: Couldn't find file at https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/amazon_polarity/amazon_polarity.py ``` and ```bash FileNotFoundError: Couldn't find file locally at amazon_polarity/amazon_polarity.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/amazon_polarity/amazon_polarity.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/amazon_polarity/amazon_polarity.py ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1585/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1585/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1584
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1584/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1584/comments
https://api.github.com/repos/huggingface/datasets/issues/1584/events
https://github.com/huggingface/datasets/pull/1584
768,820,406
MDExOlB1bGxSZXF1ZXN0NTQxMTM2OTQ5
1,584
Load hind encorp
{ "login": "rahul-art", "id": 56379013, "node_id": "MDQ6VXNlcjU2Mzc5MDEz", "avatar_url": "https://avatars.githubusercontent.com/u/56379013?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rahul-art", "html_url": "https://github.com/rahul-art", "followers_url": "https://api.github.com/users/rahul-art/followers", "following_url": "https://api.github.com/users/rahul-art/following{/other_user}", "gists_url": "https://api.github.com/users/rahul-art/gists{/gist_id}", "starred_url": "https://api.github.com/users/rahul-art/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rahul-art/subscriptions", "organizations_url": "https://api.github.com/users/rahul-art/orgs", "repos_url": "https://api.github.com/users/rahul-art/repos", "events_url": "https://api.github.com/users/rahul-art/events{/privacy}", "received_events_url": "https://api.github.com/users/rahul-art/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,122,318,000
1,608,258,444,000
1,608,258,444,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1584", "html_url": "https://github.com/huggingface/datasets/pull/1584", "diff_url": "https://github.com/huggingface/datasets/pull/1584.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1584.patch", "merged_at": null }
reformated well documented, yaml tags added, code
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1584/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1584/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1583
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1583/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1583/comments
https://api.github.com/repos/huggingface/datasets/issues/1583/events
https://github.com/huggingface/datasets/pull/1583
768,795,986
MDExOlB1bGxSZXF1ZXN0NTQxMTIyODEz
1,583
Update metrics docstrings.
{ "login": "Fraser-Greenlee", "id": 8402500, "node_id": "MDQ6VXNlcjg0MDI1MDA=", "avatar_url": "https://avatars.githubusercontent.com/u/8402500?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Fraser-Greenlee", "html_url": "https://github.com/Fraser-Greenlee", "followers_url": "https://api.github.com/users/Fraser-Greenlee/followers", "following_url": "https://api.github.com/users/Fraser-Greenlee/following{/other_user}", "gists_url": "https://api.github.com/users/Fraser-Greenlee/gists{/gist_id}", "starred_url": "https://api.github.com/users/Fraser-Greenlee/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Fraser-Greenlee/subscriptions", "organizations_url": "https://api.github.com/users/Fraser-Greenlee/orgs", "repos_url": "https://api.github.com/users/Fraser-Greenlee/repos", "events_url": "https://api.github.com/users/Fraser-Greenlee/events{/privacy}", "received_events_url": "https://api.github.com/users/Fraser-Greenlee/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,120,858,000
1,608,316,746,000
1,608,316,746,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1583", "html_url": "https://github.com/huggingface/datasets/pull/1583", "diff_url": "https://github.com/huggingface/datasets/pull/1583.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1583.patch", "merged_at": 1608316746000 }
#1478 Correcting the argument descriptions for metrics. Let me know if there's any issues.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1583/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1583/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1582
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1582/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1582/comments
https://api.github.com/repos/huggingface/datasets/issues/1582/events
https://github.com/huggingface/datasets/pull/1582
768,776,617
MDExOlB1bGxSZXF1ZXN0NTQxMTEwODU1
1,582
Adding wiki lingua dataset as new branch
{ "login": "katnoria", "id": 7674948, "node_id": "MDQ6VXNlcjc2NzQ5NDg=", "avatar_url": "https://avatars.githubusercontent.com/u/7674948?v=4", "gravatar_id": "", "url": "https://api.github.com/users/katnoria", "html_url": "https://github.com/katnoria", "followers_url": "https://api.github.com/users/katnoria/followers", "following_url": "https://api.github.com/users/katnoria/following{/other_user}", "gists_url": "https://api.github.com/users/katnoria/gists{/gist_id}", "starred_url": "https://api.github.com/users/katnoria/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/katnoria/subscriptions", "organizations_url": "https://api.github.com/users/katnoria/orgs", "repos_url": "https://api.github.com/users/katnoria/repos", "events_url": "https://api.github.com/users/katnoria/events{/privacy}", "received_events_url": "https://api.github.com/users/katnoria/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,119,587,000
1,608,228,406,000
1,608,228,405,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1582", "html_url": "https://github.com/huggingface/datasets/pull/1582", "diff_url": "https://github.com/huggingface/datasets/pull/1582.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1582.patch", "merged_at": 1608228405000 }
Adding the dataset as new branch as advised here: #1470
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1582/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1582/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1581
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1581/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1581/comments
https://api.github.com/repos/huggingface/datasets/issues/1581/events
https://github.com/huggingface/datasets/issues/1581
768,320,594
MDU6SXNzdWU3NjgzMjA1OTQ=
1,581
Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'
{ "login": "eduardofv", "id": 702586, "node_id": "MDQ6VXNlcjcwMjU4Ng==", "avatar_url": "https://avatars.githubusercontent.com/u/702586?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eduardofv", "html_url": "https://github.com/eduardofv", "followers_url": "https://api.github.com/users/eduardofv/followers", "following_url": "https://api.github.com/users/eduardofv/following{/other_user}", "gists_url": "https://api.github.com/users/eduardofv/gists{/gist_id}", "starred_url": "https://api.github.com/users/eduardofv/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eduardofv/subscriptions", "organizations_url": "https://api.github.com/users/eduardofv/orgs", "repos_url": "https://api.github.com/users/eduardofv/repos", "events_url": "https://api.github.com/users/eduardofv/events{/privacy}", "received_events_url": "https://api.github.com/users/eduardofv/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nYou can override the directory in which cache file are stored using for example\r\n```\r\nENV HF_HOME=\"/root/cache/hf_cache_home\"\r\n```\r\n\r\nThis way both `transformers` and `datasets` will use this directory instead of the default `.cache`", "Great, thanks. I didn't see documentation about than ENV variable, looks like an obvious solution. ", "> Thanks for reporting !\r\n> You can override the directory in which cache file are stored using for example\r\n> \r\n> ```\r\n> ENV HF_HOME=\"/root/cache/hf_cache_home\"\r\n> ```\r\n> \r\n> This way both `transformers` and `datasets` will use this directory instead of the default `.cache`\r\n\r\ncan we disable caching directly?", "Hi ! Unfortunately no since we need this directory to load datasets.\r\nWhen you load a dataset, it downloads the raw data files in the cache directory inside <cache_dir>/downloads. Then it builds the dataset and saves it as arrow data inside <cache_dir>/<dataset_name>.\r\n\r\nHowever you can specify the directory of your choice, and it can be a temporary directory if you want to clean everything up at one point.", "I'm closing this to keep issues a bit cleaner" ]
1,608,076,941,000
1,623,944,445,000
1,623,944,445,000
NONE
null
null
null
I am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `/.cache`: ``` $ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)/data:/root/data -v $(pwd):/root -v $(pwd)/models/:/root/models -v $(pwd)/saved_models/:/root/saved_models -e "HOST_HOSTNAME=$(hostname)" hf-error:latest /bin/bash ________ _______________ ___ __/__________________________________ ____/__ /________ __ __ / _ _ \_ __ \_ ___/ __ \_ ___/_ /_ __ /_ __ \_ | /| / / _ / / __/ / / /(__ )/ /_/ / / _ __/ _ / / /_/ /_ |/ |/ / /_/ \___//_/ /_//____/ \____//_/ /_/ /_/ \____/____/|__/ You are running this container as user with ID 1000 and group 1000, which should map to the ID and group for your user on the Docker host. Great! tf-docker /root > python Python 3.6.9 (default, Oct 8 2020, 12:12:24) [GCC 8.4.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import transformers 2020-12-15 23:53:21.165827: I tensorflow/stream_executor/platform/default/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0 Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/local/lib/python3.6/dist-packages/transformers/__init__.py", line 22, in <module> from .integrations import ( # isort:skip File "/usr/local/lib/python3.6/dist-packages/transformers/integrations.py", line 5, in <module> from .trainer_utils import EvaluationStrategy File "/usr/local/lib/python3.6/dist-packages/transformers/trainer_utils.py", line 25, in <module> from .file_utils import is_tf_available, is_torch_available, is_torch_tpu_available File "/usr/local/lib/python3.6/dist-packages/transformers/file_utils.py", line 88, in <module> import datasets # noqa: F401 File "/usr/local/lib/python3.6/dist-packages/datasets/__init__.py", line 26, in <module> from .arrow_dataset import Dataset, concatenate_datasets File "/usr/local/lib/python3.6/dist-packages/datasets/arrow_dataset.py", line 40, in <module> from .arrow_reader import ArrowReader File "/usr/local/lib/python3.6/dist-packages/datasets/arrow_reader.py", line 31, in <module> from .utils import cached_path, logging File "/usr/local/lib/python3.6/dist-packages/datasets/utils/__init__.py", line 20, in <module> from .download_manager import DownloadManager, GenerateMode File "/usr/local/lib/python3.6/dist-packages/datasets/utils/download_manager.py", line 25, in <module> from .file_utils import HF_DATASETS_CACHE, cached_path, get_from_cache, hash_url_to_filename File "/usr/local/lib/python3.6/dist-packages/datasets/utils/file_utils.py", line 118, in <module> os.makedirs(HF_MODULES_CACHE, exist_ok=True) File "/usr/lib/python3.6/os.py", line 210, in makedirs makedirs(head, mode, exist_ok) File "/usr/lib/python3.6/os.py", line 210, in makedirs makedirs(head, mode, exist_ok) File "/usr/lib/python3.6/os.py", line 220, in makedirs mkdir(name, mode) PermissionError: [Errno 13] Permission denied: '/.cache' ``` I've pinned the problem to `RUN pip install datasets`, and by commenting it you can actually import transformers correctly. Another workaround I've found is creating the directory and giving permissions to it directly on the Dockerfile. ``` FROM tensorflow/tensorflow:latest-gpu-jupyter WORKDIR /root EXPOSE 80 EXPOSE 8888 EXPOSE 6006 ENV SHELL /bin/bash ENV PATH="/root/.local/bin:${PATH}" ENV CUDA_CACHE_PATH="/root/cache/cuda" ENV CUDA_CACHE_MAXSIZE="4294967296" ENV TFHUB_CACHE_DIR="/root/cache/tfhub" RUN pip install --upgrade pip RUN apt update -y && apt upgrade -y RUN pip install transformers #Installing datasets will throw the error, try commenting and rebuilding RUN pip install datasets #Another workaround is creating the directory and give permissions explicitly #RUN mkdir /.cache #RUN chmod 777 /.cache ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1581/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1581/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/1580
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1580/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1580/comments
https://api.github.com/repos/huggingface/datasets/issues/1580/events
https://github.com/huggingface/datasets/pull/1580
768,111,377
MDExOlB1bGxSZXF1ZXN0NTQwNjQxNDQ3
1,580
made suggested changes in diplomacy_detection.py
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,061,920,000
1,608,114,472,000
1,608,114,472,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1580", "html_url": "https://github.com/huggingface/datasets/pull/1580", "diff_url": "https://github.com/huggingface/datasets/pull/1580.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1580.patch", "merged_at": 1608114472000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1580/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1580/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1579
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1579/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1579/comments
https://api.github.com/repos/huggingface/datasets/issues/1579/events
https://github.com/huggingface/datasets/pull/1579
767,808,465
MDExOlB1bGxSZXF1ZXN0NTQwMzk5OTY5
1,579
Adding CLIMATE-FEVER dataset
{ "login": "tdiggelm", "id": 1658969, "node_id": "MDQ6VXNlcjE2NTg5Njk=", "avatar_url": "https://avatars.githubusercontent.com/u/1658969?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tdiggelm", "html_url": "https://github.com/tdiggelm", "followers_url": "https://api.github.com/users/tdiggelm/followers", "following_url": "https://api.github.com/users/tdiggelm/following{/other_user}", "gists_url": "https://api.github.com/users/tdiggelm/gists{/gist_id}", "starred_url": "https://api.github.com/users/tdiggelm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tdiggelm/subscriptions", "organizations_url": "https://api.github.com/users/tdiggelm/orgs", "repos_url": "https://api.github.com/users/tdiggelm/repos", "events_url": "https://api.github.com/users/tdiggelm/events{/privacy}", "received_events_url": "https://api.github.com/users/tdiggelm/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I `git rebase`ed my branch to `upstream/master` as suggested in point 7 of <https://huggingface.co/docs/datasets/share_dataset.html> and subsequently used `git pull` to be able to push to my remote branch. However, I think this messed up the history.\r\n\r\nPlease let me know if I should create a clean new PR with my changes.\r\n\r\nUpdate: I also fixed the dataset name in the Dataset Card.", "Dear @SBrandeis , @lhoestq . I am not sure how to fix the PR with respect to the additional files that are currently included in the commits. Could you provide me with an example? Otherwise I would be happy to close/re-open another PR. Please let me know if anything is missing for the review.", "Hi @tdiggelm, thanks for the contribution! This dataset is really awesome.\r\nI believe creating a new branch from master and opening a new PR with your changes is the simplest option since no review has been done yet. Feel free to ping us when it's done.", "> Hi @tdiggelm, thanks for the contribution! This dataset is really awesome.\r\n> I believe creating a new branch from master and opening a new PR with your changes is the simplest option since no review has been done yet. Feel free to ping us when it's done.\r\n\r\nThank you very much for your quick reply! Will do ASAP and ping you when done.", "closing in favor of #1623" ]
1,608,050,962,000
1,608,644,596,000
1,608,644,595,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1579", "html_url": "https://github.com/huggingface/datasets/pull/1579", "diff_url": "https://github.com/huggingface/datasets/pull/1579.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1579.patch", "merged_at": null }
This PR request the addition of the CLIMATE-FEVER dataset: A dataset adopting the FEVER methodology that consists of 1,535 real-world claims regarding climate-change collected on the internet. Each claim is accompanied by five manually annotated evidence sentences retrieved from the English Wikipedia that support, refute or do not give enough information to validate the claim totalling in 7,675 claim-evidence pairs. The dataset features challenging claims that relate multiple facets and disputed cases of claims where both supporting and refuting evidence are present. More information can be found at: - Homepage: <http://climatefever.ai> - Paper: <https://arxiv.org/abs/2012.00614>
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1579/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1579/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1578
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1578/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1578/comments
https://api.github.com/repos/huggingface/datasets/issues/1578/events
https://github.com/huggingface/datasets/pull/1578
767,760,513
MDExOlB1bGxSZXF1ZXN0NTQwMzY1NzYz
1,578
update multiwozv22 checksums
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}", "starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yjernite/subscriptions", "organizations_url": "https://api.github.com/users/yjernite/orgs", "repos_url": "https://api.github.com/users/yjernite/repos", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "received_events_url": "https://api.github.com/users/yjernite/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,608,048,832,000
1,608,051,989,000
1,608,051,989,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1578", "html_url": "https://github.com/huggingface/datasets/pull/1578", "diff_url": "https://github.com/huggingface/datasets/pull/1578.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1578.patch", "merged_at": 1608051989000 }
a file was updated on the GitHub repo for the dataset
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1578/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1578/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1577
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1577/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1577/comments
https://api.github.com/repos/huggingface/datasets/issues/1577/events
https://github.com/huggingface/datasets/pull/1577
767,342,432
MDExOlB1bGxSZXF1ZXN0NTQwMDg2MzY5
1,577
Add comet metric
{ "login": "ricardorei", "id": 17256847, "node_id": "MDQ6VXNlcjE3MjU2ODQ3", "avatar_url": "https://avatars.githubusercontent.com/u/17256847?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ricardorei", "html_url": "https://github.com/ricardorei", "followers_url": "https://api.github.com/users/ricardorei/followers", "following_url": "https://api.github.com/users/ricardorei/following{/other_user}", "gists_url": "https://api.github.com/users/ricardorei/gists{/gist_id}", "starred_url": "https://api.github.com/users/ricardorei/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ricardorei/subscriptions", "organizations_url": "https://api.github.com/users/ricardorei/orgs", "repos_url": "https://api.github.com/users/ricardorei/repos", "events_url": "https://api.github.com/users/ricardorei/events{/privacy}", "received_events_url": "https://api.github.com/users/ricardorei/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I also thought a bit about the fact that \"sources\" can't be added to the batch.. but changing that would require a lot more changes. And I agree that the idea of adding them as part of the references is not ideal. Conceptually they are not references.\r\n\r\nI would keep it like this for now.. And in the future, work on a more consistent batch interface." ]
1,608,022,560,000
1,610,631,190,000
1,610,631,190,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1577", "html_url": "https://github.com/huggingface/datasets/pull/1577", "diff_url": "https://github.com/huggingface/datasets/pull/1577.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1577.patch", "merged_at": 1610631190000 }
Hey! I decided to add our new Crosslingual Optimized Metric for Evaluation of Translation (COMET) to the list of the available metrics. COMET was [presented at EMNLP20](https://www.aclweb.org/anthology/2020.emnlp-main.213/) and it is the highest performing metric, so far, on the WMT19 benchmark. We also participated in the [WMT20 Metrics shared task ](http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf) where once again COMET was validated as a top-performing metric. I hope that this metric will help researcher's and industry workers to better validate their MT systems in the future πŸ€— ! Cheers, Ricardo
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1577/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1577/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1576
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1576/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1576/comments
https://api.github.com/repos/huggingface/datasets/issues/1576/events
https://github.com/huggingface/datasets/pull/1576
767,080,645
MDExOlB1bGxSZXF1ZXN0NTM5OTE3MTA0
1,576
Remove the contributors section
{ "login": "clmnt", "id": 821155, "node_id": "MDQ6VXNlcjgyMTE1NQ==", "avatar_url": "https://avatars.githubusercontent.com/u/821155?v=4", "gravatar_id": "", "url": "https://api.github.com/users/clmnt", "html_url": "https://github.com/clmnt", "followers_url": "https://api.github.com/users/clmnt/followers", "following_url": "https://api.github.com/users/clmnt/following{/other_user}", "gists_url": "https://api.github.com/users/clmnt/gists{/gist_id}", "starred_url": "https://api.github.com/users/clmnt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/clmnt/subscriptions", "organizations_url": "https://api.github.com/users/clmnt/orgs", "repos_url": "https://api.github.com/users/clmnt/repos", "events_url": "https://api.github.com/users/clmnt/events{/privacy}", "received_events_url": "https://api.github.com/users/clmnt/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,996,835,000
1,608,036,827,000
1,608,036,826,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1576", "html_url": "https://github.com/huggingface/datasets/pull/1576", "diff_url": "https://github.com/huggingface/datasets/pull/1576.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1576.patch", "merged_at": 1608036826000 }
sourcerer is down
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1576/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1576/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1575
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1575/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1575/comments
https://api.github.com/repos/huggingface/datasets/issues/1575/events
https://github.com/huggingface/datasets/pull/1575
767,076,374
MDExOlB1bGxSZXF1ZXN0NTM5OTEzNzgx
1,575
Hind_Encorp all done
{ "login": "rahul-art", "id": 56379013, "node_id": "MDQ6VXNlcjU2Mzc5MDEz", "avatar_url": "https://avatars.githubusercontent.com/u/56379013?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rahul-art", "html_url": "https://github.com/rahul-art", "followers_url": "https://api.github.com/users/rahul-art/followers", "following_url": "https://api.github.com/users/rahul-art/following{/other_user}", "gists_url": "https://api.github.com/users/rahul-art/gists{/gist_id}", "starred_url": "https://api.github.com/users/rahul-art/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rahul-art/subscriptions", "organizations_url": "https://api.github.com/users/rahul-art/orgs", "repos_url": "https://api.github.com/users/rahul-art/repos", "events_url": "https://api.github.com/users/rahul-art/events{/privacy}", "received_events_url": "https://api.github.com/users/rahul-art/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "ALL TEST PASSED locally @yjernite ", "@rahul-art kindly run the following from the datasets folder \r\n\r\n```\r\nmake style \r\nflake8 datasets\r\n\r\n```\r\n", "@skyprince999 I did that before it says all done \r\n", "I did that again it gives the same output all done and then I synchronized my changes with this branch ", "@lhoestq i did all the changes you suggested but at the time of load_dataset it is giving me error\r\n`**`datasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=76591256, num_examples=1, dataset_name='hind_encorp'), 'recorded': SplitInfo(name='train', num_bytes=78945714, num_examples=273885, dataset_name='hind_encorp')}]`**`", "\r\n\r\n\r\nI cloned the branch and it seems to work fine at my end. try to clear the cache - \r\n\r\n```\r\nrm -rf /home/ubuntu/.cache/huggingface/datasets/\r\nrm -rf /home/ubuntu/.cache/huggingface/modules/datasets_modules//datasets/\r\n```\r\nBut the dataset has only one record. Is that correct? \r\n![image](https://user-images.githubusercontent.com/9033954/102331376-c7929b00-3fb0-11eb-8a6c-81b2cf47bc2a.png)\r\n", "> @lhoestq i did all the changes you suggested but at the time of load_dataset it is giving me error\r\n> `**`datasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=76591256, num_examples=1, dataset_name='hind_encorp'), 'recorded': SplitInfo(name='train', num_bytes=78945714, num_examples=273885, dataset_name='hind_encorp')}]`**`\r\n\r\nYou can ignore this error by adding `ignore_verifications=True` to `load_dataset`.\r\n\r\nThis error is raised because you're loading a dataset that you've already loaded once in the past. Therefore the library does some verifications to make sure it's generated the same way. \r\n\r\nHowever since you've done changes in the dataset script you should ignore these verifications.\r\n\r\nYou can regenerate the dataset_infos.json with\r\n```\r\ndatasets-cli test ./datasets/hindi_encorp --save_infos --all_configs --ignore_verifications\r\n```\r\n\r\n\r\n> I cloned the branch and it seems to work fine at my end. try to clear the cache -\r\n> \r\n> ```\r\n> rm -rf /home/ubuntu/.cache/huggingface/datasets/\r\n> rm -rf /home/ubuntu/.cache/huggingface/modules/datasets_modules//datasets/\r\n> ```\r\n> \r\n> But the dataset has only one record. Is that correct?\r\n> ![image](https://user-images.githubusercontent.com/9033954/102331376-c7929b00-3fb0-11eb-8a6c-81b2cf47bc2a.png)\r\n\r\nYes the current parsing is wrong, I've already given @rahul-art some suggestions and it looks like it works way better now (num_examples=273885).\r\n\r\nThanks for fixing the parsing @rahul-art !\r\nFeel free to commit and push your changes once it's ready :) ", "i ran the command you provided datasets-cli test ./datasets/hindi_encorp --save_infos --all_configs --ignore_verifications \r\nbut now its giving this error @lhoestq \r\n\r\nFileNotFoundError: Couldn't find file locally at ./datasets/hindi_encorp/hindi_encorp.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/master/datasets/./datasets/hindi_encorp/hindi_encorp.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/./datasets/hindi_encorp/hindi_encorp.py.\r\nIf the dataset was added recently, you may need to to pass script_version=\"master\" to find the loading script on the master branch.\r\n", "whoops I meant `hind_encorp` instead of `hindi_encorp` sorry", "@lhoestq all changes have done successfully in this PR #1584", "Ok thanks ! closing this one in favor of #1584 " ]
1,607,996,162,000
1,608,131,717,000
1,608,131,717,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1575", "html_url": "https://github.com/huggingface/datasets/pull/1575", "diff_url": "https://github.com/huggingface/datasets/pull/1575.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1575.patch", "merged_at": null }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1575/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1575/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1574
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1574/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1574/comments
https://api.github.com/repos/huggingface/datasets/issues/1574/events
https://github.com/huggingface/datasets/pull/1574
767,015,317
MDExOlB1bGxSZXF1ZXN0NTM5ODY1Mzcy
1,574
Diplomacy detection 3
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,988,531,000
1,607,988,572,000
1,607,988,572,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1574", "html_url": "https://github.com/huggingface/datasets/pull/1574", "diff_url": "https://github.com/huggingface/datasets/pull/1574.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1574.patch", "merged_at": null }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1574/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1574/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1573
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1573/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1573/comments
https://api.github.com/repos/huggingface/datasets/issues/1573/events
https://github.com/huggingface/datasets/pull/1573
767,011,938
MDExOlB1bGxSZXF1ZXN0NTM5ODYyNjcx
1,573
adding dataset for diplomacy detection-2
{ "login": "MisbahKhan789", "id": 15351802, "node_id": "MDQ6VXNlcjE1MzUxODAy", "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MisbahKhan789", "html_url": "https://github.com/MisbahKhan789", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,988,097,000
1,607,989,017,000
1,607,989,017,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1573", "html_url": "https://github.com/huggingface/datasets/pull/1573", "diff_url": "https://github.com/huggingface/datasets/pull/1573.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1573.patch", "merged_at": null }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1573/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1573/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1572
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1572/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1572/comments
https://api.github.com/repos/huggingface/datasets/issues/1572/events
https://github.com/huggingface/datasets/pull/1572
767,008,470
MDExOlB1bGxSZXF1ZXN0NTM5ODU5OTgx
1,572
add Gnad10 dataset
{ "login": "stevhliu", "id": 59462357, "node_id": "MDQ6VXNlcjU5NDYyMzU3", "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stevhliu", "html_url": "https://github.com/stevhliu", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "repos_url": "https://api.github.com/users/stevhliu/repos", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,987,702,000
1,631,897,677,000
1,608,137,550,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1572", "html_url": "https://github.com/huggingface/datasets/pull/1572", "diff_url": "https://github.com/huggingface/datasets/pull/1572.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1572.patch", "merged_at": 1608137550000 }
reference [PR#1317](https://github.com/huggingface/datasets/pull/1317)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1572/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1572/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1571
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1571/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1571/comments
https://api.github.com/repos/huggingface/datasets/issues/1571/events
https://github.com/huggingface/datasets/pull/1571
766,981,721
MDExOlB1bGxSZXF1ZXN0NTM5ODM5OTEw
1,571
Fixing the KILT tasks to match our current standards
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}", "starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yjernite/subscriptions", "organizations_url": "https://api.github.com/users/yjernite/orgs", "repos_url": "https://api.github.com/users/yjernite/repos", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "received_events_url": "https://api.github.com/users/yjernite/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,984,772,000
1,607,987,261,000
1,607,987,261,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1571", "html_url": "https://github.com/huggingface/datasets/pull/1571", "diff_url": "https://github.com/huggingface/datasets/pull/1571.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1571.patch", "merged_at": 1607987261000 }
This introduces a few changes to the Knowledge Intensive Learning task benchmark to bring it more in line with our current datasets, including adding the (minimal) dataset card and having one config per sub-task
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1571/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1571/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1570
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1570/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1570/comments
https://api.github.com/repos/huggingface/datasets/issues/1570/events
https://github.com/huggingface/datasets/pull/1570
766,830,545
MDExOlB1bGxSZXF1ZXN0NTM5NzM1MDY2
1,570
Documentation for loading CSV datasets misleads the user
{ "login": "onurgu", "id": 56893, "node_id": "MDQ6VXNlcjU2ODkz", "avatar_url": "https://avatars.githubusercontent.com/u/56893?v=4", "gravatar_id": "", "url": "https://api.github.com/users/onurgu", "html_url": "https://github.com/onurgu", "followers_url": "https://api.github.com/users/onurgu/followers", "following_url": "https://api.github.com/users/onurgu/following{/other_user}", "gists_url": "https://api.github.com/users/onurgu/gists{/gist_id}", "starred_url": "https://api.github.com/users/onurgu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/onurgu/subscriptions", "organizations_url": "https://api.github.com/users/onurgu/orgs", "repos_url": "https://api.github.com/users/onurgu/repos", "events_url": "https://api.github.com/users/onurgu/events{/privacy}", "received_events_url": "https://api.github.com/users/onurgu/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,972,677,000
1,608,665,412,000
1,608,558,429,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1570", "html_url": "https://github.com/huggingface/datasets/pull/1570", "diff_url": "https://github.com/huggingface/datasets/pull/1570.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1570.patch", "merged_at": 1608558429000 }
Documentation for loading CSV datasets misleads the user into thinking setting `quote_char' to False will disable quoting. There are two problems here: i) `quote_char' is misspelled, must be `quotechar' ii) the documentation should mention `quoting'
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1570/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1570/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1569
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1569/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1569/comments
https://api.github.com/repos/huggingface/datasets/issues/1569/events
https://github.com/huggingface/datasets/pull/1569
766,758,895
MDExOlB1bGxSZXF1ZXN0NTM5NjkwMjc2
1,569
added un_ga dataset
{ "login": "param087", "id": 26374564, "node_id": "MDQ6VXNlcjI2Mzc0NTY0", "avatar_url": "https://avatars.githubusercontent.com/u/26374564?v=4", "gravatar_id": "", "url": "https://api.github.com/users/param087", "html_url": "https://github.com/param087", "followers_url": "https://api.github.com/users/param087/followers", "following_url": "https://api.github.com/users/param087/following{/other_user}", "gists_url": "https://api.github.com/users/param087/gists{/gist_id}", "starred_url": "https://api.github.com/users/param087/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/param087/subscriptions", "organizations_url": "https://api.github.com/users/param087/orgs", "repos_url": "https://api.github.com/users/param087/repos", "events_url": "https://api.github.com/users/param087/events{/privacy}", "received_events_url": "https://api.github.com/users/param087/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,967,724,000
1,608,046,138,000
1,608,046,138,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1569", "html_url": "https://github.com/huggingface/datasets/pull/1569", "diff_url": "https://github.com/huggingface/datasets/pull/1569.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1569.patch", "merged_at": 1608046138000 }
Hi :hugs:, This is a PR for [United nations general assembly resolutions: A six-language parallel corpus](http://opus.nlpl.eu/UN.php) dataset. With suggested changes in #1330
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1569/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1569/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1568
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1568/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1568/comments
https://api.github.com/repos/huggingface/datasets/issues/1568/events
https://github.com/huggingface/datasets/pull/1568
766,722,994
MDExOlB1bGxSZXF1ZXN0NTM5NjY2ODg1
1,568
Added the dataset clickbait_news_bg
{ "login": "tsvm", "id": 1083319, "node_id": "MDQ6VXNlcjEwODMzMTk=", "avatar_url": "https://avatars.githubusercontent.com/u/1083319?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tsvm", "html_url": "https://github.com/tsvm", "followers_url": "https://api.github.com/users/tsvm/followers", "following_url": "https://api.github.com/users/tsvm/following{/other_user}", "gists_url": "https://api.github.com/users/tsvm/gists{/gist_id}", "starred_url": "https://api.github.com/users/tsvm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tsvm/subscriptions", "organizations_url": "https://api.github.com/users/tsvm/orgs", "repos_url": "https://api.github.com/users/tsvm/repos", "events_url": "https://api.github.com/users/tsvm/events{/privacy}", "received_events_url": "https://api.github.com/users/tsvm/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi @tsvm Great work! \r\nSince you have raised a clean PR could you close the earlier one - #1445 ? \r\n", "> Hi @tsvm Great work!\r\n> Since you have raised a clean PR could you close the earlier one - #1445 ?\r\n\r\nDone." ]
1,607,965,380,000
1,608,056,936,000
1,608,056,936,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1568", "html_url": "https://github.com/huggingface/datasets/pull/1568", "diff_url": "https://github.com/huggingface/datasets/pull/1568.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1568.patch", "merged_at": 1608056936000 }
There was a problem with my [previous PR 1445](https://github.com/huggingface/datasets/pull/1445) after rebasing, so I'm copying the dataset code into a new branch and submitting a new PR.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1568/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1568/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1567
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1567/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1567/comments
https://api.github.com/repos/huggingface/datasets/issues/1567/events
https://github.com/huggingface/datasets/pull/1567
766,382,609
MDExOlB1bGxSZXF1ZXN0NTM5NDE3NzI5
1,567
[wording] Update Readme.md
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "https://api.github.com/users/thomwolf/gists{/gist_id}", "starred_url": "https://api.github.com/users/thomwolf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomwolf/subscriptions", "organizations_url": "https://api.github.com/users/thomwolf/orgs", "repos_url": "https://api.github.com/users/thomwolf/repos", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "received_events_url": "https://api.github.com/users/thomwolf/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,949,292,000
1,608,036,847,000
1,608,036,846,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1567", "html_url": "https://github.com/huggingface/datasets/pull/1567", "diff_url": "https://github.com/huggingface/datasets/pull/1567.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1567.patch", "merged_at": 1608036846000 }
Make the features of the library clearer.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1567/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1567/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1566
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1566/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1566/comments
https://api.github.com/repos/huggingface/datasets/issues/1566/events
https://github.com/huggingface/datasets/pull/1566
766,354,236
MDExOlB1bGxSZXF1ZXN0NTM5Mzk5NTg4
1,566
Add Microsoft Research Sequential Question Answering (SQA) Dataset
{ "login": "mattbui", "id": 46804938, "node_id": "MDQ6VXNlcjQ2ODA0OTM4", "avatar_url": "https://avatars.githubusercontent.com/u/46804938?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mattbui", "html_url": "https://github.com/mattbui", "followers_url": "https://api.github.com/users/mattbui/followers", "following_url": "https://api.github.com/users/mattbui/following{/other_user}", "gists_url": "https://api.github.com/users/mattbui/gists{/gist_id}", "starred_url": "https://api.github.com/users/mattbui/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mattbui/subscriptions", "organizations_url": "https://api.github.com/users/mattbui/orgs", "repos_url": "https://api.github.com/users/mattbui/repos", "events_url": "https://api.github.com/users/mattbui/events{/privacy}", "received_events_url": "https://api.github.com/users/mattbui/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I proposed something a few weeks ago in #898 (un-merged) but I think that the way that @mattbui added the dataset in the present PR is smarter and simpler should replace my PR #898.\r\n\r\n(Narrator voice: *And it was around that time that Thomas realized that the community was now a lot smarter than him and he should hand-over the library he had started with @lhoestq to the community and stop pretending he knew everything about it.*)" ]
1,607,947,350,000
1,608,045,862,000
1,608,045,862,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1566", "html_url": "https://github.com/huggingface/datasets/pull/1566", "diff_url": "https://github.com/huggingface/datasets/pull/1566.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1566.patch", "merged_at": 1608045862000 }
For more information: https://msropendata.com/datasets/b25190ed-0f59-47b1-9211-5962858142c2
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1566/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1566/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1565
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1565/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1565/comments
https://api.github.com/repos/huggingface/datasets/issues/1565/events
https://github.com/huggingface/datasets/pull/1565
766,333,940
MDExOlB1bGxSZXF1ZXN0NTM5Mzg2MzEx
1,565
Create README.md
{ "login": "ManuelFay", "id": 43467008, "node_id": "MDQ6VXNlcjQzNDY3MDA4", "avatar_url": "https://avatars.githubusercontent.com/u/43467008?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ManuelFay", "html_url": "https://github.com/ManuelFay", "followers_url": "https://api.github.com/users/ManuelFay/followers", "following_url": "https://api.github.com/users/ManuelFay/following{/other_user}", "gists_url": "https://api.github.com/users/ManuelFay/gists{/gist_id}", "starred_url": "https://api.github.com/users/ManuelFay/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ManuelFay/subscriptions", "organizations_url": "https://api.github.com/users/ManuelFay/orgs", "repos_url": "https://api.github.com/users/ManuelFay/repos", "events_url": "https://api.github.com/users/ManuelFay/events{/privacy}", "received_events_url": "https://api.github.com/users/ManuelFay/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "@ManuelFay thanks you so much for adding a dataset card, this is such a cool contribution!\r\n\r\nThis looks like it uses an old template for the card we've moved things around a bit and we have an app you should be using to get the tags and the structure of the Data Fields paragraph :) Would you mind moving your text to the newer format (we're also asking contributors to keep the full template structure, even if some sections still have [More Information Needed] for the time being)\r\n\r\nHere's the link to the instructions:\r\nhttps://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md#tag-the-dataset-and-write-the-dataset-card\r\n\r\nOut of curiosity, what was your landing point for filling out the card? Did you follow the \"Update on Github\" when navigating the datasets? Trying to make the instructions as clear as possible :) ", "@yjernite \r\n\r\nPerfect, I'll follow the instructions when I have a bit more time tomorrow ! I was actually browsing the new contributions after the dataset sprint and realized most of the \"old\" datasets were not tagged, so I just copied and pasted the readme from another dataset and was not aware there was precise instructions... Will fix !\r\n\r\nBTW, amazing job with the retriBert work, I used the contrastive + in-batch negative quite a bit for various projects. Probably neither the time nor place to talk about that but I was curious as to why, in your original work, you prefered using a simple projection in the last layer to differentiate the question vs answer embedding, rather than allowing for bias in the dense layer or even just to fine-tune 2 different embedders for question + answer ? ", "Cool! Looking forward to the next version!\r\n\r\nQuick answer for retriBERT is that I expected a simple projection to generalize better and more importantly only having to store the gradients for the proj means training with larger batches :) If you want to keep chatting about it, feel free to send me an email!", "Hi @ManuelFay ! \r\nIf you're still interested in completing the FQuAD dataset card, note that we've generated one that is pre-filled.\r\nTherefore feel free to complete it with the content you already have in your README.md.\r\nThis would be awesome ! And thanks again for your contribution :)", "Yo @lhoestq , just not sure about the tag table at the top, I used @yjernite eli5 template so hope it's okay ! Also want to signal the streamlit app for dataset tagging has a weird behavior with the size categories when filling in the form. \r\n\r\nThanks to you guys for doing that and sorry about the time it took, i completely forgot about it ! \r\n" ]
1,607,946,023,000
1,616,680,909,000
1,616,680,909,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1565", "html_url": "https://github.com/huggingface/datasets/pull/1565", "diff_url": "https://github.com/huggingface/datasets/pull/1565.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1565.patch", "merged_at": 1616680909000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1565/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1565/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1564
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1564/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1564/comments
https://api.github.com/repos/huggingface/datasets/issues/1564/events
https://github.com/huggingface/datasets/pull/1564
766,266,609
MDExOlB1bGxSZXF1ZXN0NTM5MzQzMjAy
1,564
added saudinewsnet
{ "login": "abdulelahsm", "id": 28743265, "node_id": "MDQ6VXNlcjI4NzQzMjY1", "avatar_url": "https://avatars.githubusercontent.com/u/28743265?v=4", "gravatar_id": "", "url": "https://api.github.com/users/abdulelahsm", "html_url": "https://github.com/abdulelahsm", "followers_url": "https://api.github.com/users/abdulelahsm/followers", "following_url": "https://api.github.com/users/abdulelahsm/following{/other_user}", "gists_url": "https://api.github.com/users/abdulelahsm/gists{/gist_id}", "starred_url": "https://api.github.com/users/abdulelahsm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/abdulelahsm/subscriptions", "organizations_url": "https://api.github.com/users/abdulelahsm/orgs", "repos_url": "https://api.github.com/users/abdulelahsm/repos", "events_url": "https://api.github.com/users/abdulelahsm/events{/privacy}", "received_events_url": "https://api.github.com/users/abdulelahsm/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi @abdulelahsm - This is an interesting dataset! But there are multiple issues with the PR. Some of them are listed below: \r\n- default builder config is not defined. There should be atleast one builder config \r\n- URL is incorrectly constructed so the data files are not being downloaded \r\n- dataset_info.json file was not created\r\n\r\nPlease have a look at some existing merged datasets to get a reference on building the data loader. If you are still stuck, reach out. \r\n", "@skyprince999 I totally agree. Thx for the feedback!", "Hi @abdulelahsm ! Thanks for adding this one :) \r\nyou don't actually have to add builder configurations if you don't need them. It's fine as it is now\r\n\r\nAnd as @skyprince999 noticed, the current URLs don't work. to download files.\r\nYou can use this one for example for the first batch instead:\r\nhttps://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-21.zip\r\n\r\nFeel free to ping me if you have questions or if you're ready for a review :) ", "@lhoestq Hey, I tried using the first batch instead, the data was downloaded but I got this error, not sure why it can't find the path?\r\n\r\nfor content, I ran ``` \"./datasets/saudinewsnet/test.py\"```\r\n\r\nwhich is a local test I'm running for the dataset, it contains the following code\r\n\r\n```\r\nfrom datasets import load_dataset\r\n\r\ndata = load_dataset(\"./datasets/saudinewsnet\", split= \"train\")\r\n\r\nprint(data)\r\n\r\nprint(data[1])\r\n```\r\n\r\nthis is the error I got \r\n\r\n```\r\n2020-12-18 21:45:39.403908: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcudart.so.10.1'; dlerror: libcudart.so.10.1: cannot open shared object file: No such file or directory\r\n2020-12-18 21:45:39.403953: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.\r\nUsing custom data configuration default\r\nDownloading and preparing dataset saudi_news_net/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /home/mesfas/.cache/huggingface/datasets/saudi_news_net/default/0.0.0/62ece5ef0a991415352d4b1efac681d75b5b3404064fd4f6a1d659499dab18f4...\r\nDownloading: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3.42M/3.42M [00:03<00:00, 1.03MB/s]\r\nTraceback (most recent call last):\r\n File \"/home/mesfas/opensource/datasets/src/datasets/builder.py\", line 604, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n File \"/home/mesfas/opensource/datasets/src/datasets/builder.py\", line 902, in _prepare_split\r\n for key, record in utils.tqdm(\r\n File \"/home/mesfas/environments/ar_res_reviews/lib/python3.8/site-packages/tqdm/std.py\", line 1133, in __iter__\r\n for obj in iterable:\r\n File \"/home/mesfas/.cache/huggingface/modules/datasets_modules/datasets/saudinewsnet/62ece5ef0a991415352d4b1efac681d75b5b3404064fd4f6a1d659499dab18f4/saudinewsnet.py\", line 108, in _generate_examples\r\n with open(filepath, encoding=\"utf-8\").read() as f:\r\nIsADirectoryError: [Errno 21] Is a directory: '/home/mesfas/.cache/huggingface/datasets/downloads/extracted/314fd983aa07d3dada9429911a805270c3285f48759d3584a1343c2d86260765'\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"./datasets/saudinewsnet/test.py\", line 3, in <module>\r\n data = load_dataset(\"./datasets/saudinewsnet\", split= \"train\")\r\n File \"/home/mesfas/opensource/datasets/src/datasets/load.py\", line 607, in load_dataset\r\n builder_instance.download_and_prepare(\r\n File \"/home/mesfas/opensource/datasets/src/datasets/builder.py\", line 526, in download_and_prepare\r\n self._download_and_prepare(\r\n File \"/home/mesfas/opensource/datasets/src/datasets/builder.py\", line 606, in _download_and_prepare\r\n raise OSError(\r\nOSError: Cannot find data file. \r\nOriginal error:\r\n[Errno 21] Is a directory: '/home/mesfas/.cache/huggingface/datasets/downloads/extracted/314fd983aa07d3dada9429911a805270c3285f48759d3584a1343c2d86260765'\r\n```\r\n\r\n\r\nthis is the split code \r\n\r\n```\r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n my_urls = _URL\r\n datadir = dl_manager.download_and_extract(my_urls)\r\n return [\r\n datasets.SplitGenerator(\r\n name=datasets.Split.TRAIN,\r\n # These kwargs will be passed to _generate_examples\r\n gen_kwargs={\r\n \"filepath\": datadir,\r\n \"split\": \"train\"\r\n },\r\n ),\r\n ]\r\n```\r\nand this is how I'm generating the examples\r\n\r\n```\r\n def _generate_examples(self, filepath, split):\r\n \r\n #logging.info(\"generating examples from = %s\", filepath)\r\n with open(filepath, encoding=\"utf-8\") as f:\r\n articles = json.load(f)\r\n for article in articles:\r\n title = article.get(\"title\", \"\").strip()\r\n source = article.get(\"source\", \"\").strip()\r\n date = article.get(\"date_extracted\", \"\").strip()\r\n link = article.get(\"url\", \"\").strip()\r\n author = article.get(\"author\", \"\").strip()\r\n content = article.get(\"content\", \"\").strip()\r\n\r\n yield id_, {\r\n \"title\": title,\r\n \"source\": source,\r\n \"date\": date,\r\n \"link\": link,\r\n \"author\": author,\r\n \"content\": content\r\n }\r\n```", "What's `_URL` ?\r\n\r\nIt looks like you are downloading an archive.\r\nTherefore you may need to get to the file path using `filepath = os.path.join(datadir, \"actual_file_name_inside_the_downloaded_archive\")`", "@lhoestq you were 100% right. Thank you. All fixed", "@lhoestq ping!", "@lhoestq added the remaining 17 batches and modified the readme.md to reflect that + resolved the camel case comment", "merging since the CI is fixed on master" ]
1,607,942,109,000
1,608,630,664,000
1,608,630,664,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1564", "html_url": "https://github.com/huggingface/datasets/pull/1564", "diff_url": "https://github.com/huggingface/datasets/pull/1564.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1564.patch", "merged_at": 1608630664000 }
I'm having issues in creating the dummy data. I'm still investigating how to fix it. I'll close the PR if I couldn't find a solution
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1564/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1564/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1563
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1563/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1563/comments
https://api.github.com/repos/huggingface/datasets/issues/1563/events
https://github.com/huggingface/datasets/pull/1563
766,211,931
MDExOlB1bGxSZXF1ZXN0NTM5MzA4Mzg4
1,563
adding tmu-gfm-dataset
{ "login": "forest1988", "id": 2755894, "node_id": "MDQ6VXNlcjI3NTU4OTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2755894?v=4", "gravatar_id": "", "url": "https://api.github.com/users/forest1988", "html_url": "https://github.com/forest1988", "followers_url": "https://api.github.com/users/forest1988/followers", "following_url": "https://api.github.com/users/forest1988/following{/other_user}", "gists_url": "https://api.github.com/users/forest1988/gists{/gist_id}", "starred_url": "https://api.github.com/users/forest1988/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/forest1988/subscriptions", "organizations_url": "https://api.github.com/users/forest1988/orgs", "repos_url": "https://api.github.com/users/forest1988/repos", "events_url": "https://api.github.com/users/forest1988/events{/privacy}", "received_events_url": "https://api.github.com/users/forest1988/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "@lhoestq Thank you for your code review! I think I could do the necessary corrections. Could you please check it again when you have time?", "Thank you for merging!" ]
1,607,939,130,000
1,608,546,064,000
1,608,545,233,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1563", "html_url": "https://github.com/huggingface/datasets/pull/1563", "diff_url": "https://github.com/huggingface/datasets/pull/1563.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1563.patch", "merged_at": 1608545233000 }
Adding TMU-GFM-Dataset for Grammatical Error Correction. https://github.com/tmu-nlp/TMU-GFM-Dataset A dataset for GEC metrics with manual evaluations of grammaticality, fluency, and meaning preservation for system outputs. More detail about the creation of the dataset can be found in [Yoshimura et al. (2020)](https://www.aclweb.org/anthology/2020.coling-main.573.pdf).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1563/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1563/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1562
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1562/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1562/comments
https://api.github.com/repos/huggingface/datasets/issues/1562/events
https://github.com/huggingface/datasets/pull/1562
765,981,749
MDExOlB1bGxSZXF1ZXN0NTM5MTc5ODc3
1,562
Add dataset COrpus of Urdu News TExt Reuse (COUNTER).
{ "login": "arkhalid", "id": 14899066, "node_id": "MDQ6VXNlcjE0ODk5MDY2", "avatar_url": "https://avatars.githubusercontent.com/u/14899066?v=4", "gravatar_id": "", "url": "https://api.github.com/users/arkhalid", "html_url": "https://github.com/arkhalid", "followers_url": "https://api.github.com/users/arkhalid/followers", "following_url": "https://api.github.com/users/arkhalid/following{/other_user}", "gists_url": "https://api.github.com/users/arkhalid/gists{/gist_id}", "starred_url": "https://api.github.com/users/arkhalid/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/arkhalid/subscriptions", "organizations_url": "https://api.github.com/users/arkhalid/orgs", "repos_url": "https://api.github.com/users/arkhalid/repos", "events_url": "https://api.github.com/users/arkhalid/events{/privacy}", "received_events_url": "https://api.github.com/users/arkhalid/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Just a small revision from simon's review: 20KB for the dummy_data.zip is fine, you can keep them this way.", "Also the CI is failing because of an error `tests/test_file_utils.py::TempSeedTest::test_tensorflow` that is not related to your dataset and is fixed on master. You can ignore it", "merging since the Ci is fixed on master" ]
1,607,927,568,000
1,608,556,486,000
1,608,556,486,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1562", "html_url": "https://github.com/huggingface/datasets/pull/1562", "diff_url": "https://github.com/huggingface/datasets/pull/1562.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1562.patch", "merged_at": 1608556486000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1562/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1562/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1561
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1561/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1561/comments
https://api.github.com/repos/huggingface/datasets/issues/1561/events
https://github.com/huggingface/datasets/pull/1561
765,831,436
MDExOlB1bGxSZXF1ZXN0NTM5MTAwNjAy
1,561
Lama
{ "login": "ontocord", "id": 8900094, "node_id": "MDQ6VXNlcjg5MDAwOTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8900094?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ontocord", "html_url": "https://github.com/ontocord", "followers_url": "https://api.github.com/users/ontocord/followers", "following_url": "https://api.github.com/users/ontocord/following{/other_user}", "gists_url": "https://api.github.com/users/ontocord/gists{/gist_id}", "starred_url": "https://api.github.com/users/ontocord/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ontocord/subscriptions", "organizations_url": "https://api.github.com/users/ontocord/orgs", "repos_url": "https://api.github.com/users/ontocord/repos", "events_url": "https://api.github.com/users/ontocord/events{/privacy}", "received_events_url": "https://api.github.com/users/ontocord/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Let me know why the pyarrow test is failing. For one of the config \"trex\", I had to load an initial datafile for a dictionary which is used to augment the rest of the datasets. In the dummy data, the dictionary file was truncated so I had to fudge that. I'm not sure if that is the issue.\r\n", "@ontocord it just needs a rerun and it will be good to go.", "THanks @tanmoyio. How do I do a rerun?", "@ontocord contributor can’t rerun it, the maintainers will rerun it, it may take lil bit of time as there are so many PRs left to be reviewed and merged ", "@lhoestq not sure why it is failing. i've made all modifications. ", "merging since the CI is fixed on master" ]
1,607,916,430,000
1,609,149,107,000
1,609,149,107,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1561", "html_url": "https://github.com/huggingface/datasets/pull/1561", "diff_url": "https://github.com/huggingface/datasets/pull/1561.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1561.patch", "merged_at": 1609149107000 }
This the LAMA dataset for probing facts and common sense from language models. See https://github.com/facebookresearch/LAMA for more details.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1561/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1561/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/1560
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1560/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1560/comments
https://api.github.com/repos/huggingface/datasets/issues/1560/events
https://github.com/huggingface/datasets/pull/1560
765,814,964
MDExOlB1bGxSZXF1ZXN0NTM5MDkzMzky
1,560
Adding the BrWaC dataset
{ "login": "jonatasgrosman", "id": 5097052, "node_id": "MDQ6VXNlcjUwOTcwNTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5097052?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonatasgrosman", "html_url": "https://github.com/jonatasgrosman", "followers_url": "https://api.github.com/users/jonatasgrosman/followers", "following_url": "https://api.github.com/users/jonatasgrosman/following{/other_user}", "gists_url": "https://api.github.com/users/jonatasgrosman/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonatasgrosman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonatasgrosman/subscriptions", "organizations_url": "https://api.github.com/users/jonatasgrosman/orgs", "repos_url": "https://api.github.com/users/jonatasgrosman/repos", "events_url": "https://api.github.com/users/jonatasgrosman/events{/privacy}", "received_events_url": "https://api.github.com/users/jonatasgrosman/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,607,915,036,000
1,608,307,016,000
1,608,307,015,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/1560", "html_url": "https://github.com/huggingface/datasets/pull/1560", "diff_url": "https://github.com/huggingface/datasets/pull/1560.diff", "patch_url": "https://github.com/huggingface/datasets/pull/1560.patch", "merged_at": 1608307015000 }
Adding the BrWaC dataset, a large corpus of Portuguese language texts
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1560/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1560/timeline
null
true