url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
2.12B
node_id
stringlengths
18
32
number
int64
1
6.65k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
int64
0
70
created_at
unknown
updated_at
unknown
closed_at
unknown
author_association
stringclasses
3 values
active_lock_reason
float64
draft
float64
0
1
pull_request
dict
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
float64
state_reason
stringclasses
3 values
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/6342
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6342/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6342/comments
https://api.github.com/repos/huggingface/datasets/issues/6342/events
https://github.com/huggingface/datasets/pull/6342
1,957,344,445
PR_kwDODunzps5dijxt
6,342
Release: 2.14.6
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
5
"2023-10-23T14:43:26Z"
"2023-10-23T15:21:54Z"
"2023-10-23T15:07:25Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6342.diff", "html_url": "https://github.com/huggingface/datasets/pull/6342", "merged_at": "2023-10-23T15:07:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/6342.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6342" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6342/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6342/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6340
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6340/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6340/comments
https://api.github.com/repos/huggingface/datasets/issues/6340/events
https://github.com/huggingface/datasets/pull/6340
1,956,917,893
PR_kwDODunzps5dhGpW
6,340
Release 2.14.5
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
1
"2023-10-23T11:10:22Z"
"2023-10-23T14:20:46Z"
"2023-10-23T11:12:40Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6340.diff", "html_url": "https://github.com/huggingface/datasets/pull/6340", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6340.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6340" }
(wrong release number - I was continuing the 2.14 branch but 2.14.5 was released from `main`)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6340/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6340/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6339
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6339/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6339/comments
https://api.github.com/repos/huggingface/datasets/issues/6339/events
https://github.com/huggingface/datasets/pull/6339
1,956,912,627
PR_kwDODunzps5dhFfg
6,339
minor release step improvement
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
3
"2023-10-23T11:07:04Z"
"2023-11-07T10:38:54Z"
"2023-11-07T10:32:41Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6339.diff", "html_url": "https://github.com/huggingface/datasets/pull/6339", "merged_at": "2023-11-07T10:32:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/6339.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6339" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6339/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6339/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6338
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6338/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6338/comments
https://api.github.com/repos/huggingface/datasets/issues/6338/events
https://github.com/huggingface/datasets/pull/6338
1,956,886,072
PR_kwDODunzps5dg_sb
6,338
pin fsspec before it switches to glob.glob
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
2
"2023-10-23T10:50:54Z"
"2024-01-11T06:32:56Z"
"2023-10-23T10:51:52Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6338.diff", "html_url": "https://github.com/huggingface/datasets/pull/6338", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6338.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6338" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6338/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6338/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6337
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6337/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6337/comments
https://api.github.com/repos/huggingface/datasets/issues/6337/events
https://github.com/huggingface/datasets/pull/6337
1,956,875,259
PR_kwDODunzps5dg9Uu
6,337
Pin supported upper version of fsspec
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
6
"2023-10-23T10:44:16Z"
"2023-10-23T12:13:20Z"
"2023-10-23T12:04:36Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6337.diff", "html_url": "https://github.com/huggingface/datasets/pull/6337", "merged_at": "2023-10-23T12:04:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/6337.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6337" }
Pin upper version of `fsspec` to avoid disruptions introduced by breaking changes (and the need of urgent patch releases with hotfixes) on each release on their side. See: - #6331 - #6210 - #5731 - #5617 - #5447 I propose that we explicitly test, introduce fixes and support each new `fsspec` version release. CC: @LysandreJik
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6337/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6337/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6336
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6336/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6336/comments
https://api.github.com/repos/huggingface/datasets/issues/6336/events
https://github.com/huggingface/datasets/pull/6336
1,956,827,232
PR_kwDODunzps5dgy0w
6,336
unpin-fsspec
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
3
"2023-10-23T10:16:46Z"
"2024-02-07T12:41:35Z"
"2023-10-23T10:17:48Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6336.diff", "html_url": "https://github.com/huggingface/datasets/pull/6336", "merged_at": "2023-10-23T10:17:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/6336.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6336" }
Close #6333.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6336/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6336/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6335
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6335/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6335/comments
https://api.github.com/repos/huggingface/datasets/issues/6335/events
https://github.com/huggingface/datasets/pull/6335
1,956,740,818
PR_kwDODunzps5dggIV
6,335
Support fsspec 2023.10.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
7
"2023-10-23T09:29:17Z"
"2024-01-11T06:33:35Z"
"2023-11-14T14:17:40Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6335.diff", "html_url": "https://github.com/huggingface/datasets/pull/6335", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6335.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6335" }
Fix #6333.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6335/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6335/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6334
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6334/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6334/comments
https://api.github.com/repos/huggingface/datasets/issues/6334/events
https://github.com/huggingface/datasets/pull/6334
1,956,719,774
PR_kwDODunzps5dgbpR
6,334
datasets.filesystems: fix is_remote_filesystems
{ "avatar_url": "https://avatars.githubusercontent.com/u/1463443?v=4", "events_url": "https://api.github.com/users/ap--/events{/privacy}", "followers_url": "https://api.github.com/users/ap--/followers", "following_url": "https://api.github.com/users/ap--/following{/other_user}", "gists_url": "https://api.github.com/users/ap--/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ap--", "id": 1463443, "login": "ap--", "node_id": "MDQ6VXNlcjE0NjM0NDM=", "organizations_url": "https://api.github.com/users/ap--/orgs", "received_events_url": "https://api.github.com/users/ap--/received_events", "repos_url": "https://api.github.com/users/ap--/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ap--/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ap--/subscriptions", "type": "User", "url": "https://api.github.com/users/ap--" }
[]
closed
false
null
[]
null
3
"2023-10-23T09:17:54Z"
"2024-02-07T12:41:15Z"
"2023-10-23T10:14:10Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6334.diff", "html_url": "https://github.com/huggingface/datasets/pull/6334", "merged_at": "2023-10-23T10:14:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/6334.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6334" }
Close #6330, close #6333. `fsspec.implementations.LocalFilesystem.protocol` was changed from `str` "file" to `tuple[str,...]` ("file", "local") in `fsspec>=2023.10.0` This commit supports both styles.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/6334/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6334/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6333
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6333/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6333/comments
https://api.github.com/repos/huggingface/datasets/issues/6333/events
https://github.com/huggingface/datasets/issues/6333
1,956,714,423
I_kwDODunzps50oRe3
6,333
Support fsspec 2023.10.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
4
"2023-10-23T09:14:53Z"
"2024-02-07T12:39:58Z"
"2024-02-07T12:39:58Z"
MEMBER
null
null
null
Once root issue is fixed, remove temporary pin of fsspec < 2023.10.0 introduced by: - #6331 Related to issue: - #6330 As @ZachNagengast suggested, the issue might be related to: - https://github.com/fsspec/filesystem_spec/pull/1381
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6333/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6333/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6332
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6332/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6332/comments
https://api.github.com/repos/huggingface/datasets/issues/6332/events
https://github.com/huggingface/datasets/pull/6332
1,956,697,328
PR_kwDODunzps5dgW3w
6,332
Replace deprecated license_file in setup.cfg
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
4
"2023-10-23T09:05:26Z"
"2023-11-07T08:23:10Z"
"2023-11-07T08:09:06Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6332.diff", "html_url": "https://github.com/huggingface/datasets/pull/6332", "merged_at": "2023-11-07T08:09:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/6332.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6332" }
Replace deprecated license_file in `setup.cfg`. See: https://github.com/huggingface/datasets/actions/runs/6610930650/job/17953825724?pr=6331 ``` /tmp/pip-build-env-a51hls20/overlay/lib/python3.8/site-packages/setuptools/config/setupcfg.py:293: _DeprecatedConfig: Deprecated config in `setup.cfg` !! ******************************************************************************** The license_file parameter is deprecated, use license_files instead. By 2023-Oct-30, you need to update your project and remove deprecated calls or your builds will no longer be supported. See https://setuptools.pypa.io/en/latest/userguide/declarative_config.html for details. ******************************************************************************** !! ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6332/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6332/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6331
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6331/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6331/comments
https://api.github.com/repos/huggingface/datasets/issues/6331/events
https://github.com/huggingface/datasets/pull/6331
1,956,671,256
PR_kwDODunzps5dgRQt
6,331
Temporarily pin fsspec < 2023.10.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
3
"2023-10-23T08:51:50Z"
"2023-10-23T09:26:42Z"
"2023-10-23T09:17:55Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6331.diff", "html_url": "https://github.com/huggingface/datasets/pull/6331", "merged_at": "2023-10-23T09:17:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/6331.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6331" }
Temporarily pin fsspec < 2023.10.0 until permanent solution is found. Hot fix #6330. See: https://github.com/huggingface/datasets/actions/runs/6610904287/job/17953774987 ``` ... ERROR tests/test_iterable_dataset.py::test_iterable_dataset_from_file - NotImplementedError: Loading a dataset cached in a LocalFileSystem is not supported. = 373 failed, 2055 passed, 17 skipped, 8 warnings, 6 errors in 228.14s (0:03:48) = ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6331/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6331/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6330
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6330/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6330/comments
https://api.github.com/repos/huggingface/datasets/issues/6330/events
https://github.com/huggingface/datasets/issues/6330
1,956,053,294
I_kwDODunzps50lwEu
6,330
Latest fsspec==2023.10.0 issue with streaming datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/1981179?v=4", "events_url": "https://api.github.com/users/ZachNagengast/events{/privacy}", "followers_url": "https://api.github.com/users/ZachNagengast/followers", "following_url": "https://api.github.com/users/ZachNagengast/following{/other_user}", "gists_url": "https://api.github.com/users/ZachNagengast/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZachNagengast", "id": 1981179, "login": "ZachNagengast", "node_id": "MDQ6VXNlcjE5ODExNzk=", "organizations_url": "https://api.github.com/users/ZachNagengast/orgs", "received_events_url": "https://api.github.com/users/ZachNagengast/received_events", "repos_url": "https://api.github.com/users/ZachNagengast/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZachNagengast/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZachNagengast/subscriptions", "type": "User", "url": "https://api.github.com/users/ZachNagengast" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
6
"2023-10-22T20:57:10Z"
"2023-11-07T10:02:14Z"
"2023-10-23T09:17:56Z"
CONTRIBUTOR
null
null
null
### Describe the bug Loading a streaming dataset with this version of fsspec fails with the following error: `NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.` I suspect the issue is with this PR https://github.com/fsspec/filesystem_spec/pull/1381 ### Steps to reproduce the bug 1. Upgrade fsspec to version `2023.10.0` 2. Attempt to load a streaming dataset e.g. `load_dataset("laion/gpt4v-emotion-dataset", split="train", streaming=True)` 3. Observe the following exception: ``` File "/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/site-packages/datasets/load.py", line 2146, in load_dataset return builder_instance.as_streaming_dataset(split=split) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/site-packages/datasets/builder.py", line 1318, in as_streaming_dataset raise NotImplementedError( NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet. ``` ### Expected behavior Should stream the dataset as normal. ### Environment info datasets@main fsspec==2023.10.0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6330/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6330/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6329
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6329/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6329/comments
https://api.github.com/repos/huggingface/datasets/issues/6329/events
https://github.com/huggingface/datasets/issues/6329
1,955,858,020
I_kwDODunzps50lAZk
6,329
شبکه های متن به گفتار ابتدا متن داده شده را به بازنمایی میانی
{ "avatar_url": "https://avatars.githubusercontent.com/u/147399213?v=4", "events_url": "https://api.github.com/users/shabnam706/events{/privacy}", "followers_url": "https://api.github.com/users/shabnam706/followers", "following_url": "https://api.github.com/users/shabnam706/following{/other_user}", "gists_url": "https://api.github.com/users/shabnam706/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shabnam706", "id": 147399213, "login": "shabnam706", "node_id": "U_kgDOCMkiLQ", "organizations_url": "https://api.github.com/users/shabnam706/orgs", "received_events_url": "https://api.github.com/users/shabnam706/received_events", "repos_url": "https://api.github.com/users/shabnam706/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shabnam706/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shabnam706/subscriptions", "type": "User", "url": "https://api.github.com/users/shabnam706" }
[]
closed
false
null
[]
null
0
"2023-10-22T11:07:46Z"
"2023-10-23T09:22:58Z"
"2023-10-23T09:22:58Z"
NONE
null
null
null
شبکه های متن به گفتار ابتدا متن داده شده را به بازنمایی میانی
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6329/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6329/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6328
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6328/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6328/comments
https://api.github.com/repos/huggingface/datasets/issues/6328/events
https://github.com/huggingface/datasets/issues/6328
1,955,857,904
I_kwDODunzps50lAXw
6,328
شبکه های متن به گفتار ابتدا متن داده شده را به بازنمایی میانی
{ "avatar_url": "https://avatars.githubusercontent.com/u/147399213?v=4", "events_url": "https://api.github.com/users/shabnam706/events{/privacy}", "followers_url": "https://api.github.com/users/shabnam706/followers", "following_url": "https://api.github.com/users/shabnam706/following{/other_user}", "gists_url": "https://api.github.com/users/shabnam706/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shabnam706", "id": 147399213, "login": "shabnam706", "node_id": "U_kgDOCMkiLQ", "organizations_url": "https://api.github.com/users/shabnam706/orgs", "received_events_url": "https://api.github.com/users/shabnam706/received_events", "repos_url": "https://api.github.com/users/shabnam706/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shabnam706/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shabnam706/subscriptions", "type": "User", "url": "https://api.github.com/users/shabnam706" }
[]
closed
false
null
[]
null
1
"2023-10-22T11:07:21Z"
"2023-10-23T09:22:38Z"
"2023-10-23T09:22:38Z"
NONE
null
null
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6328/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6328/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6327
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6327/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6327/comments
https://api.github.com/repos/huggingface/datasets/issues/6327/events
https://github.com/huggingface/datasets/issues/6327
1,955,470,755
I_kwDODunzps50jh2j
6,327
FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)`
{ "avatar_url": "https://avatars.githubusercontent.com/u/18402347?v=4", "events_url": "https://api.github.com/users/yzhangcs/events{/privacy}", "followers_url": "https://api.github.com/users/yzhangcs/followers", "following_url": "https://api.github.com/users/yzhangcs/following{/other_user}", "gists_url": "https://api.github.com/users/yzhangcs/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yzhangcs", "id": 18402347, "login": "yzhangcs", "node_id": "MDQ6VXNlcjE4NDAyMzQ3", "organizations_url": "https://api.github.com/users/yzhangcs/orgs", "received_events_url": "https://api.github.com/users/yzhangcs/received_events", "repos_url": "https://api.github.com/users/yzhangcs/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yzhangcs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yzhangcs/subscriptions", "type": "User", "url": "https://api.github.com/users/yzhangcs" }
[]
closed
false
null
[]
null
3
"2023-10-21T12:27:03Z"
"2023-10-23T18:50:07Z"
"2023-10-23T18:50:07Z"
NONE
null
null
null
### Describe the bug Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs. ### Steps to reproduce the bug I've downloaded the dataset and save it to the cache dir in advance. My hope is loading the files in offline environment and without taking too much hours to prepross the entire data before running into the training process. So I try the following code to load the files streamingly ```py dataset = load_dataset('togethercomputer/RedPajama-Data-1T-Sample', streaming=True) print(next(iter(dataset['train']))) ``` Sadly, it raises the following: ``` FileNotFoundError: [Errno 2] No such file or directory: 'CURRENT_CODE_PATH/arxiv_sample.jsonl' ``` I've noticed that the dataset can be properly found in the begining ``` Using the latest cached version of the module from /root/.cache/huggingface/modules/datasets_modules/datasets/togethercomputer--RedPajama-Data-1T-Sample/6ea3bc8ec2e84ec6d2df1930942e9028ace8c5b9d9143823cf911c50bbd92039 (last modified on Sat Oct 21 20:12:57 2023) since it couldn't be found locally at togethercomputer/RedPajama-Data-1T-Sample., or remotely on the Hugging Face Hub. ``` But it seems that the paths couldn't be properly parsed when loading iteratively. How should I fix this error. I've tried specifying `data_files` or `data_dir` as `.../arxiv_sample.jsonl` but none of them works. Thanks. ### Expected behavior Properly load the dataset. ### Environment info `datasets==2.14.5`
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6327/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6327/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6326
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6326/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6326/comments
https://api.github.com/repos/huggingface/datasets/issues/6326/events
https://github.com/huggingface/datasets/pull/6326
1,955,420,536
PR_kwDODunzps5dcSRa
6,326
Create battery_analysis.py
{ "avatar_url": "https://avatars.githubusercontent.com/u/130216732?v=4", "events_url": "https://api.github.com/users/vinitkm/events{/privacy}", "followers_url": "https://api.github.com/users/vinitkm/followers", "following_url": "https://api.github.com/users/vinitkm/following{/other_user}", "gists_url": "https://api.github.com/users/vinitkm/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vinitkm", "id": 130216732, "login": "vinitkm", "node_id": "U_kgDOB8LzHA", "organizations_url": "https://api.github.com/users/vinitkm/orgs", "received_events_url": "https://api.github.com/users/vinitkm/received_events", "repos_url": "https://api.github.com/users/vinitkm/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vinitkm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vinitkm/subscriptions", "type": "User", "url": "https://api.github.com/users/vinitkm" }
[]
closed
false
null
[]
null
0
"2023-10-21T10:07:48Z"
"2023-10-23T14:56:20Z"
"2023-10-23T14:56:20Z"
NONE
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6326.diff", "html_url": "https://github.com/huggingface/datasets/pull/6326", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6326.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6326" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6326/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6326/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6325
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6325/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6325/comments
https://api.github.com/repos/huggingface/datasets/issues/6325/events
https://github.com/huggingface/datasets/pull/6325
1,955,420,178
PR_kwDODunzps5dcSM3
6,325
Create battery_analysis.py
{ "avatar_url": "https://avatars.githubusercontent.com/u/130216732?v=4", "events_url": "https://api.github.com/users/vinitkm/events{/privacy}", "followers_url": "https://api.github.com/users/vinitkm/followers", "following_url": "https://api.github.com/users/vinitkm/following{/other_user}", "gists_url": "https://api.github.com/users/vinitkm/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vinitkm", "id": 130216732, "login": "vinitkm", "node_id": "U_kgDOB8LzHA", "organizations_url": "https://api.github.com/users/vinitkm/orgs", "received_events_url": "https://api.github.com/users/vinitkm/received_events", "repos_url": "https://api.github.com/users/vinitkm/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vinitkm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vinitkm/subscriptions", "type": "User", "url": "https://api.github.com/users/vinitkm" }
[]
closed
false
null
[]
null
0
"2023-10-21T10:06:37Z"
"2023-10-23T14:55:58Z"
"2023-10-23T14:55:58Z"
NONE
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6325.diff", "html_url": "https://github.com/huggingface/datasets/pull/6325", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6325.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6325" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6325/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6325/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6324
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6324/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6324/comments
https://api.github.com/repos/huggingface/datasets/issues/6324/events
https://github.com/huggingface/datasets/issues/6324
1,955,126,687
I_kwDODunzps50iN2f
6,324
Conversion to Arrow fails due to wrong type heuristic
{ "avatar_url": "https://avatars.githubusercontent.com/u/2862336?v=4", "events_url": "https://api.github.com/users/jphme/events{/privacy}", "followers_url": "https://api.github.com/users/jphme/followers", "following_url": "https://api.github.com/users/jphme/following{/other_user}", "gists_url": "https://api.github.com/users/jphme/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jphme", "id": 2862336, "login": "jphme", "node_id": "MDQ6VXNlcjI4NjIzMzY=", "organizations_url": "https://api.github.com/users/jphme/orgs", "received_events_url": "https://api.github.com/users/jphme/received_events", "repos_url": "https://api.github.com/users/jphme/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jphme/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jphme/subscriptions", "type": "User", "url": "https://api.github.com/users/jphme" }
[]
closed
false
null
[]
null
2
"2023-10-20T23:20:58Z"
"2023-10-23T20:52:57Z"
"2023-10-23T20:52:57Z"
NONE
null
null
null
### Describe the bug I have a list of dictionaries with valid/JSON-serializable values. One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on. If trying to convert this list to a dataset with `Dataset.from_list()`, I always get `ArrowInvalid: Could not convert '1' with type str: tried to convert to int64`, presumably because pyarrow tries to convert the keys to integers. Is there any way to circumvent this and fix dtypes? I didn't find anything in the documentation. ### Steps to reproduce the bug * create a list of dicts with one key being a string of an integer for the first few thousand occurences and try to convert to dataset. ### Expected behavior There shouldn't be an error (e.g. some flag to turn off automatic str to numeric conversion). ### Environment info - `datasets` version: 2.14.5 - Platform: Linux-5.15.0-84-generic-x86_64-with-glibc2.35 - Python version: 3.9.18 - Huggingface_hub version: 0.17.3 - PyArrow version: 13.0.0 - Pandas version: 2.1.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6324/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6324/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6323
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6323/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6323/comments
https://api.github.com/repos/huggingface/datasets/issues/6323/events
https://github.com/huggingface/datasets/issues/6323
1,954,245,980
I_kwDODunzps50e21c
6,323
Loading dataset from large GCS bucket very slow since 2.14
{ "avatar_url": "https://avatars.githubusercontent.com/u/6209990?v=4", "events_url": "https://api.github.com/users/jbcdnr/events{/privacy}", "followers_url": "https://api.github.com/users/jbcdnr/followers", "following_url": "https://api.github.com/users/jbcdnr/following{/other_user}", "gists_url": "https://api.github.com/users/jbcdnr/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jbcdnr", "id": 6209990, "login": "jbcdnr", "node_id": "MDQ6VXNlcjYyMDk5OTA=", "organizations_url": "https://api.github.com/users/jbcdnr/orgs", "received_events_url": "https://api.github.com/users/jbcdnr/received_events", "repos_url": "https://api.github.com/users/jbcdnr/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jbcdnr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jbcdnr/subscriptions", "type": "User", "url": "https://api.github.com/users/jbcdnr" }
[]
open
false
null
[]
null
0
"2023-10-20T12:59:55Z"
"2023-10-20T12:59:55Z"
null
NONE
null
null
null
### Describe the bug Since updating to >2.14 we have very slow access to our parquet files on GCS when loading a dataset (>30 min vs 3s). Our GCS bucket has many objects and resolving globs is very slow. I could track down the problem to this change: https://github.com/huggingface/datasets/blame/bade7af74437347a760830466eb74f7a8ce0d799/src/datasets/data_files.py#L348 The underlying implementation with gcsfs is really slow. Could you go back to the old way if we are simply giving the parquet files and no glob pattern? Thank you. ### Steps to reproduce the bug Load a dataset from a GCS bucket that has many files. ### Expected behavior Used to be fast (3s) in 2.13 ### Environment info datasets==2.14.5
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6323/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6323/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6322
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6322/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6322/comments
https://api.github.com/repos/huggingface/datasets/issues/6322/events
https://github.com/huggingface/datasets/pull/6322
1,952,947,461
PR_kwDODunzps5dT5vG
6,322
Fix regex `get_data_files` formatting for base paths
{ "avatar_url": "https://avatars.githubusercontent.com/u/1981179?v=4", "events_url": "https://api.github.com/users/ZachNagengast/events{/privacy}", "followers_url": "https://api.github.com/users/ZachNagengast/followers", "following_url": "https://api.github.com/users/ZachNagengast/following{/other_user}", "gists_url": "https://api.github.com/users/ZachNagengast/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZachNagengast", "id": 1981179, "login": "ZachNagengast", "node_id": "MDQ6VXNlcjE5ODExNzk=", "organizations_url": "https://api.github.com/users/ZachNagengast/orgs", "received_events_url": "https://api.github.com/users/ZachNagengast/received_events", "repos_url": "https://api.github.com/users/ZachNagengast/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZachNagengast/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZachNagengast/subscriptions", "type": "User", "url": "https://api.github.com/users/ZachNagengast" }
[]
closed
false
null
[]
null
4
"2023-10-19T19:45:10Z"
"2023-10-23T14:40:45Z"
"2023-10-23T14:31:21Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6322.diff", "html_url": "https://github.com/huggingface/datasets/pull/6322", "merged_at": "2023-10-23T14:31:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/6322.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6322" }
With this pr https://github.com/huggingface/datasets/pull/6309, it is formatting the entire base path into regex, which results in the undesired formatting error `doesn't match the pattern` because of the line in `glob_pattern_to_regex`: `.replace("//", "/")`: - Input: `hf://datasets/...` - Output: `hf:/datasets/...` This fix will only convert the `split_pattern` to regex and keep the `base_path` unchanged. cc @albertvillanova hopefully this still works with your implementation
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6322/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6322/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6321
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6321/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6321/comments
https://api.github.com/repos/huggingface/datasets/issues/6321/events
https://github.com/huggingface/datasets/pull/6321
1,952,643,483
PR_kwDODunzps5dS3Mc
6,321
Fix typos
{ "avatar_url": "https://avatars.githubusercontent.com/u/3097956?v=4", "events_url": "https://api.github.com/users/python273/events{/privacy}", "followers_url": "https://api.github.com/users/python273/followers", "following_url": "https://api.github.com/users/python273/following{/other_user}", "gists_url": "https://api.github.com/users/python273/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/python273", "id": 3097956, "login": "python273", "node_id": "MDQ6VXNlcjMwOTc5NTY=", "organizations_url": "https://api.github.com/users/python273/orgs", "received_events_url": "https://api.github.com/users/python273/received_events", "repos_url": "https://api.github.com/users/python273/repos", "site_admin": false, "starred_url": "https://api.github.com/users/python273/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/python273/subscriptions", "type": "User", "url": "https://api.github.com/users/python273" }
[]
closed
false
null
[]
null
2
"2023-10-19T16:24:35Z"
"2023-10-19T17:18:00Z"
"2023-10-19T17:07:35Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6321.diff", "html_url": "https://github.com/huggingface/datasets/pull/6321", "merged_at": "2023-10-19T17:07:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/6321.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6321" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6321/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6321/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6320
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6320/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6320/comments
https://api.github.com/repos/huggingface/datasets/issues/6320/events
https://github.com/huggingface/datasets/issues/6320
1,952,618,316
I_kwDODunzps50YpdM
6,320
Dataset slice splits can't load training and validation at the same time
{ "avatar_url": "https://avatars.githubusercontent.com/u/32488097?v=4", "events_url": "https://api.github.com/users/timlac/events{/privacy}", "followers_url": "https://api.github.com/users/timlac/followers", "following_url": "https://api.github.com/users/timlac/following{/other_user}", "gists_url": "https://api.github.com/users/timlac/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/timlac", "id": 32488097, "login": "timlac", "node_id": "MDQ6VXNlcjMyNDg4MDk3", "organizations_url": "https://api.github.com/users/timlac/orgs", "received_events_url": "https://api.github.com/users/timlac/received_events", "repos_url": "https://api.github.com/users/timlac/repos", "site_admin": false, "starred_url": "https://api.github.com/users/timlac/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/timlac/subscriptions", "type": "User", "url": "https://api.github.com/users/timlac" }
[]
closed
false
null
[]
null
1
"2023-10-19T16:09:22Z"
"2023-11-30T16:21:15Z"
"2023-11-30T16:21:15Z"
NONE
null
null
null
### Describe the bug According to the [documentation](https://huggingface.co/docs/datasets/v2.14.5/loading#slice-splits) is should be possible to run the following command: `train_test_ds = datasets.load_dataset("bookcorpus", split="train+test")` to load the train and test sets from the dataset. However executing the equivalent code: `speech_commands_v1 = load_dataset("superb", "ks", split="train+test")` only yields the following output: > Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 54175 > }) Where loading the dataset without the split argument yields: > DatasetDict({ > train: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 51094 > }) > validation: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 6798 > }) > test: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 3081 > }) > }) Thus, the API seems to be broken in this regard. This is a bit annoying since I want to be able to use the split argument with `split="train[:10%]+test[:10%]"` to have smaller dataset to work with when validating my model is working correctly. ### Steps to reproduce the bug `speech_commands_v1 = load_dataset("superb", "ks", split="train+test")` ### Expected behavior > DatasetDict({ > train: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 51094 > }) > test: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 3081 > }) > }) ### Environment info ``` import datasets print(datasets.__version__) ``` > 2.14.5 ``` import sys print(sys.version) ``` > 3.9.17 (main, Jul 5 2023, 20:41:20) > [GCC 11.2.0]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6320/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6320/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6319
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6319/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6319/comments
https://api.github.com/repos/huggingface/datasets/issues/6319/events
https://github.com/huggingface/datasets/issues/6319
1,952,101,717
I_kwDODunzps50WrVV
6,319
Datasets.map is severely broken
{ "avatar_url": "https://avatars.githubusercontent.com/u/4603365?v=4", "events_url": "https://api.github.com/users/phalexo/events{/privacy}", "followers_url": "https://api.github.com/users/phalexo/followers", "following_url": "https://api.github.com/users/phalexo/following{/other_user}", "gists_url": "https://api.github.com/users/phalexo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/phalexo", "id": 4603365, "login": "phalexo", "node_id": "MDQ6VXNlcjQ2MDMzNjU=", "organizations_url": "https://api.github.com/users/phalexo/orgs", "received_events_url": "https://api.github.com/users/phalexo/received_events", "repos_url": "https://api.github.com/users/phalexo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/phalexo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/phalexo/subscriptions", "type": "User", "url": "https://api.github.com/users/phalexo" }
[]
open
false
null
[]
null
11
"2023-10-19T12:19:33Z"
"2023-11-30T03:27:26Z"
null
NONE
null
null
null
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing something, but it never ends. I saw some comments about fast tokenizers using Rust and all and tried different variations. NOTHING works. ### Steps to reproduce the bug Running it without breaking the dataset into parts results in the same behavior. The loop was an attempt to see if this was a RAM issue. for idx in range(100): dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=cache_dir, split=f'train[{idx}%:{idx+1}%]') dataset = dataset.map(partial(tokenize_fn, tokenizer), batched=False, num_proc=1, remove_columns=["text", "meta"]) dataset.save_to_disk(training_args.cache_dir + f"/training_data_{idx}") ### Expected behavior I expect map to run at more or less the same speed it starts with and FINISH its processing. ### Environment info Python 3.8, same with 3.10 makes no difference. Ubuntu 20.04,
{ "+1": 3, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/6319/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6319/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6318
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6318/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6318/comments
https://api.github.com/repos/huggingface/datasets/issues/6318/events
https://github.com/huggingface/datasets/pull/6318
1,952,100,706
PR_kwDODunzps5dRC9V
6,318
Deterministic set hash
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
3
"2023-10-19T12:19:13Z"
"2023-10-19T16:27:20Z"
"2023-10-19T16:16:31Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6318.diff", "html_url": "https://github.com/huggingface/datasets/pull/6318", "merged_at": "2023-10-19T16:16:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/6318.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6318" }
Sort the items in a set according to their `datasets.fingerprint.Hasher.hash` hash to get a deterministic hash of sets. This is useful to get deterministic hashes of tokenizers that use a trie based on python sets. reported in https://github.com/huggingface/datasets/issues/3847
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6318/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6318/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6317
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6317/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6317/comments
https://api.github.com/repos/huggingface/datasets/issues/6317/events
https://github.com/huggingface/datasets/issues/6317
1,951,965,668
I_kwDODunzps50WKHk
6,317
sentiment140 dataset unavailable
{ "avatar_url": "https://avatars.githubusercontent.com/u/52670382?v=4", "events_url": "https://api.github.com/users/AndreasKarasenko/events{/privacy}", "followers_url": "https://api.github.com/users/AndreasKarasenko/followers", "following_url": "https://api.github.com/users/AndreasKarasenko/following{/other_user}", "gists_url": "https://api.github.com/users/AndreasKarasenko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AndreasKarasenko", "id": 52670382, "login": "AndreasKarasenko", "node_id": "MDQ6VXNlcjUyNjcwMzgy", "organizations_url": "https://api.github.com/users/AndreasKarasenko/orgs", "received_events_url": "https://api.github.com/users/AndreasKarasenko/received_events", "repos_url": "https://api.github.com/users/AndreasKarasenko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AndreasKarasenko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AndreasKarasenko/subscriptions", "type": "User", "url": "https://api.github.com/users/AndreasKarasenko" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2
"2023-10-19T11:25:21Z"
"2023-10-19T13:04:56Z"
"2023-10-19T13:04:56Z"
NONE
null
null
null
### Describe the bug loading the dataset using load_dataset("sentiment140") returns the following error ConnectionError: Couldn't reach http://cs.stanford.edu/people/alecmgo/trainingandtestdata.zip (error 403) ### Steps to reproduce the bug Run the following code (version should not matter). ``` from datasets import load_dataset data = load_dataset("sentiment140") ``` ### Expected behavior The dataset should be loaded just like any other. The main issue is that it is no longer hosted by stanford. It is still available from a [Google Drive Link](https://docs.google.com/file/d/0B04GJPshIjmPRnZManQwWEdTZjg/edit). ### Environment info - `datasets` version: 2.14.5 - Platform: Windows-10-10.0.19045-SP0 - Python version: 3.10.8 - Huggingface_hub version: 0.17.3 - PyArrow version: 13.0.0 - Pandas version: 2.1.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6317/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6317/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6316
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6316/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6316/comments
https://api.github.com/repos/huggingface/datasets/issues/6316/events
https://github.com/huggingface/datasets/pull/6316
1,951,819,869
PR_kwDODunzps5dQGpg
6,316
Fix loading Hub datasets with CSV metadata file
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
4
"2023-10-19T10:21:34Z"
"2023-10-20T06:23:21Z"
"2023-10-20T06:14:09Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6316.diff", "html_url": "https://github.com/huggingface/datasets/pull/6316", "merged_at": "2023-10-20T06:14:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/6316.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6316" }
Currently, the reading of the metadata file infers the file extension (.jsonl or .csv) from the passed filename. However, downloaded files from the Hub don't have file extension. For example: - the original file: `hf://datasets/__DUMMY_TRANSFORMERS_USER__/test-dataset-5916a4-16977085077831/metadata.jsonl` - corresponds to the downloaded path: `/tmp/pytest-of-username/pytest-46/cache/datasets/downloads/9f5374dbb470f711f6b89d66a5eec1f19cc96324b26bcbebe29138bda6cb20e6`, which does not have extension In the case where the metadata file does not have an extension, the reader assumes it is a JSONL file, thus the reported error when trying to read a CSV file as a JSONL one: `ArrowInvalid: JSON parse error: Invalid value. in row 0` This behavior was introduced by: - #4837 This PR extracts the metadata file extension from the original filename (instead of the downloaded one) and passes it as a parameter to the read_metadata function. Fix #6315.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6316/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6316/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6315
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6315/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6315/comments
https://api.github.com/repos/huggingface/datasets/issues/6315/events
https://github.com/huggingface/datasets/issues/6315
1,951,800,819
I_kwDODunzps50Vh3z
6,315
Hub datasets with CSV metadata raise ArrowInvalid: JSON parse error: Invalid value. in row 0
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
0
"2023-10-19T10:11:29Z"
"2023-10-20T06:14:10Z"
"2023-10-20T06:14:10Z"
MEMBER
null
null
null
When trying to load a Hub dataset that contains a CSV metadata file, it raises an `ArrowInvalid` error: ``` E pyarrow.lib.ArrowInvalid: JSON parse error: Invalid value. in row 0 pyarrow/error.pxi:100: ArrowInvalid ``` See: https://huggingface.co/datasets/lukarape/public_small_papers/discussions/1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6315/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6315/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6314
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6314/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6314/comments
https://api.github.com/repos/huggingface/datasets/issues/6314/events
https://github.com/huggingface/datasets/pull/6314
1,951,684,763
PR_kwDODunzps5dPo25
6,314
Support creating new branch in push_to_hub
{ "avatar_url": "https://avatars.githubusercontent.com/u/1000442?v=4", "events_url": "https://api.github.com/users/jmif/events{/privacy}", "followers_url": "https://api.github.com/users/jmif/followers", "following_url": "https://api.github.com/users/jmif/following{/other_user}", "gists_url": "https://api.github.com/users/jmif/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jmif", "id": 1000442, "login": "jmif", "node_id": "MDQ6VXNlcjEwMDA0NDI=", "organizations_url": "https://api.github.com/users/jmif/orgs", "received_events_url": "https://api.github.com/users/jmif/received_events", "repos_url": "https://api.github.com/users/jmif/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jmif/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmif/subscriptions", "type": "User", "url": "https://api.github.com/users/jmif" }
[]
closed
false
null
[]
null
0
"2023-10-19T09:12:39Z"
"2023-10-19T09:20:06Z"
"2023-10-19T09:19:48Z"
NONE
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6314.diff", "html_url": "https://github.com/huggingface/datasets/pull/6314", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6314.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6314" }
This adds support for creating a new branch when pushing a dataset to the hub. Tested both methods locally and branches are created.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6314/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6314/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6313
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6313/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6313/comments
https://api.github.com/repos/huggingface/datasets/issues/6313/events
https://github.com/huggingface/datasets/pull/6313
1,951,527,712
PR_kwDODunzps5dPGmL
6,313
Fix commit message formatting in multi-commit uploads
{ "avatar_url": "https://avatars.githubusercontent.com/u/45557362?v=4", "events_url": "https://api.github.com/users/qgallouedec/events{/privacy}", "followers_url": "https://api.github.com/users/qgallouedec/followers", "following_url": "https://api.github.com/users/qgallouedec/following{/other_user}", "gists_url": "https://api.github.com/users/qgallouedec/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/qgallouedec", "id": 45557362, "login": "qgallouedec", "node_id": "MDQ6VXNlcjQ1NTU3MzYy", "organizations_url": "https://api.github.com/users/qgallouedec/orgs", "received_events_url": "https://api.github.com/users/qgallouedec/received_events", "repos_url": "https://api.github.com/users/qgallouedec/repos", "site_admin": false, "starred_url": "https://api.github.com/users/qgallouedec/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qgallouedec/subscriptions", "type": "User", "url": "https://api.github.com/users/qgallouedec" }
[]
closed
false
null
[]
null
2
"2023-10-19T07:53:56Z"
"2023-10-20T14:06:13Z"
"2023-10-20T13:57:39Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6313.diff", "html_url": "https://github.com/huggingface/datasets/pull/6313", "merged_at": "2023-10-20T13:57:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/6313.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6313" }
Currently, the commit message keeps on adding: - `Upload dataset (part 00000-of-00002)` - `Upload dataset (part 00000-of-00002) (part 00001-of-00002)` Introduced in https://github.com/huggingface/datasets/pull/6269 This PR fixes this issue to have - `Upload dataset (part 00000-of-00002)` - `Upload dataset (part 00001-of-00002)`
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6313/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6313/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6312
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6312/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6312/comments
https://api.github.com/repos/huggingface/datasets/issues/6312/events
https://github.com/huggingface/datasets/pull/6312
1,950,128,416
PR_kwDODunzps5dKWDF
6,312
docs: resolving namespace conflict, refactored variable
{ "avatar_url": "https://avatars.githubusercontent.com/u/74114936?v=4", "events_url": "https://api.github.com/users/smty2018/events{/privacy}", "followers_url": "https://api.github.com/users/smty2018/followers", "following_url": "https://api.github.com/users/smty2018/following{/other_user}", "gists_url": "https://api.github.com/users/smty2018/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/smty2018", "id": 74114936, "login": "smty2018", "node_id": "MDQ6VXNlcjc0MTE0OTM2", "organizations_url": "https://api.github.com/users/smty2018/orgs", "received_events_url": "https://api.github.com/users/smty2018/received_events", "repos_url": "https://api.github.com/users/smty2018/repos", "site_admin": false, "starred_url": "https://api.github.com/users/smty2018/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/smty2018/subscriptions", "type": "User", "url": "https://api.github.com/users/smty2018" }
[]
closed
false
null
[]
null
1
"2023-10-18T16:10:59Z"
"2023-10-19T16:31:59Z"
"2023-10-19T16:23:07Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6312.diff", "html_url": "https://github.com/huggingface/datasets/pull/6312", "merged_at": "2023-10-19T16:23:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/6312.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6312" }
In docs of about_arrow.md, in the below example code ![image](https://github.com/huggingface/datasets/assets/74114936/fc70e152-e15f-422e-949a-1c4c4c9aa116) The variable name 'time' was being used in a way that could potentially lead to a namespace conflict with Python's built-in 'time' module. It is not a good convention and can lead to unintended variable shadowing for any user re-using the example code. To ensure code clarity, and prevent potential naming conflicts renamed the variable 'time' to 'elapsed_time' in the example code.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6312/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6312/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6311
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6311/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6311/comments
https://api.github.com/repos/huggingface/datasets/issues/6311/events
https://github.com/huggingface/datasets/issues/6311
1,949,304,993
I_kwDODunzps50MAih
6,311
cast_column to Sequence with length=4 occur exception raise in datasets/table.py:2146
{ "avatar_url": "https://avatars.githubusercontent.com/u/16574677?v=4", "events_url": "https://api.github.com/users/neiblegy/events{/privacy}", "followers_url": "https://api.github.com/users/neiblegy/followers", "following_url": "https://api.github.com/users/neiblegy/following{/other_user}", "gists_url": "https://api.github.com/users/neiblegy/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/neiblegy", "id": 16574677, "login": "neiblegy", "node_id": "MDQ6VXNlcjE2NTc0Njc3", "organizations_url": "https://api.github.com/users/neiblegy/orgs", "received_events_url": "https://api.github.com/users/neiblegy/received_events", "repos_url": "https://api.github.com/users/neiblegy/repos", "site_admin": false, "starred_url": "https://api.github.com/users/neiblegy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neiblegy/subscriptions", "type": "User", "url": "https://api.github.com/users/neiblegy" }
[]
closed
false
null
[]
null
4
"2023-10-18T09:38:05Z"
"2024-02-06T19:24:20Z"
"2024-02-06T19:24:20Z"
NONE
null
null
null
### Describe the bug i load a dataset from local csv file which has 187383612 examples, then use `map` to generate new columns for test. here is my code : ``` import os from datasets import load_dataset from datasets.features import Sequence, Value def add_new_path(example): example["ais_bbox"] = [100,100,200,200] example["ais_image_path"] = os.path.join("images", example["image_path"]) if example["image_path"] else "" return example ais_dataset = load_dataset("/data/ryan.gao/ais_dataset_cache/raw/1749/") hf_ds = ais_dataset.map(add_new_path, batched=False, num_proc=32) ds = hf_ds.cast_column("ais_bbox", Sequence(Value("int32"), length=4)) ``` and the `cast_column` raise an exception ``` Casting the dataset: 3%|███▉ ... File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 2110, in cast_column return self.cast(features) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 2055, in cast dataset = dataset.map( File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 592, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3097, in map for rank, done, content in Dataset._map_single(**dataset_kwargs): File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3474, in _map_single batch = apply_function_on_filtered_inputs( File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3353, in apply_function_on_filtered_inputs processed_inputs = function(*fn_args, *additional_args, **fn_kwargs) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2329, in table_cast return cast_table_to_schema(table, schema) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2288, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2288, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 1831, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 1831, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2145, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") TypeError: Couldn't cast array of type list<item: int64> to Sequence(feature=Value(dtype='int32', id=None), length=4, id=None) ``` i check the source code and make debug info: in datasets/table.py:2092 ``` 2091 if feature.length > -1: 2092 if feature.length * len(array) == len(array.values): 2093 return pa.FixedSizeListArray.from_arrays(_c(array.values, feature.feature), feature.length) 2094 print(len(array)) 2095 print(len(array.values)) ``` my feature.length is 4. but feature.length * len(array) == len(array.values) is false. print(len(array)) is 262 print(len(array.values)) is 4000 then I use "for item in array" to print each item then get 262 * [100,100,200,200] and use "for item in array.values" to print each item and get 4000 int32 which are 1000 * [100,100,200,200] i'm wondering the `chunk` in each `array.chunks`, the "chunk.values" may get all the chunks's value rather than single chunk? but i check the pyarrow's doc seems chunk.values is chunk's value not all. ### Steps to reproduce the bug code provided above. ### Expected behavior feature.length * len(array) == len(array.values) should be true. and there should not has Exception. ### Environment info python3.9 x86_64 datasets: 2.14.4 pyarrow: 13.0.0 or 10.0.0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6311/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6311/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6310
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6310/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6310/comments
https://api.github.com/repos/huggingface/datasets/issues/6310/events
https://github.com/huggingface/datasets/pull/6310
1,947,457,988
PR_kwDODunzps5dBPnY
6,310
Add return_file_name in load_dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/40604584?v=4", "events_url": "https://api.github.com/users/juliendenize/events{/privacy}", "followers_url": "https://api.github.com/users/juliendenize/followers", "following_url": "https://api.github.com/users/juliendenize/following{/other_user}", "gists_url": "https://api.github.com/users/juliendenize/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/juliendenize", "id": 40604584, "login": "juliendenize", "node_id": "MDQ6VXNlcjQwNjA0NTg0", "organizations_url": "https://api.github.com/users/juliendenize/orgs", "received_events_url": "https://api.github.com/users/juliendenize/received_events", "repos_url": "https://api.github.com/users/juliendenize/repos", "site_admin": false, "starred_url": "https://api.github.com/users/juliendenize/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/juliendenize/subscriptions", "type": "User", "url": "https://api.github.com/users/juliendenize" }
[]
open
false
null
[]
null
3
"2023-10-17T13:36:57Z"
"2023-11-27T21:11:14Z"
null
NONE
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6310.diff", "html_url": "https://github.com/huggingface/datasets/pull/6310", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6310.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6310" }
Proposition to fix #5806. Added an optional parameter `return_file_name` in the dataset builder config. When set to `True`, the function will include the file name corresponding to the sample in the returned output. There is a difference between arrow-based and folder-based datasets to return the file name: - for arrow-based: a column is concatenated after the table is cast. - for folder-based: `dataset.info.features` has the entry `file_name` and the original file name is passed to the `sample_metadata` dictionary. The difference in behavior might be a concern, also I do not know whether the `file_name` should return the original file path or the downloaded one for folder-based datasets. I added some tests for the datasets that already had a test file.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6310/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6310/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6309
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6309/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6309/comments
https://api.github.com/repos/huggingface/datasets/issues/6309/events
https://github.com/huggingface/datasets/pull/6309
1,946,916,969
PR_kwDODunzps5c_YcX
6,309
Fix get_data_patterns for directories with the word data twice
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
7
"2023-10-17T09:00:39Z"
"2023-10-18T14:01:52Z"
"2023-10-18T13:50:35Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6309.diff", "html_url": "https://github.com/huggingface/datasets/pull/6309", "merged_at": "2023-10-18T13:50:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/6309.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6309" }
Before the fix, `get_data_patterns` inferred wrongly the split name for paths with the word "data" twice: - For the URL path: `hf://datasets/piuba-bigdata/articles_and_comments@f328d536425ae8fcac5d098c8408f437bffdd357/data/train-00001-of-00009.parquet` (note the org name `piuba-bigdata/` ending with `data/`) - The inferred split name was: `articles_and_comments@f328d536425ae8fcac5d098c8408f437bffdd357/data/train` instead of `train` This PR fixes this issue by passing the `base_path` (`hf://datasets/piuba-bigdata/articles_and_comments@f328d536425ae8fcac5d098c8408f437bffdd357`) to `_get_data_files_patterns` and prepending it to the regex split pattern (`data/{split}-[0-9][0-9][0-9][0-9][0-9]-of-[0-9][0-9][0-9][0-9][0-9].*\\..*`). Fix #6305. Fix https://huggingface.co/datasets/piuba-bigdata/articles_and_comments/discussions/1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6309/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6309/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6308
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6308/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6308/comments
https://api.github.com/repos/huggingface/datasets/issues/6308/events
https://github.com/huggingface/datasets/issues/6308
1,946,810,625
I_kwDODunzps50CfkB
6,308
module 'resource' has no attribute 'error'
{ "avatar_url": "https://avatars.githubusercontent.com/u/48009681?v=4", "events_url": "https://api.github.com/users/NeoWang9999/events{/privacy}", "followers_url": "https://api.github.com/users/NeoWang9999/followers", "following_url": "https://api.github.com/users/NeoWang9999/following{/other_user}", "gists_url": "https://api.github.com/users/NeoWang9999/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NeoWang9999", "id": 48009681, "login": "NeoWang9999", "node_id": "MDQ6VXNlcjQ4MDA5Njgx", "organizations_url": "https://api.github.com/users/NeoWang9999/orgs", "received_events_url": "https://api.github.com/users/NeoWang9999/received_events", "repos_url": "https://api.github.com/users/NeoWang9999/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NeoWang9999/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NeoWang9999/subscriptions", "type": "User", "url": "https://api.github.com/users/NeoWang9999" }
[]
closed
false
null
[]
null
4
"2023-10-17T08:08:54Z"
"2023-10-25T17:09:22Z"
"2023-10-25T17:09:22Z"
NONE
null
null
null
### Describe the bug just run import: `from datasets import load_dataset` and then: ``` File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\__init__.py", line 22, in <module> from .arrow_dataset import Dataset File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\arrow_dataset.py", line 66, in <module> from .arrow_reader import ArrowReader File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\arrow_reader.py", line 30, in <module> from .download.download_config import DownloadConfig File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\download\__init__.py", line 10, in <module> from .streaming_download_manager import StreamingDownloadManager File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\download\streaming_download_manager.py", line 21, in <module> from ..filesystems import COMPRESSION_FILESYSTEMS File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\filesystems\__init__.py", line 8, in <module> import fsspec.asyn File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\fsspec\asyn.py", line 157, in <module> ResourceEror = resource.error AttributeError: module 'resource' has no attribute 'error' Process finished with exit code 1 ``` and the error codes are: ``` try: import resource except ImportError: resource = None ResourceError = OSError else: ResourceEror = resource.error ``` 1. miss spelling : "ResourceEror " should be "ResourceErorr" 2. module 'resource' has no attribute 'error' ### Steps to reproduce the bug only one step: `from datasets import load_dataset` ### Expected behavior slove error: module 'resource' has no attribute 'error' ### Environment info python=3.10 datasets==2.14.5
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6308/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6308/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6307
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6307/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6307/comments
https://api.github.com/repos/huggingface/datasets/issues/6307/events
https://github.com/huggingface/datasets/pull/6307
1,946,414,808
PR_kwDODunzps5c9s0j
6,307
Fix typo in code example in docs
{ "avatar_url": "https://avatars.githubusercontent.com/u/3905501?v=4", "events_url": "https://api.github.com/users/bryant1410/events{/privacy}", "followers_url": "https://api.github.com/users/bryant1410/followers", "following_url": "https://api.github.com/users/bryant1410/following{/other_user}", "gists_url": "https://api.github.com/users/bryant1410/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bryant1410", "id": 3905501, "login": "bryant1410", "node_id": "MDQ6VXNlcjM5MDU1MDE=", "organizations_url": "https://api.github.com/users/bryant1410/orgs", "received_events_url": "https://api.github.com/users/bryant1410/received_events", "repos_url": "https://api.github.com/users/bryant1410/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bryant1410/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bryant1410/subscriptions", "type": "User", "url": "https://api.github.com/users/bryant1410" }
[]
closed
false
null
[]
null
2
"2023-10-17T02:28:50Z"
"2023-10-17T12:59:26Z"
"2023-10-17T06:36:19Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6307.diff", "html_url": "https://github.com/huggingface/datasets/pull/6307", "merged_at": "2023-10-17T06:36:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/6307.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6307" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6307/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6307/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6306
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6306/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6306/comments
https://api.github.com/repos/huggingface/datasets/issues/6306/events
https://github.com/huggingface/datasets/issues/6306
1,946,363,452
I_kwDODunzps50AyY8
6,306
pyinstaller : OSError: could not get source code
{ "avatar_url": "https://avatars.githubusercontent.com/u/57702070?v=4", "events_url": "https://api.github.com/users/dusk877647949/events{/privacy}", "followers_url": "https://api.github.com/users/dusk877647949/followers", "following_url": "https://api.github.com/users/dusk877647949/following{/other_user}", "gists_url": "https://api.github.com/users/dusk877647949/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dusk877647949", "id": 57702070, "login": "dusk877647949", "node_id": "MDQ6VXNlcjU3NzAyMDcw", "organizations_url": "https://api.github.com/users/dusk877647949/orgs", "received_events_url": "https://api.github.com/users/dusk877647949/received_events", "repos_url": "https://api.github.com/users/dusk877647949/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dusk877647949/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dusk877647949/subscriptions", "type": "User", "url": "https://api.github.com/users/dusk877647949" }
[]
closed
false
null
[]
null
5
"2023-10-17T01:41:51Z"
"2023-11-02T07:24:51Z"
"2023-10-18T14:03:42Z"
NONE
null
null
null
### Describe the bug I ran a package with pyinstaller and got the following error: ### Steps to reproduce the bug ``` ... File "datasets\__init__.py", line 52, in <module> File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "PyInstaller\loader\pyimod02_importers.py", line 499, in exec_module File "datasets\inspect.py", line 30, in <module> File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "PyInstaller\loader\pyimod02_importers.py", line 499, in exec_module File "datasets\load.py", line 58, in <module> File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "PyInstaller\loader\pyimod02_importers.py", line 499, in exec_module File "datasets\packaged_modules\__init__.py", line 31, in <module> File "inspect.py", line 1147, in getsource File "inspect.py", line 1129, in getsourcelines File "inspect.py", line 958, in findsource OSError: could not get source code ``` ### Expected behavior I have looked up the relevant information, but I can't find a suitable reason ### Environment info ```python python 3.10 datasets 2.14.4 pyinstaller 5.6.2 ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6306/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6306/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6305
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6305/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6305/comments
https://api.github.com/repos/huggingface/datasets/issues/6305/events
https://github.com/huggingface/datasets/issues/6305
1,946,010,912
I_kwDODunzps5z_cUg
6,305
Cannot load dataset with `2.14.5`: `FileNotFound` error
{ "avatar_url": "https://avatars.githubusercontent.com/u/167943?v=4", "events_url": "https://api.github.com/users/finiteautomata/events{/privacy}", "followers_url": "https://api.github.com/users/finiteautomata/followers", "following_url": "https://api.github.com/users/finiteautomata/following{/other_user}", "gists_url": "https://api.github.com/users/finiteautomata/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/finiteautomata", "id": 167943, "login": "finiteautomata", "node_id": "MDQ6VXNlcjE2Nzk0Mw==", "organizations_url": "https://api.github.com/users/finiteautomata/orgs", "received_events_url": "https://api.github.com/users/finiteautomata/received_events", "repos_url": "https://api.github.com/users/finiteautomata/repos", "site_admin": false, "starred_url": "https://api.github.com/users/finiteautomata/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/finiteautomata/subscriptions", "type": "User", "url": "https://api.github.com/users/finiteautomata" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2
"2023-10-16T20:11:27Z"
"2023-10-18T13:50:36Z"
"2023-10-18T13:50:36Z"
NONE
null
null
null
### Describe the bug I'm trying to load [piuba-bigdata/articles_and_comments] and I'm stumbling with this error on `2.14.5`. However, this works on `2.10.0`. ### Steps to reproduce the bug [Colab link](https://colab.research.google.com/drive/1SAftFMQnFE708ikRnJJHIXZV7R5IBOCE#scrollTo=r2R2ipCCDmsg) ```python Downloading readme: 100% 1.19k/1.19k [00:00<00:00, 30.9kB/s] --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) [<ipython-input-2-807c3583d297>](https://localhost:8080/#) in <cell line: 3>() 1 from datasets import load_dataset 2 ----> 3 load_dataset("piuba-bigdata/articles_and_comments", split="train") 2 frames [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://localhost:8080/#) in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, ignore_verifications, keep_in_memory, save_infos, revision, token, use_auth_token, task, streaming, num_proc, storage_options, **config_kwargs) 2127 2128 # Create a dataset builder -> 2129 builder_instance = load_dataset_builder( 2130 path=path, 2131 name=name, [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://localhost:8080/#) in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, token, use_auth_token, storage_options, **config_kwargs) 1813 download_config = download_config.copy() if download_config else DownloadConfig() 1814 download_config.storage_options.update(storage_options) -> 1815 dataset_module = dataset_module_factory( 1816 path, 1817 revision=revision, [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://localhost:8080/#) in dataset_module_factory(path, revision, download_config, download_mode, dynamic_modules_path, data_dir, data_files, **download_kwargs) 1506 raise e1 from None 1507 if isinstance(e1, FileNotFoundError): -> 1508 raise FileNotFoundError( 1509 f"Couldn't find a dataset script at {relative_to_absolute_path(combined_path)} or any data file in the same directory. " 1510 f"Couldn't find '{path}' on the Hugging Face Hub either: {type(e1).__name__}: {e1}" FileNotFoundError: Couldn't find a dataset script at /content/piuba-bigdata/articles_and_comments/articles_and_comments.py or any data file in the same directory. Couldn't find 'piuba-bigdata/articles_and_comments' on the Hugging Face Hub either: FileNotFoundError: No (supported) data files or dataset script found in piuba-bigdata/articles_and_comments. ``` ### Expected behavior It should load normally. ### Environment info ``` - `datasets` version: 2.14.5 - Platform: Linux-5.15.120+-x86_64-with-glibc2.35 - Python version: 3.10.12 - Huggingface_hub version: 0.18.0 - PyArrow version: 9.0.0 - Pandas version: 1.5.3 ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6305/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6305/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6304
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6304/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6304/comments
https://api.github.com/repos/huggingface/datasets/issues/6304/events
https://github.com/huggingface/datasets/pull/6304
1,945,913,521
PR_kwDODunzps5c7-4q
6,304
Update README.md
{ "avatar_url": "https://avatars.githubusercontent.com/u/74114936?v=4", "events_url": "https://api.github.com/users/smty2018/events{/privacy}", "followers_url": "https://api.github.com/users/smty2018/followers", "following_url": "https://api.github.com/users/smty2018/following{/other_user}", "gists_url": "https://api.github.com/users/smty2018/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/smty2018", "id": 74114936, "login": "smty2018", "node_id": "MDQ6VXNlcjc0MTE0OTM2", "organizations_url": "https://api.github.com/users/smty2018/orgs", "received_events_url": "https://api.github.com/users/smty2018/received_events", "repos_url": "https://api.github.com/users/smty2018/repos", "site_admin": false, "starred_url": "https://api.github.com/users/smty2018/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/smty2018/subscriptions", "type": "User", "url": "https://api.github.com/users/smty2018" }
[]
closed
false
null
[]
null
1
"2023-10-16T19:10:39Z"
"2023-10-17T15:13:37Z"
"2023-10-17T15:04:52Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6304.diff", "html_url": "https://github.com/huggingface/datasets/pull/6304", "merged_at": "2023-10-17T15:04:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/6304.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6304" }
Fixed typos in ReadMe and added punctuation marks Tensorflow --> TensorFlow
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6304/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6304/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6303
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6303/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6303/comments
https://api.github.com/repos/huggingface/datasets/issues/6303/events
https://github.com/huggingface/datasets/issues/6303
1,943,466,532
I_kwDODunzps5z1vIk
6,303
Parquet uploads off-by-one naming scheme
{ "avatar_url": "https://avatars.githubusercontent.com/u/1981179?v=4", "events_url": "https://api.github.com/users/ZachNagengast/events{/privacy}", "followers_url": "https://api.github.com/users/ZachNagengast/followers", "following_url": "https://api.github.com/users/ZachNagengast/following{/other_user}", "gists_url": "https://api.github.com/users/ZachNagengast/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZachNagengast", "id": 1981179, "login": "ZachNagengast", "node_id": "MDQ6VXNlcjE5ODExNzk=", "organizations_url": "https://api.github.com/users/ZachNagengast/orgs", "received_events_url": "https://api.github.com/users/ZachNagengast/received_events", "repos_url": "https://api.github.com/users/ZachNagengast/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZachNagengast/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZachNagengast/subscriptions", "type": "User", "url": "https://api.github.com/users/ZachNagengast" }
[]
open
false
null
[]
null
4
"2023-10-14T18:31:03Z"
"2023-10-16T16:33:21Z"
null
CONTRIBUTOR
null
null
null
### Describe the bug I noticed this numbering scheme not matching up in a different project and wanted to raise it as an issue for discussion, what is the actual proper way to have these stored? <img width="425" alt="image" src="https://github.com/huggingface/datasets/assets/1981179/3ffa2144-7c9a-446f-b521-a5e9db71e7ce"> The `-SSSSS-of-NNNNN` seems to be used widely across the codebase. The section that creates the part in my screenshot is here https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L5287 There are also some edits to this section in the single commit branch. ### Steps to reproduce the bug 1. Upload a dataset that requires at least two parquet files in it 2. Observe the naming scheme ### Expected behavior The couple options here are of course **1. keeping it as is** **2. Starting the index at 1:** train-00001-of-00002-{hash}.parquet train-00002-of-00002-{hash}.parquet **3. My preferred option** (which would solve my specific issue), dropping the total entirely: train-00000-{hash}.parquet train-00001-{hash}.parquet This also solves an issue that will occur with an `append` variable for `push_to_hub` (see https://github.com/huggingface/datasets/issues/6290) where as you add a new parquet file, you need to rename everything in the repo as well. However, I know there are parts of the repo that use 0 as the starting file or may require the total, so raising the question for discussion. ### Environment info - `datasets` version: 2.14.6.dev0 - Platform: macOS-14.0-arm64-arm-64bit - Python version: 3.10.12 - Huggingface_hub version: 0.18.0 - PyArrow version: 12.0.1 - Pandas version: 1.5.3
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6303/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6303/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6302
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6302/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6302/comments
https://api.github.com/repos/huggingface/datasets/issues/6302/events
https://github.com/huggingface/datasets/issues/6302
1,942,096,078
I_kwDODunzps5zwgjO
6,302
ArrowWriter/ParquetWriter `write` method does not increase `_num_bytes` and hence datasets not sharding at `max_shard_size`
{ "avatar_url": "https://avatars.githubusercontent.com/u/2855550?v=4", "events_url": "https://api.github.com/users/Rassibassi/events{/privacy}", "followers_url": "https://api.github.com/users/Rassibassi/followers", "following_url": "https://api.github.com/users/Rassibassi/following{/other_user}", "gists_url": "https://api.github.com/users/Rassibassi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rassibassi", "id": 2855550, "login": "Rassibassi", "node_id": "MDQ6VXNlcjI4NTU1NTA=", "organizations_url": "https://api.github.com/users/Rassibassi/orgs", "received_events_url": "https://api.github.com/users/Rassibassi/received_events", "repos_url": "https://api.github.com/users/Rassibassi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rassibassi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rassibassi/subscriptions", "type": "User", "url": "https://api.github.com/users/Rassibassi" }
[]
closed
false
null
[]
null
2
"2023-10-13T14:43:36Z"
"2023-10-17T06:52:12Z"
"2023-10-17T06:52:11Z"
NONE
null
null
null
### Describe the bug An example from [1], does not work when limiting shards with `max_shard_size`. Try the following example with low `max_shard_size`, such as: ```python builder.download_and_prepare(output_dir, storage_options=storage_options, file_format="parquet", max_shard_size="10MB") ``` The reason for this is that, in line [2] `writer._num_bytes > max_shard_size` is never true, because the `write` method of `ArrowWriter` [3] does not increase `self._num_bytes`. Such that respective Arrow/Parquet shards are only written to file based on the `writer_batch_size` or `config.DEFAULT_MAX_BATCH_SIZE`, but not based on `max_shard_size`. [1] https://huggingface.co/docs/datasets/filesystems#download-and-prepare-a-dataset-into-a-cloud-storage [2] https://github.com/huggingface/datasets/blob/3e8d420808718c9a1453a2e7ee3484ca12c9c70d/src/datasets/builder.py#L1677 [3] https://github.com/huggingface/datasets/blob/3e8d420808718c9a1453a2e7ee3484ca12c9c70d/src/datasets/arrow_writer.py#L459 ### Steps to reproduce the bug Get example from: https://huggingface.co/docs/datasets/filesystems#download-and-prepare-a-dataset-into-a-cloud-storage Call `builder.download_and_prepare` with low `max_shard_size` such as `10MB`, e.g.: ```python builder.download_and_prepare(output_dir, storage_options=storage_options, file_format="parquet", max_shard_size="10MB") ``` ### Expected behavior Shards should be written based on `max_shard_size` instead of batch size. ### Environment info ``` >>> import datasets >>> datasets.__version__ '2.14.6.dev0 ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6302/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6302/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6301
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6301/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6301/comments
https://api.github.com/repos/huggingface/datasets/issues/6301/events
https://github.com/huggingface/datasets/pull/6301
1,940,183,999
PR_kwDODunzps5cpPVh
6,301
Unpin `tensorflow` maximum version
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
3
"2023-10-12T14:58:07Z"
"2023-10-12T15:58:20Z"
"2023-10-12T15:49:54Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6301.diff", "html_url": "https://github.com/huggingface/datasets/pull/6301", "merged_at": "2023-10-12T15:49:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/6301.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6301" }
Removes the temporary pin introduced in #6264
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6301/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6301/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6300
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6300/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6300/comments
https://api.github.com/repos/huggingface/datasets/issues/6300/events
https://github.com/huggingface/datasets/pull/6300
1,940,153,432
PR_kwDODunzps5cpIoG
6,300
Unpin `jax` maximum version
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
6
"2023-10-12T14:42:40Z"
"2023-10-12T16:37:55Z"
"2023-10-12T16:28:57Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6300.diff", "html_url": "https://github.com/huggingface/datasets/pull/6300", "merged_at": "2023-10-12T16:28:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/6300.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6300" }
fix #6299 fix #6202
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6300/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6300/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6299
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6299/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6299/comments
https://api.github.com/repos/huggingface/datasets/issues/6299/events
https://github.com/huggingface/datasets/issues/6299
1,939,649,238
I_kwDODunzps5znLLW
6,299
Support for newer versions of JAX
{ "avatar_url": "https://avatars.githubusercontent.com/u/25456859?v=4", "events_url": "https://api.github.com/users/ddrous/events{/privacy}", "followers_url": "https://api.github.com/users/ddrous/followers", "following_url": "https://api.github.com/users/ddrous/following{/other_user}", "gists_url": "https://api.github.com/users/ddrous/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ddrous", "id": 25456859, "login": "ddrous", "node_id": "MDQ6VXNlcjI1NDU2ODU5", "organizations_url": "https://api.github.com/users/ddrous/orgs", "received_events_url": "https://api.github.com/users/ddrous/received_events", "repos_url": "https://api.github.com/users/ddrous/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ddrous/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ddrous/subscriptions", "type": "User", "url": "https://api.github.com/users/ddrous" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
0
"2023-10-12T10:03:46Z"
"2023-10-12T16:28:59Z"
"2023-10-12T16:28:59Z"
NONE
null
null
null
### Feature request Hi, I like your idea of adapting the datasets library to be usable with JAX. Thank you for that. However, in your [setup.py](https://github.com/huggingface/datasets/blob/main/setup.py), you enforce old versions of JAX <= 0.3... It is very cumbersome ! What is the rationale for such a limitation ? Can you remove it please ? Thanks, ### Motivation This library is unusable with new versions of JAX ? ### Your contribution Yes.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6299/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6299/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6298
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6298/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6298/comments
https://api.github.com/repos/huggingface/datasets/issues/6298/events
https://github.com/huggingface/datasets/pull/6298
1,938,797,389
PR_kwDODunzps5ckg6j
6,298
Doc readme improvements
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
2
"2023-10-11T21:51:12Z"
"2023-10-12T12:47:15Z"
"2023-10-12T12:38:19Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6298.diff", "html_url": "https://github.com/huggingface/datasets/pull/6298", "merged_at": "2023-10-12T12:38:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/6298.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6298" }
Changes in the doc READMe: * adds two new sections (to be aligned with `transformers` and `hfh`): "Previewing the documentation" and "Writing documentation examples" * replaces the mentions of `transformers` with `datasets` * fixes some dead links
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6298/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6298/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6297
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6297/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6297/comments
https://api.github.com/repos/huggingface/datasets/issues/6297/events
https://github.com/huggingface/datasets/pull/6297
1,938,752,707
PR_kwDODunzps5ckXBa
6,297
Fix ArrayXD cast
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
2
"2023-10-11T21:14:59Z"
"2023-10-13T13:54:00Z"
"2023-10-13T13:45:30Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6297.diff", "html_url": "https://github.com/huggingface/datasets/pull/6297", "merged_at": "2023-10-13T13:45:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/6297.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6297" }
Fix #6291
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6297/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6297/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6296
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6296/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6296/comments
https://api.github.com/repos/huggingface/datasets/issues/6296/events
https://github.com/huggingface/datasets/pull/6296
1,938,453,845
PR_kwDODunzps5cjUs1
6,296
Move `exceptions.py` to `utils/exceptions.py`
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
open
false
null
[]
null
4
"2023-10-11T18:28:00Z"
"2023-10-17T13:25:33Z"
null
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6296.diff", "html_url": "https://github.com/huggingface/datasets/pull/6296", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6296.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6296" }
I didn't notice the path while reviewing the PR yesterday :(
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6296/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6296/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6295
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6295/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6295/comments
https://api.github.com/repos/huggingface/datasets/issues/6295/events
https://github.com/huggingface/datasets/pull/6295
1,937,362,102
PR_kwDODunzps5cfiW8
6,295
Fix parquet columns argument in streaming mode
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
3
"2023-10-11T10:01:01Z"
"2023-10-11T16:30:24Z"
"2023-10-11T16:21:36Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6295.diff", "html_url": "https://github.com/huggingface/datasets/pull/6295", "merged_at": "2023-10-11T16:21:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/6295.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6295" }
It was failing when there's a DatasetInfo with non-None info.features from the YAML (therefore containing columns that should be ignored) Fix https://github.com/huggingface/datasets/issues/6293
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6295/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6295/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6294
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6294/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6294/comments
https://api.github.com/repos/huggingface/datasets/issues/6294/events
https://github.com/huggingface/datasets/issues/6294
1,937,359,605
I_kwDODunzps5zecL1
6,294
IndexError: Invalid key is out of bounds for size 0 despite having a populated dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/61892155?v=4", "events_url": "https://api.github.com/users/ZYM66/events{/privacy}", "followers_url": "https://api.github.com/users/ZYM66/followers", "following_url": "https://api.github.com/users/ZYM66/following{/other_user}", "gists_url": "https://api.github.com/users/ZYM66/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZYM66", "id": 61892155, "login": "ZYM66", "node_id": "MDQ6VXNlcjYxODkyMTU1", "organizations_url": "https://api.github.com/users/ZYM66/orgs", "received_events_url": "https://api.github.com/users/ZYM66/received_events", "repos_url": "https://api.github.com/users/ZYM66/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZYM66/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZYM66/subscriptions", "type": "User", "url": "https://api.github.com/users/ZYM66" }
[]
closed
false
null
[]
null
1
"2023-10-11T09:59:38Z"
"2023-10-17T11:24:06Z"
"2023-10-17T11:24:06Z"
NONE
null
null
null
### Describe the bug I am encountering an `IndexError` when trying to access data from a DataLoader which wraps around a dataset I've loaded using the `datasets` library. The error suggests that the dataset size is `0`, but when I check the length and print the dataset, it's clear that it has `1166` entries. ### Steps to reproduce the bug 1. Load a dataset with `1166` entries. 2. Create a DataLoader using this dataset. 3. Try iterating over the DataLoader. code: ```python def get_train_dataloader(self) -> DataLoader: if self.train_dataset is None: raise ValueError("Trainer: training requires a train_dataset.") train_dataset = self.train_dataset data_collator = self.data_collator print(len(train_dataset)) print(train_dataset) if is_datasets_available() and isinstance(train_dataset, datasets.Dataset): train_dataset = self._remove_unused_columns(train_dataset, description="training") else: data_collator = self._get_collator_with_removed_columns(data_collator, description="training") train_sampler = self._get_train_sampler() dl = DataLoader( train_dataset, batch_size=self._train_batch_size, sampler=train_sampler, collate_fn=data_collator, drop_last=self.args.dataloader_drop_last, num_workers=self.args.dataloader_num_workers, pin_memory=self.args.dataloader_pin_memory, worker_init_fn=seed_worker, ) print(dl) print(len(dl)) for i in dl: print(i) break return dl ``` output : ``` 1166 Dataset({ features: ['input_ids', 'special_tokens_mask'], num_rows: 1166 }) <torch.utils.data.dataloader.DataLoader object ...> 146 ``` Error: ``` Traceback (most recent call last): File "/home/dl/zym/llamaJP/TestUseContinuePretrainLlama.py", line 266, in <module> train() File "/home/dl/zym/llamaJP/TestUseContinuePretrainLlama.py", line 260, in train trainer.train() File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/transformers/trainer.py", line 1506, in train return inner_training_loop( File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/transformers/trainer.py", line 1520, in _inner_training_loop train_dataloader = self.get_train_dataloader() File "/home/dl/zym/llamaJP/TestUseContinuePretrainLlama.py", line 80, in get_train_dataloader for i in dl: File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 630, in __next__ data = self._next_data() File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 674, in _next_data data = self._dataset_fetcher.fetch(index) # may raise StopIteration File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/torch/utils/data/_utils/fetch.py", line 49, in fetch data = self.dataset.__getitems__(possibly_batched_index) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2807, in __getitems__ batch = self.__getitem__(keys) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2803, in __getitem__ return self._getitem(key) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2787, in _getitem pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/formatting/formatting.py", line 583, in query_table _check_valid_index_key(key, size) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/formatting/formatting.py", line 536, in _check_valid_index_key _check_valid_index_key(int(max(key)), size=size) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/formatting/formatting.py", line 526, in _check_valid_index_key raise IndexError(f"Invalid key: {key} is out of bounds for size {size}") IndexError: Invalid key: 1116 is out of bounds for size 0 ``` ### Expected behavior I expect to be able to iterate over the DataLoader without encountering an IndexError since the dataset is populated. ### Environment info - `datasets` library version: [2.14.5] - Platform: [Linux] - Python version: 3.10 - Other libraries involved: HuggingFace Transformers
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6294/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6294/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6293
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6293/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6293/comments
https://api.github.com/repos/huggingface/datasets/issues/6293/events
https://github.com/huggingface/datasets/issues/6293
1,937,238,047
I_kwDODunzps5zd-gf
6,293
Choose columns to stream parquet data in streaming mode
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
0
"2023-10-11T08:59:36Z"
"2023-10-11T16:21:38Z"
"2023-10-11T16:21:38Z"
MEMBER
null
null
null
Currently passing columns= to load_dataset in streaming mode fails ``` Tried to load parquet data with columns '['link']' with mismatching features '{'caption': Value(dtype='string', id=None), 'image': {'bytes': Value(dtype='binary', id=None), 'path': Value(dtype='null', id=None)}, 'link': Value(dtype='string', id=None), 'message_id': Value(dtype='string', id=None), 'timestamp': Value(dtype='string', id=None)}' ``` similar to https://github.com/huggingface/datasets/issues/6039 reported at https://huggingface.co/datasets/laion/dalle-3-dataset/discussions/3#65259a09617407d4520f4ad9
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6293/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6293/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6292
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6292/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6292/comments
https://api.github.com/repos/huggingface/datasets/issues/6292/events
https://github.com/huggingface/datasets/issues/6292
1,937,050,470
I_kwDODunzps5zdQtm
6,292
how to load the image of dtype float32 or float64
{ "avatar_url": "https://avatars.githubusercontent.com/u/26437644?v=4", "events_url": "https://api.github.com/users/wanglaofei/events{/privacy}", "followers_url": "https://api.github.com/users/wanglaofei/followers", "following_url": "https://api.github.com/users/wanglaofei/following{/other_user}", "gists_url": "https://api.github.com/users/wanglaofei/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wanglaofei", "id": 26437644, "login": "wanglaofei", "node_id": "MDQ6VXNlcjI2NDM3NjQ0", "organizations_url": "https://api.github.com/users/wanglaofei/orgs", "received_events_url": "https://api.github.com/users/wanglaofei/received_events", "repos_url": "https://api.github.com/users/wanglaofei/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wanglaofei/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wanglaofei/subscriptions", "type": "User", "url": "https://api.github.com/users/wanglaofei" }
[]
open
false
null
[]
null
1
"2023-10-11T07:27:16Z"
"2023-10-11T13:19:11Z"
null
NONE
null
null
null
_FEATURES = datasets.Features( { "image": datasets.Image(), "text": datasets.Value("string"), }, ) The datasets builder seems only support the unit8 data. How to load the float dtype data?
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6292/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6292/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6291
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6291/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6291/comments
https://api.github.com/repos/huggingface/datasets/issues/6291/events
https://github.com/huggingface/datasets/issues/6291
1,936,129,871
I_kwDODunzps5zZv9P
6,291
Casting type from Array2D int to Array2D float crashes
{ "avatar_url": "https://avatars.githubusercontent.com/u/22567306?v=4", "events_url": "https://api.github.com/users/AlanBlanchet/events{/privacy}", "followers_url": "https://api.github.com/users/AlanBlanchet/followers", "following_url": "https://api.github.com/users/AlanBlanchet/following{/other_user}", "gists_url": "https://api.github.com/users/AlanBlanchet/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AlanBlanchet", "id": 22567306, "login": "AlanBlanchet", "node_id": "MDQ6VXNlcjIyNTY3MzA2", "organizations_url": "https://api.github.com/users/AlanBlanchet/orgs", "received_events_url": "https://api.github.com/users/AlanBlanchet/received_events", "repos_url": "https://api.github.com/users/AlanBlanchet/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AlanBlanchet/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AlanBlanchet/subscriptions", "type": "User", "url": "https://api.github.com/users/AlanBlanchet" }
[]
closed
false
null
[]
null
1
"2023-10-10T20:10:10Z"
"2023-10-13T13:45:31Z"
"2023-10-13T13:45:31Z"
NONE
null
null
null
### Describe the bug I am on a school project and the initial type for feature annotations are `Array2D(shape=(None, 4))`. I am trying to cast this type to a `float64` and pyarrow gives me this error : ``` Traceback (most recent call last): File "/home/alan/dev/ClassezDesImagesAvecDesAlgorithmesDeDeeplearning/src/sdd/data/dataset.py", line 141, in <module> dataset = StanfordDogsDataset(size, 5).original(True).demo() File "<attrs generated init __main__.StanfordDogsDataset>", line 4, in __init__ File "/home/alan/dev/ClassezDesImagesAvecDesAlgorithmesDeDeeplearning/src/sdd/data/dataset.py", line 33, in __attrs_post_init__ self.dataset = self.dataset.cast_column( File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/fingerprint.py", line 511, in wrapper out = func(dataset, *args, **kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2110, in cast_column return self.cast(features) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2055, in cast dataset = dataset.map( File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 592, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 3097, in map for rank, done, content in Dataset._map_single(**dataset_kwargs): File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 3474, in _map_single batch = apply_function_on_filtered_inputs( File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 3353, in apply_function_on_filtered_inputs processed_inputs = function(*fn_args, *additional_args, **fn_kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 2328, in table_cast return cast_table_to_schema(table, schema) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 2287, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 2287, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 1831, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 1831, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 2143, in cast_array_to_feature return array_cast(array, feature(), allow_number_to_str=allow_number_to_str) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 1833, in wrapper return func(array, *args, **kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 1967, in array_cast return pa_type.wrap_array(array) File "pyarrow/types.pxi", line 1369, in pyarrow.lib.BaseExtensionType.wrap_array TypeError: Incompatible storage type for extension<arrow.py_extension_type<Array2DExtensionType>>: expected list<item: list<item: double>>, got list<item: list<item: int32>> ``` ### Steps to reproduce the bug ```python dataset = datasets.load_dataset("Alanox/stanford-dogs", split="full") dataset = dataset.cast_column("annotations", Array2D((None, 4), "float64")) ``` ### Expected behavior It should simply cast the column feature type to a `float64` without error ### Environment info datasets == 2.14.5
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6291/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6291/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6290
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6290/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6290/comments
https://api.github.com/repos/huggingface/datasets/issues/6290/events
https://github.com/huggingface/datasets/issues/6290
1,935,629,679
I_kwDODunzps5zX11v
6,290
Incremental dataset (e.g. `.push_to_hub(..., append=True)`)
{ "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Wauplin", "id": 11801849, "login": "Wauplin", "node_id": "MDQ6VXNlcjExODAxODQ5", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "repos_url": "https://api.github.com/users/Wauplin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "type": "User", "url": "https://api.github.com/users/Wauplin" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
1
"2023-10-10T15:18:03Z"
"2023-10-13T16:05:26Z"
null
CONTRIBUTOR
null
null
null
### Feature request Have the possibility to do `ds.push_to_hub(..., append=True)`. ### Motivation Requested in this [comment](https://huggingface.co/datasets/laion/dalle-3-dataset/discussions/3#65252597c4edc168202a5eaa) and this [comment](https://huggingface.co/datasets/laion/dalle-3-dataset/discussions/4#6524f675c9607bdffb208d8f). Discussed internally on [slack](https://huggingface.slack.com/archives/C02EMARJ65P/p1696950642610639?thread_ts=1690554266.830949&cid=C02EMARJ65P). ### Your contribution What I suggest to do for parquet datasets is to use `CommitOperationCopy` + `CommitOperationDelete` from `huggingface_hub`: 1. list files 2. copy files from parquet-0001-of-0004 to parquet-0001-of-0005 3. delete files like parquet-0001-of-0004 4. generate + add last parquet file parquet-0005-of-0005 => make a single commit with all commit operations at once I think it should be quite straightforward to implement. Happy to review a PR (maybe conflicting with the ongoing "1 commit push_to_hub" PR https://github.com/huggingface/datasets/pull/6269)
{ "+1": 3, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/6290/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6290/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6289
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6289/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6289/comments
https://api.github.com/repos/huggingface/datasets/issues/6289/events
https://github.com/huggingface/datasets/pull/6289
1,935,628,506
PR_kwDODunzps5cZiay
6,289
testing doc-builder
{ "avatar_url": "https://avatars.githubusercontent.com/u/11827707?v=4", "events_url": "https://api.github.com/users/mishig25/events{/privacy}", "followers_url": "https://api.github.com/users/mishig25/followers", "following_url": "https://api.github.com/users/mishig25/following{/other_user}", "gists_url": "https://api.github.com/users/mishig25/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mishig25", "id": 11827707, "login": "mishig25", "node_id": "MDQ6VXNlcjExODI3NzA3", "organizations_url": "https://api.github.com/users/mishig25/orgs", "received_events_url": "https://api.github.com/users/mishig25/received_events", "repos_url": "https://api.github.com/users/mishig25/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mishig25/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mishig25/subscriptions", "type": "User", "url": "https://api.github.com/users/mishig25" }
[]
closed
false
null
[]
null
2
"2023-10-10T15:17:29Z"
"2023-10-13T08:57:14Z"
"2023-10-13T08:56:48Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6289.diff", "html_url": "https://github.com/huggingface/datasets/pull/6289", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6289.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6289" }
testing https://github.com/huggingface/doc-builder/pull/426
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6289/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6289/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6288
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6288/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6288/comments
https://api.github.com/repos/huggingface/datasets/issues/6288/events
https://github.com/huggingface/datasets/issues/6288
1,935,005,457
I_kwDODunzps5zVdcR
6,288
Dataset.from_pandas with a DataFrame of PIL.Images
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
2
"2023-10-10T10:29:16Z"
"2023-10-20T18:23:05Z"
null
MEMBER
null
null
null
Currently type inference doesn't know what to do with a Pandas Series of PIL.Image objects, though it would be nice to get a Dataset with the Image type this way
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6288/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6288/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6287
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6287/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6287/comments
https://api.github.com/repos/huggingface/datasets/issues/6287/events
https://github.com/huggingface/datasets/issues/6287
1,932,758,192
I_kwDODunzps5zM4yw
6,287
map() not recognizing "text"
{ "avatar_url": "https://avatars.githubusercontent.com/u/5688359?v=4", "events_url": "https://api.github.com/users/EngineerKhan/events{/privacy}", "followers_url": "https://api.github.com/users/EngineerKhan/followers", "following_url": "https://api.github.com/users/EngineerKhan/following{/other_user}", "gists_url": "https://api.github.com/users/EngineerKhan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/EngineerKhan", "id": 5688359, "login": "EngineerKhan", "node_id": "MDQ6VXNlcjU2ODgzNTk=", "organizations_url": "https://api.github.com/users/EngineerKhan/orgs", "received_events_url": "https://api.github.com/users/EngineerKhan/received_events", "repos_url": "https://api.github.com/users/EngineerKhan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/EngineerKhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EngineerKhan/subscriptions", "type": "User", "url": "https://api.github.com/users/EngineerKhan" }
[]
closed
false
null
[]
null
1
"2023-10-09T10:27:30Z"
"2023-10-11T20:28:45Z"
"2023-10-11T20:28:45Z"
NONE
null
null
null
### Describe the bug The [map() documentation](https://huggingface.co/docs/datasets/v2.14.5/en/package_reference/main_classes#datasets.Dataset.map) reads: ` ds = ds.map(lambda x: tokenizer(x['text'], truncation=True, padding=True), batched=True)` I have been trying to reproduce it in my code as: `tokenizedDataset = dataset.map(lambda x: tokenizer(x['text']), batched=True)` But it doesn't work as it throws the error: > KeyError: 'text' Can you please guide me on how to fix it? ### Steps to reproduce the bug 1. `from datasets import load_dataset dataset = load_dataset("amazon_reviews_multi")` 2. Then this code: `from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("bert-base-cased")` 3. The line I quoted above (which I have been trying) ### Expected behavior As mentioned in the documentation, it should run without any error and map the tokenization on the whole dataset. ### Environment info Python 3.10.2
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6287/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6287/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6286
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6286/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6286/comments
https://api.github.com/repos/huggingface/datasets/issues/6286/events
https://github.com/huggingface/datasets/pull/6286
1,932,640,128
PR_kwDODunzps5cPKNK
6,286
Create DefunctDatasetError
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
2
"2023-10-09T09:23:23Z"
"2023-10-10T07:13:22Z"
"2023-10-10T07:03:04Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6286.diff", "html_url": "https://github.com/huggingface/datasets/pull/6286", "merged_at": "2023-10-10T07:03:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/6286.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6286" }
Create `DefunctDatasetError` as a specific error to be raised when a dataset is defunct and no longer accessible. See Hub discussion: https://huggingface.co/datasets/the_pile_books3/discussions/7#6523c13a94f3a1a2092d251b
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6286/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6286/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6285
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6285/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6285/comments
https://api.github.com/repos/huggingface/datasets/issues/6285/events
https://github.com/huggingface/datasets/issues/6285
1,932,306,325
I_kwDODunzps5zLKeV
6,285
TypeError: expected str, bytes or os.PathLike object, not dict
{ "avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4", "events_url": "https://api.github.com/users/andysingal/events{/privacy}", "followers_url": "https://api.github.com/users/andysingal/followers", "following_url": "https://api.github.com/users/andysingal/following{/other_user}", "gists_url": "https://api.github.com/users/andysingal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/andysingal", "id": 20493493, "login": "andysingal", "node_id": "MDQ6VXNlcjIwNDkzNDkz", "organizations_url": "https://api.github.com/users/andysingal/orgs", "received_events_url": "https://api.github.com/users/andysingal/received_events", "repos_url": "https://api.github.com/users/andysingal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/andysingal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andysingal/subscriptions", "type": "User", "url": "https://api.github.com/users/andysingal" }
[]
open
false
null
[]
null
4
"2023-10-09T04:56:26Z"
"2023-10-10T13:17:33Z"
null
NONE
null
null
null
### Describe the bug my dataset is in form : train- image /n -labels and tried the code: ``` from datasets import load_dataset data_files = { "train": "/content/datasets/PotholeDetectionYOLOv8-1/train/", "validation": "/content/datasets/PotholeDetectionYOLOv8-1/valid/", "test": "/content/datasets/PotholeDetectionYOLOv8-1/test/" } dataset = load_dataset("imagefolder", data_dir=data_files) dataset ``` got error: ``` --------------------------------------------------------------------------- TypeError Traceback (most recent call last) [<ipython-input-29-2ef1926f73d9>](https://localhost:8080/#) in <cell line: 8>() 6 "test": "/content/datasets/PotholeDetectionYOLOv8-1/test/" 7 } ----> 8 dataset = load_dataset("imagefolder", data_dir=data_files) 9 dataset 6 frames [/usr/lib/python3.10/pathlib.py](https://localhost:8080/#) in _parse_args(cls, args) 576 parts += a._parts 577 else: --> 578 a = os.fspath(a) 579 if isinstance(a, str): 580 # Force-cast str subclasses to str (issue #21127) TypeError: expected str, bytes or os.PathLike object, not dict ``` ### Steps to reproduce the bug as share above ### Expected behavior load images and labels , but my dataset only uploads images - https://huggingface.co/datasets/Andyrasika/potholes-dataset ### Environment info colab pro
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6285/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6285/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6284
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6284/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6284/comments
https://api.github.com/repos/huggingface/datasets/issues/6284/events
https://github.com/huggingface/datasets/issues/6284
1,929,551,712
I_kwDODunzps5zAp9g
6,284
Add Belebele multiple-choice machine reading comprehension (MRC) dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/64583161?v=4", "events_url": "https://api.github.com/users/rajveer43/events{/privacy}", "followers_url": "https://api.github.com/users/rajveer43/followers", "following_url": "https://api.github.com/users/rajveer43/following{/other_user}", "gists_url": "https://api.github.com/users/rajveer43/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rajveer43", "id": 64583161, "login": "rajveer43", "node_id": "MDQ6VXNlcjY0NTgzMTYx", "organizations_url": "https://api.github.com/users/rajveer43/orgs", "received_events_url": "https://api.github.com/users/rajveer43/received_events", "repos_url": "https://api.github.com/users/rajveer43/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rajveer43/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rajveer43/subscriptions", "type": "User", "url": "https://api.github.com/users/rajveer43" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
1
"2023-10-06T06:58:03Z"
"2023-10-06T13:26:51Z"
"2023-10-06T13:26:51Z"
NONE
null
null
null
### Feature request Belebele is a multiple-choice machine reading comprehension (MRC) dataset spanning 122 language variants. This dataset enables the evaluation of mono- and multi-lingual models in high-, medium-, and low-resource languages. Each question has four multiple-choice answers and is linked to a short passage from the [FLORES-200](https://github.com/facebookresearch/flores/tree/main/flores200) dataset. The human annotation procedure was carefully curated to create questions that discriminate between different levels of generalizable language comprehension and is reinforced by extensive quality checks. While all questions directly relate to the passage, the English dataset on its own proves difficult enough to challenge state-of-the-art language models. Being fully parallel, this dataset enables direct comparison of model performance across all languages. Belebele opens up new avenues for evaluating and analyzing the multilingual abilities of language models and NLP systems. Please refer to paper for more details, [The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants](https://arxiv.org/abs/2308.16884). ## Composition - 900 questions per language variant - 488 distinct passages, there are 1-2 associated questions for each. - For each question, there is 4 multiple-choice answers, exactly 1 of which is correct. - 122 language/language variants (including English). - 900 x 122 = 109,800 total questions. ### Motivation official repo https://github.com/facebookresearch/belebele ### Your contribution -
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6284/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6284/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6283
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6283/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6283/comments
https://api.github.com/repos/huggingface/datasets/issues/6283/events
https://github.com/huggingface/datasets/pull/6283
1,928,552,257
PR_kwDODunzps5cBlKq
6,283
Fix array cast/embed with null values
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
9
"2023-10-05T15:24:05Z"
"2024-02-06T19:30:25Z"
"2024-02-06T19:24:19Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6283.diff", "html_url": "https://github.com/huggingface/datasets/pull/6283", "merged_at": "2024-02-06T19:24:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/6283.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6283" }
Fixes issues with casting/embedding PyArrow list arrays with null values. It also bumps the required PyArrow version to 12.0.0 (over 9 months old) to simplify the implementation. Fix #6280, fix #6311, fix #6360 (Also fixes https://github.com/huggingface/datasets/issues/5430 to make Beam compatible with PyArrow>=12.0.0)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6283/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6283/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6282
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6282/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6282/comments
https://api.github.com/repos/huggingface/datasets/issues/6282/events
https://github.com/huggingface/datasets/pull/6282
1,928,473,630
PR_kwDODunzps5cBT5p
6,282
Drop data_files duplicates
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
open
false
null
[]
null
4
"2023-10-05T14:43:08Z"
"2023-11-14T17:24:50Z"
null
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6282.diff", "html_url": "https://github.com/huggingface/datasets/pull/6282", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6282.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6282" }
I just added drop_duplicates=True to `.from_patterns`. I used a dict to deduplicate and preserve the order close https://github.com/huggingface/datasets/issues/6259 close https://github.com/huggingface/datasets/issues/6272
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6282/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6282/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6281
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6281/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6281/comments
https://api.github.com/repos/huggingface/datasets/issues/6281/events
https://github.com/huggingface/datasets/pull/6281
1,928,456,959
PR_kwDODunzps5cBQPd
6,281
Improve documentation of dataset.from_generator
{ "avatar_url": "https://avatars.githubusercontent.com/u/53510?v=4", "events_url": "https://api.github.com/users/hartmans/events{/privacy}", "followers_url": "https://api.github.com/users/hartmans/followers", "following_url": "https://api.github.com/users/hartmans/following{/other_user}", "gists_url": "https://api.github.com/users/hartmans/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hartmans", "id": 53510, "login": "hartmans", "node_id": "MDQ6VXNlcjUzNTEw", "organizations_url": "https://api.github.com/users/hartmans/orgs", "received_events_url": "https://api.github.com/users/hartmans/received_events", "repos_url": "https://api.github.com/users/hartmans/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hartmans/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hartmans/subscriptions", "type": "User", "url": "https://api.github.com/users/hartmans" }
[]
closed
false
null
[]
null
2
"2023-10-05T14:34:49Z"
"2023-10-05T19:09:07Z"
"2023-10-05T18:57:41Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6281.diff", "html_url": "https://github.com/huggingface/datasets/pull/6281", "merged_at": "2023-10-05T18:57:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/6281.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6281" }
Improve documentation to clarify sharding behavior (#6270)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6281/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6281/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6280
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6280/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6280/comments
https://api.github.com/repos/huggingface/datasets/issues/6280/events
https://github.com/huggingface/datasets/issues/6280
1,928,215,278
I_kwDODunzps5y7jru
6,280
Couldn't cast array of type fixed_size_list to Sequence(Value(float64))
{ "avatar_url": "https://avatars.githubusercontent.com/u/1000442?v=4", "events_url": "https://api.github.com/users/jmif/events{/privacy}", "followers_url": "https://api.github.com/users/jmif/followers", "following_url": "https://api.github.com/users/jmif/following{/other_user}", "gists_url": "https://api.github.com/users/jmif/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jmif", "id": 1000442, "login": "jmif", "node_id": "MDQ6VXNlcjEwMDA0NDI=", "organizations_url": "https://api.github.com/users/jmif/orgs", "received_events_url": "https://api.github.com/users/jmif/received_events", "repos_url": "https://api.github.com/users/jmif/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jmif/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmif/subscriptions", "type": "User", "url": "https://api.github.com/users/jmif" }
[]
closed
false
null
[]
null
4
"2023-10-05T12:48:31Z"
"2024-02-06T19:24:20Z"
"2024-02-06T19:24:20Z"
NONE
null
null
null
### Describe the bug I have a dataset with an embedding column, when I try to map that dataset I get the following exception: ``` Traceback (most recent call last): File "/Users/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 3189, in map for rank, done, content in iflatmap_unordered( File "/Users/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/utils/py_utils.py", line 1387, in iflatmap_unordered [async_result.get(timeout=0.05) for async_result in async_results] File "/Users/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/utils/py_utils.py", line 1387, in <listcomp> [async_result.get(timeout=0.05) for async_result in async_results] File "/Users/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/multiprocess/pool.py", line 774, in get raise self._value TypeError: Couldn't cast array of type fixed_size_list<item: float>[2] to Sequence(feature=Value(dtype='float32', id=None), length=2, id=None) ``` ### Steps to reproduce the bug Here's a simple repro script: ``` from datasets import Features, Value, Sequence, ClassLabel, Dataset dataset_features = Features({ 'text': Value('string'), 'embedding': Sequence(Value('double'), length=2), 'categories': Sequence(ClassLabel(names=sorted([ 'one', 'two', 'three' ]))), }) dataset = Dataset.from_dict( { 'text': ['A'] * 10000, 'embedding': [[0.0, 0.1]] * 10000, 'categories': [[0]] * 10000, }, features=dataset_features ) def test_mapper(r): r['text'] = list(map(lambda t: t + ' b', r['text'])) return r dataset = dataset.map(test_mapper, batched=True, batch_size=10, features=dataset_features, num_proc=2) ``` Removing the embedding column fixes the issue! ### Expected behavior The mapping completes successfully. ### Environment info - `datasets` version: 2.14.4 - Platform: macOS-14.0-arm64-arm-64bit - Python version: 3.10.12 - Huggingface_hub version: 0.17.1 - PyArrow version: 13.0.0 - Pandas version: 2.0.3
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6280/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6280/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6279
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6279/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6279/comments
https://api.github.com/repos/huggingface/datasets/issues/6279/events
https://github.com/huggingface/datasets/issues/6279
1,928,028,226
I_kwDODunzps5y62BC
6,279
Batched IterableDataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/7010688?v=4", "events_url": "https://api.github.com/users/lneukom/events{/privacy}", "followers_url": "https://api.github.com/users/lneukom/followers", "following_url": "https://api.github.com/users/lneukom/following{/other_user}", "gists_url": "https://api.github.com/users/lneukom/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lneukom", "id": 7010688, "login": "lneukom", "node_id": "MDQ6VXNlcjcwMTA2ODg=", "organizations_url": "https://api.github.com/users/lneukom/orgs", "received_events_url": "https://api.github.com/users/lneukom/received_events", "repos_url": "https://api.github.com/users/lneukom/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lneukom/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lneukom/subscriptions", "type": "User", "url": "https://api.github.com/users/lneukom" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
1
"2023-10-05T11:12:49Z"
"2023-10-05T11:50:28Z"
null
NONE
null
null
null
### Feature request Hi, could you add an implementation of a batched `IterableDataset`. It already support an option to do batch iteration via `.iter(batch_size=...)` but this cannot be used in combination with a torch `DataLoader` since it just returns an iterator. ### Motivation The current implementation loads each element of a batch individually which can be very slow in cases of a big batch_size. I did some experiments [here](https://discuss.huggingface.co/t/slow-dataloader-with-big-batch-size/57224) and using a batched iteration would speed up data loading significantly. ### Your contribution N/A
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6279/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6279/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6278
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6278/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6278/comments
https://api.github.com/repos/huggingface/datasets/issues/6278/events
https://github.com/huggingface/datasets/pull/6278
1,927,957,877
PR_kwDODunzps5b_iKb
6,278
No data files duplicates
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
4
"2023-10-05T10:31:58Z"
"2024-01-11T06:32:49Z"
"2023-10-05T14:43:17Z"
MEMBER
null
1
{ "diff_url": "https://github.com/huggingface/datasets/pull/6278.diff", "html_url": "https://github.com/huggingface/datasets/pull/6278", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6278.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6278" }
I added a new DataFilesSet class to disallow duplicate data files. I also deprecated DataFilesList. EDIT: actually I might just add drop_duplicates=True to `.from_patterns` close https://github.com/huggingface/datasets/issues/6259 close https://github.com/huggingface/datasets/issues/6272 TODO: - [ ] tests - [ ] preserve data files order
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6278/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6278/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6277
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6277/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6277/comments
https://api.github.com/repos/huggingface/datasets/issues/6277/events
https://github.com/huggingface/datasets/issues/6277
1,927,044,546
I_kwDODunzps5y3F3C
6,277
FileNotFoundError: Couldn't find a module script at /content/paws-x/paws-x.py. Module 'paws-x' doesn't exist on the Hugging Face Hub either.
{ "avatar_url": "https://avatars.githubusercontent.com/u/66733346?v=4", "events_url": "https://api.github.com/users/diegogonzalezc/events{/privacy}", "followers_url": "https://api.github.com/users/diegogonzalezc/followers", "following_url": "https://api.github.com/users/diegogonzalezc/following{/other_user}", "gists_url": "https://api.github.com/users/diegogonzalezc/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/diegogonzalezc", "id": 66733346, "login": "diegogonzalezc", "node_id": "MDQ6VXNlcjY2NzMzMzQ2", "organizations_url": "https://api.github.com/users/diegogonzalezc/orgs", "received_events_url": "https://api.github.com/users/diegogonzalezc/received_events", "repos_url": "https://api.github.com/users/diegogonzalezc/repos", "site_admin": false, "starred_url": "https://api.github.com/users/diegogonzalezc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/diegogonzalezc/subscriptions", "type": "User", "url": "https://api.github.com/users/diegogonzalezc" }
[]
closed
false
null
[]
null
1
"2023-10-04T22:01:25Z"
"2023-10-08T17:05:46Z"
"2023-10-08T17:05:46Z"
NONE
null
null
null
### Describe the bug I'm encountering a "FileNotFoundError" while attempting to use the "paws-x" dataset to retrain the DistilRoBERTa-base model. The error message is as follows: FileNotFoundError: Couldn't find a module script at /content/paws-x/paws-x.py. Module 'paws-x' doesn't exist on the Hugging Face Hub either. ### Steps to reproduce the bug https://colab.research.google.com/drive/11xUUFxloClpmqLvDy_Xxfmo3oUzjY5nx#scrollTo=kUn74FigzhHm ### Expected behavior The the trained model ### Environment info colab, "paws-x" dataset , DistilRoBERTa-base model
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6277/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6277/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6276
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6276/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6276/comments
https://api.github.com/repos/huggingface/datasets/issues/6276/events
https://github.com/huggingface/datasets/issues/6276
1,925,961,878
I_kwDODunzps5yy9iW
6,276
I'm trying to fine tune the openai/whisper model from huggingface using jupyter notebook and i keep getting this error
{ "avatar_url": "https://avatars.githubusercontent.com/u/50768065?v=4", "events_url": "https://api.github.com/users/valaofficial/events{/privacy}", "followers_url": "https://api.github.com/users/valaofficial/followers", "following_url": "https://api.github.com/users/valaofficial/following{/other_user}", "gists_url": "https://api.github.com/users/valaofficial/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/valaofficial", "id": 50768065, "login": "valaofficial", "node_id": "MDQ6VXNlcjUwNzY4MDY1", "organizations_url": "https://api.github.com/users/valaofficial/orgs", "received_events_url": "https://api.github.com/users/valaofficial/received_events", "repos_url": "https://api.github.com/users/valaofficial/repos", "site_admin": false, "starred_url": "https://api.github.com/users/valaofficial/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/valaofficial/subscriptions", "type": "User", "url": "https://api.github.com/users/valaofficial" }
[]
open
false
null
[]
null
3
"2023-10-04T11:03:41Z"
"2023-11-27T10:39:16Z"
null
NONE
null
null
null
### Describe the bug I'm trying to fine tune the openai/whisper model from huggingface using jupyter notebook and i keep getting this error, i'm following the steps in this blog post https://huggingface.co/blog/fine-tune-whisper I tried google collab and it works but because I'm on the free version the training doesn't complete the error comes in jupyter notebook when i run this line `common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names["train"], num_proc=4)` here is the error message ``` Map (num_proc=4): 0% 0/2506 [00:52<?, ? examples/s] The above exception was the direct cause of the following exception: NameError Traceback (most recent call last) Cell In[19], line 1 ----> 1 common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names["train"], num_proc=4) File ~\anaconda\Lib\site-packages\datasets\dataset_dict.py:853, in DatasetDict.map(self, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc) 850 if cache_file_names is None: 851 cache_file_names = {k: None for k in self} 852 return DatasetDict( --> 853 { 854 k: dataset.map( 855 function=function, 856 with_indices=with_indices, 857 with_rank=with_rank, 858 input_columns=input_columns, 859 batched=batched, 860 batch_size=batch_size, 861 drop_last_batch=drop_last_batch, 862 remove_columns=remove_columns, 863 keep_in_memory=keep_in_memory, 864 load_from_cache_file=load_from_cache_file, 865 cache_file_name=cache_file_names[k], 866 writer_batch_size=writer_batch_size, 867 features=features, 868 disable_nullable=disable_nullable, 869 fn_kwargs=fn_kwargs, 870 num_proc=num_proc, 871 desc=desc, 872 ) 873 for k, dataset in self.items() 874 } 875 ) File ~\anaconda\Lib\site-packages\datasets\dataset_dict.py:854, in <dictcomp>(.0) 850 if cache_file_names is None: 851 cache_file_names = {k: None for k in self} 852 return DatasetDict( 853 { --> 854 k: dataset.map( 855 function=function, 856 with_indices=with_indices, 857 with_rank=with_rank, 858 input_columns=input_columns, 859 batched=batched, 860 batch_size=batch_size, 861 drop_last_batch=drop_last_batch, 862 remove_columns=remove_columns, 863 keep_in_memory=keep_in_memory, 864 load_from_cache_file=load_from_cache_file, 865 cache_file_name=cache_file_names[k], 866 writer_batch_size=writer_batch_size, 867 features=features, 868 disable_nullable=disable_nullable, 869 fn_kwargs=fn_kwargs, 870 num_proc=num_proc, 871 desc=desc, 872 ) 873 for k, dataset in self.items() 874 } 875 ) File ~\anaconda\Lib\site-packages\datasets\arrow_dataset.py:592, in transmit_tasks.<locals>.wrapper(*args, **kwargs) 590 self: "Dataset" = kwargs.pop("self") 591 # apply actual function --> 592 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 593 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 594 for dataset in datasets: 595 # Remove task templates if a column mapping of the template is no longer valid File ~\anaconda\Lib\site-packages\datasets\arrow_dataset.py:557, in transmit_format.<locals>.wrapper(*args, **kwargs) 550 self_format = { 551 "type": self._format_type, 552 "format_kwargs": self._format_kwargs, 553 "columns": self._format_columns, 554 "output_all_columns": self._output_all_columns, 555 } 556 # apply actual function --> 557 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 558 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 559 # re-apply format to the output File ~\anaconda\Lib\site-packages\datasets\arrow_dataset.py:3189, in Dataset.map(self, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 3182 logger.info(f"Spawning {num_proc} processes") 3183 with logging.tqdm( 3184 disable=not logging.is_progress_bar_enabled(), 3185 unit=" examples", 3186 total=pbar_total, 3187 desc=(desc or "Map") + f" (num_proc={num_proc})", 3188 ) as pbar: -> 3189 for rank, done, content in iflatmap_unordered( 3190 pool, Dataset._map_single, kwargs_iterable=kwargs_per_job 3191 ): 3192 if done: 3193 shards_done += 1 File ~\anaconda\Lib\site-packages\datasets\utils\py_utils.py:1394, in iflatmap_unordered(pool, func, kwargs_iterable) 1391 finally: 1392 if not pool_changed: 1393 # we get the result in case there's an error to raise -> 1394 [async_result.get(timeout=0.05) for async_result in async_results] File ~\anaconda\Lib\site-packages\datasets\utils\py_utils.py:1394, in <listcomp>(.0) 1391 finally: 1392 if not pool_changed: 1393 # we get the result in case there's an error to raise -> 1394 [async_result.get(timeout=0.05) for async_result in async_results] File ~\anaconda\Lib\site-packages\multiprocess\pool.py:774, in ApplyResult.get(self, timeout) 772 return self._value 773 else: --> 774 raise self._value NameError: name 'feature_extractor' is not defined ``` ### Steps to reproduce the bug 1. follow the steps in this blog post https://huggingface.co/blog/fine-tune-whisper 2. run this line of code `common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names["train"], num_proc=4)` 3. I'm using jupyter notebook from anaconda ### Expected behavior No error message ### Environment info datasets version: 2.8.0 Python version: 3.11 Windows 10
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6276/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6276/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6275
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6275/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6275/comments
https://api.github.com/repos/huggingface/datasets/issues/6275/events
https://github.com/huggingface/datasets/issues/6275
1,921,354,680
I_kwDODunzps5yhYu4
6,275
Would like to Contribute a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/97907750?v=4", "events_url": "https://api.github.com/users/vikas70607/events{/privacy}", "followers_url": "https://api.github.com/users/vikas70607/followers", "following_url": "https://api.github.com/users/vikas70607/following{/other_user}", "gists_url": "https://api.github.com/users/vikas70607/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vikas70607", "id": 97907750, "login": "vikas70607", "node_id": "U_kgDOBdX0Jg", "organizations_url": "https://api.github.com/users/vikas70607/orgs", "received_events_url": "https://api.github.com/users/vikas70607/received_events", "repos_url": "https://api.github.com/users/vikas70607/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vikas70607/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vikas70607/subscriptions", "type": "User", "url": "https://api.github.com/users/vikas70607" }
[]
closed
false
null
[]
null
1
"2023-10-02T07:00:21Z"
"2023-10-10T16:27:54Z"
"2023-10-10T16:27:54Z"
NONE
null
null
null
I have a dataset of 2500 images that can be used for color-blind machine-learning algorithms. Since , there was no dataset available online , I made this dataset myself and would like to contribute this now to community
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6275/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6275/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6274
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6274/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6274/comments
https://api.github.com/repos/huggingface/datasets/issues/6274/events
https://github.com/huggingface/datasets/issues/6274
1,921,036,328
I_kwDODunzps5ygLAo
6,274
FileNotFoundError for dataset with multiple builder config
{ "avatar_url": "https://avatars.githubusercontent.com/u/97120485?v=4", "events_url": "https://api.github.com/users/LouisChen15/events{/privacy}", "followers_url": "https://api.github.com/users/LouisChen15/followers", "following_url": "https://api.github.com/users/LouisChen15/following{/other_user}", "gists_url": "https://api.github.com/users/LouisChen15/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LouisChen15", "id": 97120485, "login": "LouisChen15", "node_id": "U_kgDOBcnw5Q", "organizations_url": "https://api.github.com/users/LouisChen15/orgs", "received_events_url": "https://api.github.com/users/LouisChen15/received_events", "repos_url": "https://api.github.com/users/LouisChen15/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LouisChen15/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LouisChen15/subscriptions", "type": "User", "url": "https://api.github.com/users/LouisChen15" }
[]
closed
false
null
[]
null
1
"2023-10-01T23:45:56Z"
"2023-10-02T20:09:38Z"
"2023-10-02T20:09:38Z"
NONE
null
null
null
### Describe the bug When there is only one config and only the dataset name is entered when using datasets.load_dataset(), it works fine. But if I create a second builder_config for my dataset and enter the config name when using datasets.load_dataset(), the following error will happen. FileNotFoundError: [Errno 2] No such file or directory: 'C:/Users/chenx/.cache/huggingface/datasets/my_dataset/0_shot_multiple_choice/1.0.0/97c3854a012cfd6b045e3be4c864739902af2d818bb9235b047baa94c302e9a2.incomplete/my_dataset-test-00000-00000-of-NNNNN.arrow' The "XXX.incomplete folder" in the cache folder of my dataset will disappear before "generating test split", which does not happen when config name is not entered and the config name is "default" C:\Users\chenx\.cache\huggingface\datasets\my_dataset\0_shot_multiple_choice\1.0.0 The folder that is supposed to remain under the above directory will disappear, and the data generator will not have a place to generate data into. ### Steps to reproduce the bug test = load_dataset('my_dataset', '0_shot_multiple_choice') ### Expected behavior FileNotFoundError: [Errno 2] No such file or directory: 'C:/Users/chenx/.cache/huggingface/datasets/my_dataset/0_shot_multiple_choice/1.0.0/97c3854a012cfd6b045e3be4c864739902af2d818bb9235b047baa94c302e9a2.incomplete/my_dataset-test-00000-00000-of-NNNNN.arrow' ### Environment info datasets 2.14.5 python 3.8.18
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6274/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6274/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6273
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6273/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6273/comments
https://api.github.com/repos/huggingface/datasets/issues/6273/events
https://github.com/huggingface/datasets/issues/6273
1,920,922,260
I_kwDODunzps5yfvKU
6,273
Broken Link to PubMed Abstracts dataset .
{ "avatar_url": "https://avatars.githubusercontent.com/u/100606327?v=4", "events_url": "https://api.github.com/users/sameemqureshi/events{/privacy}", "followers_url": "https://api.github.com/users/sameemqureshi/followers", "following_url": "https://api.github.com/users/sameemqureshi/following{/other_user}", "gists_url": "https://api.github.com/users/sameemqureshi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sameemqureshi", "id": 100606327, "login": "sameemqureshi", "node_id": "U_kgDOBf8hdw", "organizations_url": "https://api.github.com/users/sameemqureshi/orgs", "received_events_url": "https://api.github.com/users/sameemqureshi/received_events", "repos_url": "https://api.github.com/users/sameemqureshi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sameemqureshi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sameemqureshi/subscriptions", "type": "User", "url": "https://api.github.com/users/sameemqureshi" }
[]
open
false
null
[]
null
4
"2023-10-01T19:08:48Z"
"2024-01-09T05:48:01Z"
null
NONE
null
null
null
### Describe the bug The link provided for the dataset is broken, data_files = [https://the-eye.eu/public/AI/pile_preliminary_components/PUBMED_title_abstracts_2019_baseline.jsonl.zst](url) The ### Steps to reproduce the bug Steps to reproduce: 1) Head over to [https://huggingface.co/learn/nlp-course/chapter5/4?fw=pt#big-data-datasets-to-the-rescue](url) 2) In the Section "What is the Pile?", you can see a code snippet that contains the broken link. ### Expected behavior The link should Redirect to the "PubMed Abstracts dataset" as expected . ### Environment info .
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6273/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6273/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6272
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6272/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6272/comments
https://api.github.com/repos/huggingface/datasets/issues/6272/events
https://github.com/huggingface/datasets/issues/6272
1,920,831,487
I_kwDODunzps5yfY__
6,272
Duplicate `data_files` when named `<split>/<split>.parquet`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
open
false
null
[]
null
7
"2023-10-01T15:43:56Z"
"2023-10-05T10:32:27Z"
null
MEMBER
null
null
null
e.g. with `u23429/stock_1_minute_ticker` ```ipython In [1]: from datasets import * In [2]: b = load_dataset_builder("u23429/stock_1_minute_ticker") Downloading readme: 100%|██████████████████████████| 627/627 [00:00<00:00, 246kB/s] In [3]: b.config.data_files Out[3]: {NamedSplit('train'): ['hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/train/train.parquet', 'hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/train/train.parquet'], NamedSplit('validation'): ['hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/validation/validation.parquet', 'hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/validation/validation.parquet'], NamedSplit('test'): ['hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/test/test.parquet', 'hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/test/test.parquet']} ``` This bug issue is present in the current `datasets` 2.14.5 and also on `main` even after https://github.com/huggingface/datasets/pull/6244 cc @mariosasko
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6272/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6272/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6271
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6271/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6271/comments
https://api.github.com/repos/huggingface/datasets/issues/6271/events
https://github.com/huggingface/datasets/issues/6271
1,920,420,295
I_kwDODunzps5yd0nH
6,271
Overwriting Split overwrites data but not metadata, corrupting dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/13859249?v=4", "events_url": "https://api.github.com/users/govindrai/events{/privacy}", "followers_url": "https://api.github.com/users/govindrai/followers", "following_url": "https://api.github.com/users/govindrai/following{/other_user}", "gists_url": "https://api.github.com/users/govindrai/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/govindrai", "id": 13859249, "login": "govindrai", "node_id": "MDQ6VXNlcjEzODU5MjQ5", "organizations_url": "https://api.github.com/users/govindrai/orgs", "received_events_url": "https://api.github.com/users/govindrai/received_events", "repos_url": "https://api.github.com/users/govindrai/repos", "site_admin": false, "starred_url": "https://api.github.com/users/govindrai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/govindrai/subscriptions", "type": "User", "url": "https://api.github.com/users/govindrai" }
[]
closed
false
null
[]
null
0
"2023-09-30T22:37:31Z"
"2023-10-16T13:30:50Z"
"2023-10-16T13:30:50Z"
NONE
null
null
null
### Describe the bug I want to be able to overwrite/update/delete splits in my dataset. Currently the only way to do is to manually go into the dataset and delete the split. If I try to overwrite programmatically I end up in an error state and (somewhat) corrupting the dataset. Read below. **Current Behavior** When I push to an existing split I get this error: `ValueError: Split complexRoofLocation_01Apr2023_to_31May2023test already present` This seems to suggest that the library doesn't support overwriting splits. **Potential Bug** What’s strange is that datasets, despite the operation erroring out with the ValueError above, does, in fact, overwrite the split: `Pushing dataset shards to the dataset hub: 100% [.....................] 1/1 [00:00<00:00, 55.04it/s]` Even though you got an error message and your code fails, your dataset is now changed. That seems like a bug. Either don't change the dataset, or don't throw the error and allow the script to proceed. Additional Bug While it overwrites the split, it doesn’t overwrite the split’s information. Because of this when you pull down the dataset you may end up getting a `NonMatchingSplitsSizesError` if the size of the dataset during the overwrite is different. For example, my original split had 5 rows, but on my overwrite, I only had 4. Then when I try to download the dataset, I get a `NonMatchingSplitsSizesError` because the dataset's data.json states there’s 5 but only 4 exist in the split. Expected Behavior This corrupts the dataset rendering it unusable (until you take manual intervention). Either the library should let the overwrite happen (which it does but should also update the metadata) or it shouldn’t do anything. ### Steps to reproduce the bug [Colab Notebook](https://colab.research.google.com/drive/1bqVkD06Ngs9MQNdSk_ygCG6y1UqXA4pC?usp=sharing) ### Expected behavior The split should be overwritten and I should be able to use the new version of the dataset without issue. ### Environment info - `datasets` version: 2.14.5 - Platform: Linux-5.15.120+-x86_64-with-glibc2.35 - Python version: 3.10.12 - Huggingface_hub version: 0.17.3 - PyArrow version: 9.0.0 - Pandas version: 1.5.3
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6271/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6271/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6270
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6270/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6270/comments
https://api.github.com/repos/huggingface/datasets/issues/6270/events
https://github.com/huggingface/datasets/issues/6270
1,920,329,373
I_kwDODunzps5ydead
6,270
Dataset.from_generator raises with sharded gen_args
{ "avatar_url": "https://avatars.githubusercontent.com/u/53510?v=4", "events_url": "https://api.github.com/users/hartmans/events{/privacy}", "followers_url": "https://api.github.com/users/hartmans/followers", "following_url": "https://api.github.com/users/hartmans/following{/other_user}", "gists_url": "https://api.github.com/users/hartmans/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hartmans", "id": 53510, "login": "hartmans", "node_id": "MDQ6VXNlcjUzNTEw", "organizations_url": "https://api.github.com/users/hartmans/orgs", "received_events_url": "https://api.github.com/users/hartmans/received_events", "repos_url": "https://api.github.com/users/hartmans/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hartmans/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hartmans/subscriptions", "type": "User", "url": "https://api.github.com/users/hartmans" }
[]
closed
false
null
[]
null
6
"2023-09-30T16:50:06Z"
"2023-10-11T20:29:12Z"
"2023-10-11T20:29:11Z"
CONTRIBUTOR
null
null
null
### Describe the bug According to the docs of Datasets.from_generator: ``` gen_kwargs(`dict`, *optional*): Keyword arguments to be passed to the `generator` callable. You can define a sharded dataset by passing the list of shards in `gen_kwargs`. ``` So I'd expect that if gen_kwargs was a list, then my generator would be called once for each element in the list with the dict in the list for that element. It doesn't work that way though. ### Steps to reproduce the bug ```python #!/usr/bin/python from pathlib import Path import datasets def process_yaml(file): yield dict(example=42) if __name__ == '__main__': import sys dir = Path(sys.argv[0]).parent ds = datasets.Dataset.from_generator(process_yaml, gen_kwargs=[{'file':f} for f in dir.glob('*.yml')], ) ds.to_json('training.jsonl') ``` ``` Generating train split: 0 examples [00:00, ? examples/s] Traceback (most recent call last): File "/tmp/dataset_bug.py", line 13, in <module> ds = datasets.Dataset.from_generator(process_yaml, gen_kwargs=[{'file':f} for f in dir.glob('*.yml')], ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/arrow_dataset.py", line 1072, in from_generator ).read() ^^^^^^ File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/io/generator.py", line 47, in read self.builder.download_and_prepare( File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 954, in download_and_prepare self._download_and_prepare( File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 1717, in _download_and_prepare super()._download_and_prepare( File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 1049, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 1555, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 1656, in _prepare_split_single generator = self._generate_examples(**gen_kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ TypeError: datasets.packaged_modules.generator.generator.Generator._generate_examples() argument after ** must be a ``` mapping, not list ### Expected behavior I would expect that process_yaml would be called once for each yaml file in the directory where the script is run. I also tried with the list being in gen_kwargs, but in that case process_yaml gets called with a list. ### Environment info - `datasets` version: 2.14.6.dev0 (git commit 0cc77d7f45c7369; also tested with 2.14.0) - Platform: Linux-6.1.0-10-amd64-x86_64-with-glibc2.36 - Python version: 3.11.2 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6270/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6270/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6269
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6269/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6269/comments
https://api.github.com/repos/huggingface/datasets/issues/6269/events
https://github.com/huggingface/datasets/pull/6269
1,919,572,790
PR_kwDODunzps5bjbDc
6,269
Reduce the number of commits in `push_to_hub`
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
21
"2023-09-29T16:22:31Z"
"2023-10-16T16:03:18Z"
"2023-10-16T13:30:46Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6269.diff", "html_url": "https://github.com/huggingface/datasets/pull/6269", "merged_at": "2023-10-16T13:30:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/6269.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6269" }
Reduces the number of commits in `push_to_hub` by using the `preupload` API from https://github.com/huggingface/huggingface_hub/pull/1699. Each commit contains a maximum of 50 uploaded files. A shard's fingerprint no longer needs to be added as a suffix to support resuming an upload, meaning the shards' naming scheme is the same as the initial one. Also, it adds support for the following params: `create_pr`, `commit_message` and `revision` (`branch` deprecated; unlike the previous implementation, this one creates a branch if the branch does not exist to be consistent with `transformers`). (Nit) This implementation keeps the markdown section of the generated README.md empty to enable importing the card template (when the card is accessed on the Hub). Fixes https://github.com/huggingface/datasets/issues/5492, fixes https://github.com/huggingface/datasets/issues/6257, fixes https://github.com/huggingface/datasets/issues/5045, fixes https://github.com/huggingface/datasets/issues/6271 TODO: - [x] set the minimal version to the next `hfh` release (once it's published)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 3, "laugh": 0, "rocket": 1, "total_count": 4, "url": "https://api.github.com/repos/huggingface/datasets/issues/6269/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6269/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6268
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6268/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6268/comments
https://api.github.com/repos/huggingface/datasets/issues/6268/events
https://github.com/huggingface/datasets/pull/6268
1,919,010,645
PR_kwDODunzps5bhgs7
6,268
Add repo_id to DatasetInfo
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
open
false
null
[]
null
9
"2023-09-29T10:24:55Z"
"2023-10-01T15:29:45Z"
null
MEMBER
null
1
{ "diff_url": "https://github.com/huggingface/datasets/pull/6268.diff", "html_url": "https://github.com/huggingface/datasets/pull/6268", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6268.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6268" }
```python from datasets import load_dataset ds = load_dataset("lhoestq/demo1", split="train") ds = ds.map(lambda x: {}, num_proc=2).filter(lambda x: True).remove_columns(["id"]) print(ds.repo_id) # lhoestq/demo1 ``` - repo_id is None when the dataset doesn't come from the Hub, e.g. from Dataset.from_dict - repo_id is set to None when concatenating datasets with different repo ids related to https://github.com/huggingface/datasets/issues/4129 TODO: - [ ] discuss if it's ok for now - [ ] tests
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/6268/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6268/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6267
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6267/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6267/comments
https://api.github.com/repos/huggingface/datasets/issues/6267/events
https://github.com/huggingface/datasets/issues/6267
1,916,443,262
I_kwDODunzps5yOpp-
6,267
Multi label class encoding
{ "avatar_url": "https://avatars.githubusercontent.com/u/1000442?v=4", "events_url": "https://api.github.com/users/jmif/events{/privacy}", "followers_url": "https://api.github.com/users/jmif/followers", "following_url": "https://api.github.com/users/jmif/following{/other_user}", "gists_url": "https://api.github.com/users/jmif/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jmif", "id": 1000442, "login": "jmif", "node_id": "MDQ6VXNlcjEwMDA0NDI=", "organizations_url": "https://api.github.com/users/jmif/orgs", "received_events_url": "https://api.github.com/users/jmif/received_events", "repos_url": "https://api.github.com/users/jmif/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jmif/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmif/subscriptions", "type": "User", "url": "https://api.github.com/users/jmif" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
7
"2023-09-27T22:48:08Z"
"2023-10-26T18:46:08Z"
null
NONE
null
null
null
### Feature request I have a multi label dataset and I'd like to be able to class encode the column and store the mapping directly in the features just as I can with a single label column. `class_encode_column` currently does not support multi labels. Here's an example of what I'd like to encode: ``` data = { 'text': ['one', 'two', 'three', 'four'], 'labels': [['a', 'b'], ['b'], ['b', 'c'], ['a', 'd']] } dataset = Dataset.from_dict(data) dataset = dataset.class_encode_column('labels') ``` I did some digging into the code base to evaluate the feasibility of this (note I'm very new to this code base) and from what I noticed the `ClassLabel` feature is still stored as an underlying raw data type of int so I thought a `MultiLabel` feature could similarly be stored as a Sequence of ints, thus not requiring significant serialization / conversion work to / from arrow. I did a POC of this [here](https://github.com/huggingface/datasets/commit/15443098e9ce053943172f7ec6fce3769d7dff6e) and included a simple test case (please excuse all the commented out tests, going for speed of POC here and didn't want to fight IDE to debug a single test). In the test I just assert that `num_classes` is the same to show that things are properly serializing, but if you break after loading from disk you'll see the dataset correct and the dataset feature is as expected. After digging more I did notice a few issues - After loading from disk I noticed type of the `labels` class is `Sequence` not `MultiLabel` (though the added `feature` attribute came through). This doesn't happen for `ClassLabel` but I couldn't find the encode / decode code paths that handle this. - I subclass `Sequence` in `MultiLabel` to leverage existing serialization, but this does miss the custom encode logic that `ClassLabel` has. I'm not sure of the best way to approach this as I haven't fully understood the encode / decode flow for datasets. I suspect my simple implementation will need some improvement as it'll require a significant amount of repeated logic to mimic `ClassLabel` behavior. ### Motivation See above - would like to support multi label class encodings. ### Your contribution This would be a big help for us and we're open to contributing but I'll likely need some guidance on how to implement to fit the encode / decode flow. Some suggestions on tests / would be great too, I'm guessing in addition to the class encode tests (that I'll need to expand) we'll need encode / decode tests.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6267/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6267/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6266
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6266/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6266/comments
https://api.github.com/repos/huggingface/datasets/issues/6266/events
https://github.com/huggingface/datasets/pull/6266
1,916,334,394
PR_kwDODunzps5bYYb8
6,266
Use LibYAML with PyYAML if available
{ "avatar_url": "https://avatars.githubusercontent.com/u/3905501?v=4", "events_url": "https://api.github.com/users/bryant1410/events{/privacy}", "followers_url": "https://api.github.com/users/bryant1410/followers", "following_url": "https://api.github.com/users/bryant1410/following{/other_user}", "gists_url": "https://api.github.com/users/bryant1410/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bryant1410", "id": 3905501, "login": "bryant1410", "node_id": "MDQ6VXNlcjM5MDU1MDE=", "organizations_url": "https://api.github.com/users/bryant1410/orgs", "received_events_url": "https://api.github.com/users/bryant1410/received_events", "repos_url": "https://api.github.com/users/bryant1410/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bryant1410/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bryant1410/subscriptions", "type": "User", "url": "https://api.github.com/users/bryant1410" }
[]
open
false
null
[]
null
5
"2023-09-27T21:13:36Z"
"2023-09-28T14:29:24Z"
null
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6266.diff", "html_url": "https://github.com/huggingface/datasets/pull/6266", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6266.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6266" }
PyYAML, the YAML framework used in this library, allows the use of LibYAML to accelerate the methods `load` and `dump`. To use it, a user would need to first install a PyYAML version that uses LibYAML (not available in PyPI; needs to be manually installed). Then, to actually use them, PyYAML suggests importing the LibYAML version of the `Loader` and `Dumper` and falling back to the default ones. This PR implements this change. See [PyYAML docs](https://pyyaml.org/wiki/PyYAMLDocumentation) for more info. This change was motivated after trying to use any of [the SugarCREPE datasets in the Hub](https://huggingface.co/datasets?search=sugarcrepe) provided by [the org HuggingFaceM4](https://huggingface.co/datasets/HuggingFaceM4). Such datasets save a lot of information (~1MB) in the YAML metadata from the `README.md` file and I noticed this slowed down the data loading process. BTW, I also noticed cache files for it is also slow because it tries to hash an instance of `DatasetInfo`, which in turn has all this metadata. Also, I changed two list comprehensions into generator expressions to avoid allocating extra memory unnecessarily. And BTW, there's [an issue in PyYAML suggesting to make this automatic](https://github.com/yaml/pyyaml/issues/437).
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6266/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6266/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6265
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6265/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6265/comments
https://api.github.com/repos/huggingface/datasets/issues/6265/events
https://github.com/huggingface/datasets/pull/6265
1,915,651,566
PR_kwDODunzps5bWDfc
6,265
Remove `apache_beam` import in `BeamBasedBuilder._save_info`
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
4
"2023-09-27T13:56:34Z"
"2023-09-28T18:34:02Z"
"2023-09-28T18:23:35Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6265.diff", "html_url": "https://github.com/huggingface/datasets/pull/6265", "merged_at": "2023-09-28T18:23:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/6265.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6265" }
... to avoid an `ImportError` raised in `BeamBasedBuilder._save_info` when `apache_beam` is not installed (e.g., when downloading the processed version of a dataset from the HF GCS) Fix https://github.com/huggingface/datasets/issues/6260
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6265/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6265/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6264
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6264/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6264/comments
https://api.github.com/repos/huggingface/datasets/issues/6264/events
https://github.com/huggingface/datasets/pull/6264
1,914,958,781
PR_kwDODunzps5bTvzh
6,264
Temporarily pin tensorflow < 2.14.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
4
"2023-09-27T08:16:06Z"
"2023-09-27T08:45:24Z"
"2023-09-27T08:36:39Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6264.diff", "html_url": "https://github.com/huggingface/datasets/pull/6264", "merged_at": "2023-09-27T08:36:39Z", "patch_url": "https://github.com/huggingface/datasets/pull/6264.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6264" }
Temporarily pin tensorflow < 2.14.0 until permanent solution is found. Hot fix #6263.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6264/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6264/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6263
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6263/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6263/comments
https://api.github.com/repos/huggingface/datasets/issues/6263/events
https://github.com/huggingface/datasets/issues/6263
1,914,951,043
I_kwDODunzps5yI9WD
6,263
CI is broken: ImportError: cannot import name 'context' from 'tensorflow.python'
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
0
"2023-09-27T08:12:05Z"
"2023-09-27T08:36:40Z"
"2023-09-27T08:36:40Z"
MEMBER
null
null
null
Python 3.10 CI is broken for `test_py310`. See: https://github.com/huggingface/datasets/actions/runs/6322990957/job/17169678812?pr=6262 ``` FAILED tests/test_py_utils.py::TempSeedTest::test_tensorflow - ImportError: cannot import name 'context' from 'tensorflow.python' (/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/tensorflow/python/__init__.py) ``` ``` _________________________ TempSeedTest.test_tensorflow _________________________ [gw1] linux -- Python 3.10.13 /opt/hostedtoolcache/Python/3.10.13/x64/bin/python self = <tests.test_py_utils.TempSeedTest testMethod=test_tensorflow> @require_tf def test_tensorflow(self): import tensorflow as tf from tensorflow.keras import layers model = layers.Dense(2) def gen_random_output(): x = tf.random.uniform((1, 3)) return model(x).numpy() > with temp_seed(42, set_tensorflow=True): tests/test_py_utils.py:155: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/contextlib.py:135: in __enter__ return next(self.gen) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ seed = 42, set_pytorch = False, set_tensorflow = True @contextmanager def temp_seed(seed: int, set_pytorch=False, set_tensorflow=False): """Temporarily set the random seed. This works for python numpy, pytorch and tensorflow.""" np_state = np.random.get_state() np.random.seed(seed) if set_pytorch and config.TORCH_AVAILABLE: import torch torch_state = torch.random.get_rng_state() torch.random.manual_seed(seed) if torch.cuda.is_available(): torch_cuda_states = torch.cuda.get_rng_state_all() torch.cuda.manual_seed_all(seed) if set_tensorflow and config.TF_AVAILABLE: import tensorflow as tf > from tensorflow.python import context as tfpycontext E ImportError: cannot import name 'context' from 'tensorflow.python' (/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/tensorflow/python/__init__.py) /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/datasets/utils/py_utils.py:257: ImportError ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6263/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6263/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6262
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6262/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6262/comments
https://api.github.com/repos/huggingface/datasets/issues/6262/events
https://github.com/huggingface/datasets/pull/6262
1,914,895,459
PR_kwDODunzps5bTh6H
6,262
Fix CI 404 errors
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
9
"2023-09-27T07:40:18Z"
"2023-09-28T15:39:16Z"
"2023-09-28T15:30:40Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6262.diff", "html_url": "https://github.com/huggingface/datasets/pull/6262", "merged_at": "2023-09-28T15:30:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/6262.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6262" }
Currently our CI usually raises 404 errors when trying to delete temporary repositories. See, e.g.: https://github.com/huggingface/datasets/actions/runs/6314980985/job/17146507884 ``` FAILED tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub_multiple_files_with_max_shard_size - huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6512fb99-4a52c561752ece3d77eb6d57;2b61cae4-613d-4a73-bbb1-2faf9e32b02d) Repository Not Found for url: https://hub-ci.huggingface.co/api/repos/delete. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. FAILED tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_to_hub_custom_features_audio - huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6512fbb2-0333dd666d42f0e173c2bb68;dfdc4271-b49b-4008-8c49-f05cf7c1d53d) Repository Not Found for url: https://hub-ci.huggingface.co/api/repos/delete. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. FAILED tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub_custom_splits - huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6512fbca-167690694f39770a5b3a444e;baeaa905-0a57-4585-ac97-9aaae12dd47d) Repository Not Found for url: https://hub-ci.huggingface.co/api/repos/delete. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. ``` I think this can be caused by collisions in temporary repository IDs because we create them in multiprocessing: ```python with temporary_repo(f"{CI_HUB_USER}/test-{int(time.time() * 10e3)}") as ds_name: ``` This can also be caused when there is another issue that does not allow the creation of the repository, thus making it impossible to delete it. This PR tries to fix this issue by increasing the precision of the number on the repository ID: `10e6` instead of `10e3`. Additionally, this PR catches RepositoryNotFoundError.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6262/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6262/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6261
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6261/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6261/comments
https://api.github.com/repos/huggingface/datasets/issues/6261/events
https://github.com/huggingface/datasets/issues/6261
1,913,813,178
I_kwDODunzps5yEni6
6,261
Can't load a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/37955817?v=4", "events_url": "https://api.github.com/users/joaopedrosdmm/events{/privacy}", "followers_url": "https://api.github.com/users/joaopedrosdmm/followers", "following_url": "https://api.github.com/users/joaopedrosdmm/following{/other_user}", "gists_url": "https://api.github.com/users/joaopedrosdmm/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/joaopedrosdmm", "id": 37955817, "login": "joaopedrosdmm", "node_id": "MDQ6VXNlcjM3OTU1ODE3", "organizations_url": "https://api.github.com/users/joaopedrosdmm/orgs", "received_events_url": "https://api.github.com/users/joaopedrosdmm/received_events", "repos_url": "https://api.github.com/users/joaopedrosdmm/repos", "site_admin": false, "starred_url": "https://api.github.com/users/joaopedrosdmm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joaopedrosdmm/subscriptions", "type": "User", "url": "https://api.github.com/users/joaopedrosdmm" }
[]
closed
false
null
[]
null
5
"2023-09-26T15:46:25Z"
"2023-10-05T10:23:23Z"
"2023-10-05T10:23:22Z"
NONE
null
null
null
### Describe the bug Can't seem to load the JourneyDB dataset. It throws the following error: ``` --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) Cell In[15], line 2 1 # If the dataset is gated/private, make sure you have run huggingface-cli login ----> 2 dataset = load_dataset("JourneyDB/JourneyDB", data_files="data", use_auth_token=True) File /opt/conda/lib/python3.10/site-packages/datasets/load.py:1664, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) 1661 ignore_verifications = ignore_verifications or save_infos 1663 # Create a dataset builder -> 1664 builder_instance = load_dataset_builder( 1665 path=path, 1666 name=name, 1667 data_dir=data_dir, 1668 data_files=data_files, 1669 cache_dir=cache_dir, 1670 features=features, 1671 download_config=download_config, 1672 download_mode=download_mode, 1673 revision=revision, 1674 use_auth_token=use_auth_token, 1675 **config_kwargs, 1676 ) 1678 # Return iterable dataset in case of streaming 1679 if streaming: File /opt/conda/lib/python3.10/site-packages/datasets/load.py:1490, in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, use_auth_token, **config_kwargs) 1488 download_config = download_config.copy() if download_config else DownloadConfig() 1489 download_config.use_auth_token = use_auth_token -> 1490 dataset_module = dataset_module_factory( 1491 path, 1492 revision=revision, 1493 download_config=download_config, 1494 download_mode=download_mode, 1495 data_dir=data_dir, 1496 data_files=data_files, 1497 ) 1499 # Get dataset builder class from the processing script 1500 builder_cls = import_main_class(dataset_module.module_path) File /opt/conda/lib/python3.10/site-packages/datasets/load.py:1238, in dataset_module_factory(path, revision, download_config, download_mode, force_local_path, dynamic_modules_path, data_dir, data_files, **download_kwargs) 1236 raise ConnectionError(f"Couln't reach the Hugging Face Hub for dataset '{path}': {e1}") from None 1237 if isinstance(e1, FileNotFoundError): -> 1238 raise FileNotFoundError( 1239 f"Couldn't find a dataset script at {relative_to_absolute_path(combined_path)} or any data file in the same directory. " 1240 f"Couldn't find '{path}' on the Hugging Face Hub either: {type(e1).__name__}: {e1}" 1241 ) from None 1242 raise e1 from None 1243 else: FileNotFoundError: Couldn't find a dataset script at /kaggle/working/JourneyDB/JourneyDB/JourneyDB.py or any data file in the same directory. Couldn't find 'JourneyDB/JourneyDB' on the Hugging Face Hub either: FileNotFoundError: Unable to find data in dataset repository JourneyDB/JourneyDB with any supported extension ['csv', 'tsv', 'json', 'jsonl', 'parquet', 'txt', 'blp', 'bmp', 'dib', 'bufr', 'cur', 'pcx', 'dcx', 'dds', 'ps', 'eps', 'fit', 'fits', 'fli', 'flc', 'ftc', 'ftu', 'gbr', 'gif', 'grib', 'h5', 'hdf', 'png', 'apng', 'jp2', 'j2k', 'jpc', 'jpf', 'jpx', 'j2c', 'icns', 'ico', 'im', 'iim', 'tif', 'tiff', 'jfif', 'jpe', 'jpg', 'jpeg', 'mpg', 'mpeg', 'msp', 'pcd', 'pxr', 'pbm', 'pgm', 'ppm', 'pnm', 'psd', 'bw', 'rgb', 'rgba', 'sgi', 'ras', 'tga', 'icb', 'vda', 'vst', 'webp', 'wmf', 'emf', 'xbm', 'xpm', 'zip'] ``` ### Steps to reproduce the bug 1) ``` from huggingface_hub import notebook_login notebook_login() ``` 2) ``` !pip install -q datasets from datasets import load_dataset ``` 3) `dataset = load_dataset("JourneyDB/JourneyDB", data_files="data", use_auth_token=True)` ### Expected behavior Load the dataset ### Environment info Notebook
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6261/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6261/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6260
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6260/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6260/comments
https://api.github.com/repos/huggingface/datasets/issues/6260/events
https://github.com/huggingface/datasets/issues/6260
1,912,593,466
I_kwDODunzps5x_9w6
6,260
REUSE_DATASET_IF_EXISTS don't work
{ "avatar_url": "https://avatars.githubusercontent.com/u/88258534?v=4", "events_url": "https://api.github.com/users/rangehow/events{/privacy}", "followers_url": "https://api.github.com/users/rangehow/followers", "following_url": "https://api.github.com/users/rangehow/following{/other_user}", "gists_url": "https://api.github.com/users/rangehow/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rangehow", "id": 88258534, "login": "rangehow", "node_id": "MDQ6VXNlcjg4MjU4NTM0", "organizations_url": "https://api.github.com/users/rangehow/orgs", "received_events_url": "https://api.github.com/users/rangehow/received_events", "repos_url": "https://api.github.com/users/rangehow/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rangehow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rangehow/subscriptions", "type": "User", "url": "https://api.github.com/users/rangehow" }
[]
closed
false
null
[]
null
3
"2023-09-26T03:02:16Z"
"2023-09-28T18:23:36Z"
"2023-09-28T18:23:36Z"
NONE
null
null
null
### Describe the bug I use the following code to download natural_question dataset. Even though I have completely download it, the next time I run this code, the new download procedure will start and cover the original /data/lxy/NQ config=datasets.DownloadConfig(resume_download=True,max_retries=100,cache_dir=r'/data/lxy/NQ',download_desc='NQ') data=datasets.load_dataset('natural_questions',cache_dir=r'/data/lxy/NQ',download_config=config,download_mode=DownloadMode.REUSE_DATASET_IF_EXISTS) --- Since I don't have apache_beam installed, it throw a exception. After I pip install apache_beam ,the download restart.. ![image](https://github.com/huggingface/datasets/assets/88258534/f28ce7fe-29ea-4348-b87f-e69182a8bd41) ### Steps to reproduce the bug run this two line code config=datasets.DownloadConfig(resume_download=True,max_retries=100,cache_dir=r'/data/lxy/NQ',download_desc='NQ') data=datasets.load_dataset('natural_questions',cache_dir=r'/data/lxy/NQ',download_config=config,download_mode=DownloadMode.REUSE_DATASET_IF_EXISTS) ### Expected behavior Download behavior can be correctly follow DownloadMode ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-3.10.0-1160.88.1.el7.x86_64-x86_64-with-glibc2.17 - Python version: 3.9.17 - Huggingface_hub version: 0.16.4 - PyArrow version: 11.0.0 - Pandas version: 2.0.3
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6260/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6260/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6259
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6259/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6259/comments
https://api.github.com/repos/huggingface/datasets/issues/6259/events
https://github.com/huggingface/datasets/issues/6259
1,911,965,758
I_kwDODunzps5x9kg-
6,259
Duplicated Rows When Loading Parquet Files from Root Directory with Subdirectories
{ "avatar_url": "https://avatars.githubusercontent.com/u/141304309?v=4", "events_url": "https://api.github.com/users/MF-FOOM/events{/privacy}", "followers_url": "https://api.github.com/users/MF-FOOM/followers", "following_url": "https://api.github.com/users/MF-FOOM/following{/other_user}", "gists_url": "https://api.github.com/users/MF-FOOM/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MF-FOOM", "id": 141304309, "login": "MF-FOOM", "node_id": "U_kgDOCGwh9Q", "organizations_url": "https://api.github.com/users/MF-FOOM/orgs", "received_events_url": "https://api.github.com/users/MF-FOOM/received_events", "repos_url": "https://api.github.com/users/MF-FOOM/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MF-FOOM/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MF-FOOM/subscriptions", "type": "User", "url": "https://api.github.com/users/MF-FOOM" }
[]
open
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" } ]
null
1
"2023-09-25T17:20:54Z"
"2023-09-26T17:54:08Z"
null
NONE
null
null
null
### Describe the bug When parquet files are saved in "train" and "val" subdirectories under a root directory, and datasets are then loaded using `load_dataset("parquet", data_dir="root_directory")`, the resulting dataset has duplicated rows for both the training and validation sets. ### Steps to reproduce the bug 1. Create a root directory, e.g., "testing123". 2. Under "testing123", create two subdirectories: "train" and "val". 3. Create and save a parquet file with 3 unique rows in the "train" subdirectory. 4. Create and save a parquet file with 4 unique rows in the "val" subdirectory. 5. Load the datasets from the root directory using `load_dataset("parquet", data_dir="testing123")` 6. Iterate through the datasets and print the rows Here's a collab reproducing these steps: https://colab.research.google.com/drive/11NEdImnQ3OqJlwKSHRMhr7jCBesNdLY4?usp=sharing ### Expected behavior - Training set should contain 3 unique rows. - Validation set should contain 4 unique rows. ### Environment info - `datasets` version: 2.14.5 - Platform: Linux-5.15.120+-x86_64-with-glibc2.35 - Python version: 3.10.12 - Huggingface_hub version: 0.17.2 - PyArrow version: 9.0.0 - Pandas version: 1.5.3
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6259/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6259/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6258
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6258/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6258/comments
https://api.github.com/repos/huggingface/datasets/issues/6258/events
https://github.com/huggingface/datasets/pull/6258
1,911,445,373
PR_kwDODunzps5bHxHl
6,258
[DOCS] Fix typo: Elasticsearch
{ "avatar_url": "https://avatars.githubusercontent.com/u/32779855?v=4", "events_url": "https://api.github.com/users/leemthompo/events{/privacy}", "followers_url": "https://api.github.com/users/leemthompo/followers", "following_url": "https://api.github.com/users/leemthompo/following{/other_user}", "gists_url": "https://api.github.com/users/leemthompo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/leemthompo", "id": 32779855, "login": "leemthompo", "node_id": "MDQ6VXNlcjMyNzc5ODU1", "organizations_url": "https://api.github.com/users/leemthompo/orgs", "received_events_url": "https://api.github.com/users/leemthompo/received_events", "repos_url": "https://api.github.com/users/leemthompo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/leemthompo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leemthompo/subscriptions", "type": "User", "url": "https://api.github.com/users/leemthompo" }
[]
closed
false
null
[]
null
2
"2023-09-25T12:50:59Z"
"2023-09-26T14:55:35Z"
"2023-09-26T13:36:40Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6258.diff", "html_url": "https://github.com/huggingface/datasets/pull/6258", "merged_at": "2023-09-26T13:36:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/6258.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6258" }
Not ElasticSearch :)
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6258/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6258/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6257
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6257/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6257/comments
https://api.github.com/repos/huggingface/datasets/issues/6257/events
https://github.com/huggingface/datasets/issues/6257
1,910,741,044
I_kwDODunzps5x45g0
6,257
HfHubHTTPError - exceeded our hourly quotas for action: commit
{ "avatar_url": "https://avatars.githubusercontent.com/u/57996478?v=4", "events_url": "https://api.github.com/users/yuvalkirstain/events{/privacy}", "followers_url": "https://api.github.com/users/yuvalkirstain/followers", "following_url": "https://api.github.com/users/yuvalkirstain/following{/other_user}", "gists_url": "https://api.github.com/users/yuvalkirstain/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yuvalkirstain", "id": 57996478, "login": "yuvalkirstain", "node_id": "MDQ6VXNlcjU3OTk2NDc4", "organizations_url": "https://api.github.com/users/yuvalkirstain/orgs", "received_events_url": "https://api.github.com/users/yuvalkirstain/received_events", "repos_url": "https://api.github.com/users/yuvalkirstain/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yuvalkirstain/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yuvalkirstain/subscriptions", "type": "User", "url": "https://api.github.com/users/yuvalkirstain" }
[]
closed
false
null
[]
null
4
"2023-09-25T06:11:43Z"
"2023-10-16T13:30:49Z"
"2023-10-16T13:30:48Z"
NONE
null
null
null
### Describe the bug I try to upload a very large dataset of images, and get the following error: ``` File /fsx-multigen/yuvalkirstain/miniconda/envs/pickapic/lib/python3.10/site-packages/huggingface_hub/hf_api.py:2712, in HfApi.create_commit(self, repo_id, operations, commit_message, commit_description, token, repo_type, revision, create_pr, num_threads, parent_commit, run_as_future) 2710 try: 2711 commit_resp = get_session().post(url=commit_url, headers=headers, data=data, params=params) -> 2712 hf_raise_for_status(commit_resp, endpoint_name="commit") 2713 except RepositoryNotFoundError as e: 2714 e.append_to_message(_CREATE_COMMIT_NO_REPO_ERROR_MESSAGE) File /fsx-multigen/yuvalkirstain/miniconda/envs/pickapic/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py:301, in hf_raise_for_status(response, endpoint_name) 297 raise BadRequestError(message, response=response) from e 299 # Convert `HTTPError` into a `HfHubHTTPError` to display request information 300 # as well (request id and/or server error message) --> 301 raise HfHubHTTPError(str(e), response=response) from e HfHubHTTPError: 429 Client Error: Too Many Requests for url: https://huggingface.co/api/datasets/yuvalkirstain/pickapic_v2/commit/main (Request ID: Root=1-65112399-12d63f7d7f28bfa40a36a0fd) You have exceeded our hourly quotas for action: commit. We invite you to retry later. ``` this makes it much less convenient to host large datasets on HF hub. ### Steps to reproduce the bug Upload a very large dataset of images ### Expected behavior the upload to work well ### Environment info - `datasets` version: 2.13.1 - Platform: Linux-5.15.0-1033-aws-x86_64-with-glibc2.31 - Python version: 3.10.11 - Huggingface_hub version: 0.15.1 - PyArrow version: 12.0.1 - Pandas version: 1.5.3
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6257/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6257/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6256
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6256/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6256/comments
https://api.github.com/repos/huggingface/datasets/issues/6256/events
https://github.com/huggingface/datasets/issues/6256
1,910,275,199
I_kwDODunzps5x3Hx_
6,256
load_dataset() function's cache_dir does not seems to work
{ "avatar_url": "https://avatars.githubusercontent.com/u/171831?v=4", "events_url": "https://api.github.com/users/andyzhu/events{/privacy}", "followers_url": "https://api.github.com/users/andyzhu/followers", "following_url": "https://api.github.com/users/andyzhu/following{/other_user}", "gists_url": "https://api.github.com/users/andyzhu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/andyzhu", "id": 171831, "login": "andyzhu", "node_id": "MDQ6VXNlcjE3MTgzMQ==", "organizations_url": "https://api.github.com/users/andyzhu/orgs", "received_events_url": "https://api.github.com/users/andyzhu/received_events", "repos_url": "https://api.github.com/users/andyzhu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/andyzhu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andyzhu/subscriptions", "type": "User", "url": "https://api.github.com/users/andyzhu" }
[]
open
false
null
[]
null
1
"2023-09-24T15:34:06Z"
"2023-09-27T13:40:45Z"
null
NONE
null
null
null
### Describe the bug datasets version: 2.14.5 when trying to run the following command trec = load_dataset('trec', split='train[:1000]', cache_dir='/path/to/my/dir') I keep getting error saying the command does not have permission to the default cache directory on my macbook pro machine. It seems the cache_dir parameter cannot change the dataset saving directory from the default what ever explained in the https://huggingface.co/docs/datasets/cache does not seem to work ### Steps to reproduce the bug datasets version: 2.14.5 when trying to run the following command trec = load_dataset('trec', split='train[:1000]', cache_dir='/path/to/my/dir') I keep getting error saying the command does not have permission to the default cache directory on my macbook pro machine. It seems the cache_dir parameter cannot change the dataset saving directory from the default what ever explained in the https://huggingface.co/docs/datasets/cache does not seem to work ### Expected behavior the dataset should be saved to the cache_dir points to ### Environment info datasets version: 2.14.5 macos X: Ventura 13.4.1 (c)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6256/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6256/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6255
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6255/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6255/comments
https://api.github.com/repos/huggingface/datasets/issues/6255/events
https://github.com/huggingface/datasets/pull/6255
1,909,842,977
PR_kwDODunzps5bCioS
6,255
Parallelize builder configs creation
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
5
"2023-09-23T11:56:20Z"
"2024-01-11T06:32:34Z"
"2023-09-26T15:44:19Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6255.diff", "html_url": "https://github.com/huggingface/datasets/pull/6255", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6255.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6255" }
For datasets with lots of configs defined in YAML E.g. `load_dataset("uonlp/CulturaX", "fr", revision="refs/pr/6")` from >1min to 15sec
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6255/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6255/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6254
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6254/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6254/comments
https://api.github.com/repos/huggingface/datasets/issues/6254/events
https://github.com/huggingface/datasets/issues/6254
1,909,672,104
I_kwDODunzps5x00io
6,254
Dataset.from_generator() cost much more time in vscode debugging mode then running mode
{ "avatar_url": "https://avatars.githubusercontent.com/u/56437469?v=4", "events_url": "https://api.github.com/users/dontnet-wuenze/events{/privacy}", "followers_url": "https://api.github.com/users/dontnet-wuenze/followers", "following_url": "https://api.github.com/users/dontnet-wuenze/following{/other_user}", "gists_url": "https://api.github.com/users/dontnet-wuenze/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dontnet-wuenze", "id": 56437469, "login": "dontnet-wuenze", "node_id": "MDQ6VXNlcjU2NDM3NDY5", "organizations_url": "https://api.github.com/users/dontnet-wuenze/orgs", "received_events_url": "https://api.github.com/users/dontnet-wuenze/received_events", "repos_url": "https://api.github.com/users/dontnet-wuenze/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dontnet-wuenze/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dontnet-wuenze/subscriptions", "type": "User", "url": "https://api.github.com/users/dontnet-wuenze" }
[]
closed
false
null
[]
null
1
"2023-09-23T02:07:26Z"
"2023-10-03T14:42:53Z"
"2023-10-03T14:42:53Z"
NONE
null
null
null
### Describe the bug Hey there, I’m using Dataset.from_generator() to convert a torch_dataset to the Huggingface Dataset. However, when I debug my code on vscode, I find that it runs really slow on Dataset.from_generator() which may even 20 times longer then run the script on terminal. ### Steps to reproduce the bug I write a simple test code : ```python import os from functools import partial from typing import Callable import torch import time from torch.utils.data import Dataset as TorchDataset from datasets import load_from_disk, Dataset as HFDataset import torch from torch.utils.data import Dataset class SimpleDataset(Dataset): def __init__(self, data): self.data = data self.keys = list(data[0].keys()) def __len__(self): return len(self.data) def __getitem__(self, index): sample = self.data[index] return {key: sample[key] for key in self.keys} def TorchDataset2HuggingfaceDataset(torch_dataset: TorchDataset, cache_dir: str = None ) -> HFDataset: """ convert torch dataset to huggingface dataset """ generator : Callable[[], TorchDataset] = lambda: (sample for sample in torch_dataset) return HFDataset.from_generator(generator, cache_dir=cache_dir) if __name__ == '__main__': data = [ {'id': 1, 'name': 'Alice'}, {'id': 2, 'name': 'Bob'}, {'id': 3, 'name': 'Charlie'} ] torch_dataset = SimpleDataset(data) start_time = time.time() huggingface_dataset = TorchDataset2HuggingfaceDataset(torch_dataset) end_time = time.time() print("time: ", end_time - start_time) print(huggingface_dataset) ``` ### Expected behavior this test on my machine report that the running time on terminal is 0.086, however the running time in debugging mode on vscode is 0.25, which I think is much longer than expected. I’d like to know is the anything wrong in the code or just because of debugging? I have traced the code and I find is this func which I get stuck. ```python def create_config_id( self, config_kwargs: dict, custom_features: Optional[Features] = None, ) -> str: ... # stuck in this line suffix = Hasher.hash(config_kwargs_to_add_to_suffix) ``` ### Environment info - `datasets` version: 2.12.0 - Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31 - Python version: 3.11.3 - Huggingface_hub version: 0.17.2 - PyArrow version: 11.0.0 - Pandas version: 2.0.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6254/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6254/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6253
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6253/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6253/comments
https://api.github.com/repos/huggingface/datasets/issues/6253/events
https://github.com/huggingface/datasets/pull/6253
1,906,618,910
PR_kwDODunzps5a3s__
6,253
Check builder cls default config name in inspect
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
4
"2023-09-21T10:15:32Z"
"2023-09-21T14:16:44Z"
"2023-09-21T14:08:00Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6253.diff", "html_url": "https://github.com/huggingface/datasets/pull/6253", "merged_at": "2023-09-21T14:08:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/6253.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6253" }
Fix https://github.com/huggingface/datasets-server/issues/1812 this was causing this issue: ```ipython In [1]: from datasets import * In [2]: inspect.get_dataset_config_names("aakanksha/udpos") Out[2]: ['default'] In [3]: load_dataset_builder("aakanksha/udpos").config.name Out[3]: 'en' ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6253/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6253/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6252
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6252/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6252/comments
https://api.github.com/repos/huggingface/datasets/issues/6252/events
https://github.com/huggingface/datasets/issues/6252
1,906,375,378
I_kwDODunzps5xoPrS
6,252
exif_transpose not done to Image (PIL problem)
{ "avatar_url": "https://avatars.githubusercontent.com/u/108274349?v=4", "events_url": "https://api.github.com/users/rhajou/events{/privacy}", "followers_url": "https://api.github.com/users/rhajou/followers", "following_url": "https://api.github.com/users/rhajou/following{/other_user}", "gists_url": "https://api.github.com/users/rhajou/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rhajou", "id": 108274349, "login": "rhajou", "node_id": "U_kgDOBnQirQ", "organizations_url": "https://api.github.com/users/rhajou/orgs", "received_events_url": "https://api.github.com/users/rhajou/received_events", "repos_url": "https://api.github.com/users/rhajou/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rhajou/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rhajou/subscriptions", "type": "User", "url": "https://api.github.com/users/rhajou" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
{ "closed_at": null, "closed_issues": 0, "created_at": "2023-02-13T16:22:42Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }, "description": "Next major release", "due_on": null, "html_url": "https://github.com/huggingface/datasets/milestone/10", "id": 9038583, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/10/labels", "node_id": "MI_kwDODunzps4Aier3", "number": 10, "open_issues": 4, "state": "open", "title": "3.0", "updated_at": "2023-09-22T14:07:52Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/10" }
2
"2023-09-21T08:11:46Z"
"2023-09-22T14:07:52Z"
null
NONE
null
null
null
### Feature request I noticed that some of my images loaded using PIL have some metadata related to exif that can rotate them when loading. Since the dataset.features.Image uses PIL for loading, the loaded image may be rotated (width and height will be inverted) thus for tasks as object detection and layoutLM this can create some inconsistencies (between input bboxes and input images). For now there is no option in datasets.features.Image to specify that. We need to do the following when preparing examples (when preparing images for training, test or inference): ``` from PIL import Image, ImageOps pil = ImageOps.exif_transpose(pil) ``` reference: https://stackoverflow.com/a/63950647/5720150 Is it possible to add this by default to the datasets.feature.Image ? or to add the option to do the ImageOps.exif_transpose? Thank you ### Motivation Prevent having inverted data related to exif metadata that may affect object detection tasks ### Your contribution Changing in datasets.featrues.Image I can help with that.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6252/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6252/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6251
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6251/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6251/comments
https://api.github.com/repos/huggingface/datasets/issues/6251/events
https://github.com/huggingface/datasets/pull/6251
1,904,418,426
PR_kwDODunzps5awQsy
6,251
Support streaming datasets with pyarrow.parquet.read_table
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
10
"2023-09-20T08:07:02Z"
"2023-09-27T06:37:03Z"
"2023-09-27T06:26:24Z"
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6251.diff", "html_url": "https://github.com/huggingface/datasets/pull/6251", "merged_at": "2023-09-27T06:26:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/6251.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6251" }
Support streaming datasets with `pyarrow.parquet.read_table`. See: https://huggingface.co/datasets/uonlp/CulturaX/discussions/2 CC: @AndreaFrancis
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6251/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6251/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6247
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6247/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6247/comments
https://api.github.com/repos/huggingface/datasets/issues/6247/events
https://github.com/huggingface/datasets/pull/6247
1,901,390,945
PR_kwDODunzps5amAQ1
6,247
Update create_dataset.mdx
{ "avatar_url": "https://avatars.githubusercontent.com/u/76403422?v=4", "events_url": "https://api.github.com/users/EswarDivi/events{/privacy}", "followers_url": "https://api.github.com/users/EswarDivi/followers", "following_url": "https://api.github.com/users/EswarDivi/following{/other_user}", "gists_url": "https://api.github.com/users/EswarDivi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/EswarDivi", "id": 76403422, "login": "EswarDivi", "node_id": "MDQ6VXNlcjc2NDAzNDIy", "organizations_url": "https://api.github.com/users/EswarDivi/orgs", "received_events_url": "https://api.github.com/users/EswarDivi/received_events", "repos_url": "https://api.github.com/users/EswarDivi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/EswarDivi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EswarDivi/subscriptions", "type": "User", "url": "https://api.github.com/users/EswarDivi" }
[]
closed
false
null
[]
null
2
"2023-09-18T17:06:29Z"
"2023-09-19T18:51:49Z"
"2023-09-19T18:40:10Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6247.diff", "html_url": "https://github.com/huggingface/datasets/pull/6247", "merged_at": "2023-09-19T18:40:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/6247.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6247" }
modified , as AudioFolder and ImageFolder not in Dataset Library. ``` from datasets import AudioFolder ``` and ```from datasets import ImageFolder``` to ```from datasets import load_dataset``` ``` cannot import name 'AudioFolder' from 'datasets' (/home/eswardivi/miniconda3/envs/Hugformers/lib/python3.10/site-packages/datasets/__init__.py) ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6247/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6247/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6246
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6246/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6246/comments
https://api.github.com/repos/huggingface/datasets/issues/6246/events
https://github.com/huggingface/datasets/issues/6246
1,899,848,414
I_kwDODunzps5xPWLe
6,246
Add new column to dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4", "events_url": "https://api.github.com/users/andysingal/events{/privacy}", "followers_url": "https://api.github.com/users/andysingal/followers", "following_url": "https://api.github.com/users/andysingal/following{/other_user}", "gists_url": "https://api.github.com/users/andysingal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/andysingal", "id": 20493493, "login": "andysingal", "node_id": "MDQ6VXNlcjIwNDkzNDkz", "organizations_url": "https://api.github.com/users/andysingal/orgs", "received_events_url": "https://api.github.com/users/andysingal/received_events", "repos_url": "https://api.github.com/users/andysingal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/andysingal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andysingal/subscriptions", "type": "User", "url": "https://api.github.com/users/andysingal" }
[]
closed
false
null
[]
null
4
"2023-09-17T16:59:48Z"
"2023-09-18T16:20:09Z"
"2023-09-18T16:20:09Z"
NONE
null
null
null
### Describe the bug ``` --------------------------------------------------------------------------- KeyError Traceback (most recent call last) [<ipython-input-9-bd197b36b6a0>](https://localhost:8080/#) in <cell line: 1>() ----> 1 dataset['train']['/workspace/data'] 3 frames [/usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py](https://localhost:8080/#) in _check_valid_column_key(key, columns) 518 def _check_valid_column_key(key: str, columns: List[str]) -> None: 519 if key not in columns: --> 520 raise KeyError(f"Column {key} not in the dataset. Current columns in the dataset: {columns}") 521 522 KeyError: "Column train not in the dataset. Current columns in the dataset: ['image', '/workspace/data']" ``` ### Steps to reproduce the bug please find the notebook for reference: https://colab.research.google.com/drive/10lZ_zLtU4itYVmIVTvIEVbjfOtCZaAZy?usp=sharing ### Expected behavior add column to the dataset ### Environment info colab pro
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6246/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6246/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6244
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6244/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6244/comments
https://api.github.com/repos/huggingface/datasets/issues/6244/events
https://github.com/huggingface/datasets/pull/6244
1,898,861,422
PR_kwDODunzps5adtD3
6,244
Add support for `fsspec>=2023.9.0`
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
19
"2023-09-15T17:58:25Z"
"2023-09-26T15:41:38Z"
"2023-09-26T15:32:51Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6244.diff", "html_url": "https://github.com/huggingface/datasets/pull/6244", "merged_at": "2023-09-26T15:32:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/6244.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6244" }
Fix #6214
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6244/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6244/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6243
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6243/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6243/comments
https://api.github.com/repos/huggingface/datasets/issues/6243/events
https://github.com/huggingface/datasets/pull/6243
1,898,532,784
PR_kwDODunzps5aclIy
6,243
Fix cast from fixed size list to variable size list
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
6
"2023-09-15T14:23:33Z"
"2023-09-19T18:02:21Z"
"2023-09-19T17:53:17Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6243.diff", "html_url": "https://github.com/huggingface/datasets/pull/6243", "merged_at": "2023-09-19T17:53:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/6243.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6243" }
Fix #6242
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6243/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6243/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6242
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6242/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6242/comments
https://api.github.com/repos/huggingface/datasets/issues/6242/events
https://github.com/huggingface/datasets/issues/6242
1,896,899,123
I_kwDODunzps5xEGIz
6,242
Data alteration when loading dataset with unspecified inner sequence length
{ "avatar_url": "https://avatars.githubusercontent.com/u/45557362?v=4", "events_url": "https://api.github.com/users/qgallouedec/events{/privacy}", "followers_url": "https://api.github.com/users/qgallouedec/followers", "following_url": "https://api.github.com/users/qgallouedec/following{/other_user}", "gists_url": "https://api.github.com/users/qgallouedec/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/qgallouedec", "id": 45557362, "login": "qgallouedec", "node_id": "MDQ6VXNlcjQ1NTU3MzYy", "organizations_url": "https://api.github.com/users/qgallouedec/orgs", "received_events_url": "https://api.github.com/users/qgallouedec/received_events", "repos_url": "https://api.github.com/users/qgallouedec/repos", "site_admin": false, "starred_url": "https://api.github.com/users/qgallouedec/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qgallouedec/subscriptions", "type": "User", "url": "https://api.github.com/users/qgallouedec" }
[]
closed
false
null
[]
null
2
"2023-09-14T16:12:45Z"
"2023-09-19T17:53:18Z"
"2023-09-19T17:53:18Z"
CONTRIBUTOR
null
null
null
### Describe the bug When a dataset saved with a specified inner sequence length is loaded without specifying that length, the original data is altered and becomes inconsistent. ### Steps to reproduce the bug ```python from datasets import Dataset, Features, Value, Sequence, load_dataset # Repository ID repo_id = "my_repo_id" # Define features with a specific length of 3 for each inner sequence specified_features = Features({"key": Sequence(Sequence(Value("float32"), length=3))}) # Create a dataset with the specified features data = [ [[1.0, 2.0, 3.0], [4.0, 5.0, 6.0]], [[7.0, 8.0, 9.0], [10.0, 11.0, 12.0]], ] dataset = Dataset.from_dict({"key": data}, features=specified_features) # Push the dataset to the hub dataset.push_to_hub(repo_id) # Define features without specifying the length unspecified_features = Features({"key": Sequence(Sequence(Value("float32")))}) # Load the dataset from the hub with this new feature definition dataset = load_dataset(f"qgallouedec/{repo_id}", split="train", features=unspecified_features) # The obtained data is altered print(dataset.to_dict()) # {'key': [[[1.0], [2.0]], [[3.0], [4.0]]]} ``` ### Expected behavior ```python print(dataset.to_dict()) # {'key': [[[1.0, 2.0, 3.0], [4.0, 5.0, 6.0]], [[7.0, 8.0, 9.0], [10.0, 11.0, 12.0]]]} ``` ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-6.2.0-32-generic-x86_64-with-glibc2.35 - Python version: 3.9.12 - Huggingface_hub version: 0.15.1 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6242/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6242/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6241
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6241/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6241/comments
https://api.github.com/repos/huggingface/datasets/issues/6241/events
https://github.com/huggingface/datasets/pull/6241
1,896,429,694
PR_kwDODunzps5aVfl-
6,241
Remove unused global variables in `audio.py`
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
4
"2023-09-14T12:06:32Z"
"2023-09-15T15:57:10Z"
"2023-09-15T15:46:07Z"
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6241.diff", "html_url": "https://github.com/huggingface/datasets/pull/6241", "merged_at": "2023-09-15T15:46:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/6241.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6241" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6241/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6241/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6240
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6240/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6240/comments
https://api.github.com/repos/huggingface/datasets/issues/6240/events
https://github.com/huggingface/datasets/issues/6240
1,895,723,888
I_kwDODunzps5w_nNw
6,240
Dataloader stuck on multiple GPUs
{ "avatar_url": "https://avatars.githubusercontent.com/u/40049003?v=4", "events_url": "https://api.github.com/users/kuri54/events{/privacy}", "followers_url": "https://api.github.com/users/kuri54/followers", "following_url": "https://api.github.com/users/kuri54/following{/other_user}", "gists_url": "https://api.github.com/users/kuri54/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kuri54", "id": 40049003, "login": "kuri54", "node_id": "MDQ6VXNlcjQwMDQ5MDAz", "organizations_url": "https://api.github.com/users/kuri54/orgs", "received_events_url": "https://api.github.com/users/kuri54/received_events", "repos_url": "https://api.github.com/users/kuri54/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kuri54/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kuri54/subscriptions", "type": "User", "url": "https://api.github.com/users/kuri54" }
[]
closed
false
null
[]
null
2
"2023-09-14T05:30:30Z"
"2023-09-14T23:54:42Z"
"2023-09-14T23:54:42Z"
NONE
null
null
null
### Describe the bug I am trying to get CLIP to fine-tuning with my code. When I tried to run it on multiple GPUs using accelerate, I encountered the following phenomenon. - Validation dataloader stuck in 2nd epoch only on multi-GPU Specifically, when the "for inputs in valid_loader:" process is finished, it does not proceed to the next step. train_loader process is completed. Also, both train and valid are working correctly in the first epoch. The accelerate command at that time is as follows. `accelerate launch --multi_gpu --num_processes=2 {script_name.py} {--arg1} {--arg2} ...` - This will not happen when single GPU is used. `CUDA_VISIBLE_DEVICES="0" accelerate launch {script_name.py} --arg1 --arg2 ...` - Setting num_workers=0 in dataloader did not change the result. ### Steps to reproduce the bug 1. The codes for fine-tuning the regular CLIP were updated for accelerate. 2. Run the code with the accelerate command as `accelerate launch --multi_gpu --num_processes=2 {script_name.py} {--arg1} {--arg2} ...` and the above problem will occur. 3. CUDA_VISIBLE_DEVICES="0" accelerate launch {script_name.py} --arg1 --arg2 ...` , it works fine. ### Expected behavior It Should end normally as if it was run on a single GPU. ### Environment info Since `datasets-cli env` did not work, the environment is described below. - OS: Ubuntu 22.04 with Docker - Docker: 24.0.5, build ced0996 - Python: 3.10.12 - torch==2.0.1 - accelerate==0.21.0 - transformers==4.33.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6240/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6240/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6239
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6239/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6239/comments
https://api.github.com/repos/huggingface/datasets/issues/6239/events
https://github.com/huggingface/datasets/issues/6239
1,895,349,382
I_kwDODunzps5w-LyG
6,239
Load local audio data doesn't work
{ "avatar_url": "https://avatars.githubusercontent.com/u/554032?v=4", "events_url": "https://api.github.com/users/abodacs/events{/privacy}", "followers_url": "https://api.github.com/users/abodacs/followers", "following_url": "https://api.github.com/users/abodacs/following{/other_user}", "gists_url": "https://api.github.com/users/abodacs/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/abodacs", "id": 554032, "login": "abodacs", "node_id": "MDQ6VXNlcjU1NDAzMg==", "organizations_url": "https://api.github.com/users/abodacs/orgs", "received_events_url": "https://api.github.com/users/abodacs/received_events", "repos_url": "https://api.github.com/users/abodacs/repos", "site_admin": false, "starred_url": "https://api.github.com/users/abodacs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/abodacs/subscriptions", "type": "User", "url": "https://api.github.com/users/abodacs" }
[]
closed
false
null
[]
null
2
"2023-09-13T22:30:01Z"
"2023-09-15T14:32:10Z"
"2023-09-15T14:32:10Z"
NONE
null
null
null
### Describe the bug I get a RuntimeError from the following code: ```python audio_dataset = Dataset.from_dict({"audio": ["/kaggle/input/bengaliai-speech/train_mp3s/000005f3362c.mp3"]}).cast_column("audio", Audio()) audio_dataset[0] ``` ### Traceback <details> ```python RuntimeError Traceback (most recent call last) Cell In[33], line 1 ----> 1 train_dataset[0] File /opt/conda/lib/python3.10/site-packages/datasets/arrow_dataset.py:1764, in Dataset.__getitem__(self, key) 1762 def __getitem__(self, key): # noqa: F811 1763 """Can be used to index columns (by string names) or rows (by integer index or iterable of indices or bools).""" -> 1764 return self._getitem( 1765 key, 1766 ) File /opt/conda/lib/python3.10/site-packages/datasets/arrow_dataset.py:1749, in Dataset._getitem(self, key, decoded, **kwargs) 1747 formatter = get_formatter(format_type, features=self.features, decoded=decoded, **format_kwargs) 1748 pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) -> 1749 formatted_output = format_table( 1750 pa_subtable, key, formatter=formatter, format_columns=format_columns, output_all_columns=output_all_columns 1751 ) 1752 return formatted_output File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:532, in format_table(table, key, formatter, format_columns, output_all_columns) 530 python_formatter = PythonFormatter(features=None) 531 if format_columns is None: --> 532 return formatter(pa_table, query_type=query_type) 533 elif query_type == "column": 534 if key in format_columns: File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:281, in Formatter.__call__(self, pa_table, query_type) 279 def __call__(self, pa_table: pa.Table, query_type: str) -> Union[RowFormat, ColumnFormat, BatchFormat]: 280 if query_type == "row": --> 281 return self.format_row(pa_table) 282 elif query_type == "column": 283 return self.format_column(pa_table) File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:312, in PythonFormatter.format_row(self, pa_table) 310 row = self.python_arrow_extractor().extract_row(pa_table) 311 if self.decoded: --> 312 row = self.python_features_decoder.decode_row(row) 313 return row File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:221, in PythonFeaturesDecoder.decode_row(self, row) 220 def decode_row(self, row: dict) -> dict: --> 221 return self.features.decode_example(row) if self.features else row File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1386, in Features.decode_example(self, example) 1376 def decode_example(self, example: dict): 1377 """Decode example with custom feature decoding. 1378 1379 Args: (...) 1383 :obj:`dict[str, Any]` 1384 """ -> 1386 return { 1387 column_name: decode_nested_example(feature, value) 1388 if self._column_requires_decoding[column_name] 1389 else value 1390 for column_name, (feature, value) in zip_dict( 1391 {key: value for key, value in self.items() if key in example}, example 1392 ) 1393 } File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1387, in <dictcomp>(.0) 1376 def decode_example(self, example: dict): 1377 """Decode example with custom feature decoding. 1378 1379 Args: (...) 1383 :obj:`dict[str, Any]` 1384 """ 1386 return { -> 1387 column_name: decode_nested_example(feature, value) 1388 if self._column_requires_decoding[column_name] 1389 else value 1390 for column_name, (feature, value) in zip_dict( 1391 {key: value for key, value in self.items() if key in example}, example 1392 ) 1393 } File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1087, in decode_nested_example(schema, obj) 1085 # Object with special decoding: 1086 elif isinstance(schema, (Audio, Image)): -> 1087 return schema.decode_example(obj) if obj is not None else None 1088 return obj File /opt/conda/lib/python3.10/site-packages/datasets/features/audio.py:103, in Audio.decode_example(self, value) 101 raise ValueError(f"An audio sample should have one of 'path' or 'bytes' but both are None in {value}.") 102 elif path is not None and path.endswith("mp3"): --> 103 array, sampling_rate = self._decode_mp3(file if file else path) 104 elif path is not None and path.endswith("opus"): 105 if file: File /opt/conda/lib/python3.10/site-packages/datasets/features/audio.py:241, in Audio._decode_mp3(self, path_or_file) 238 except RuntimeError as err: 239 raise ImportError("To support decoding 'mp3' audio files, please install 'sox'.") from err --> 241 array, sampling_rate = torchaudio.load(path_or_file, format="mp3") 242 if self.sampling_rate and self.sampling_rate != sampling_rate: 243 if not hasattr(self, "_resampler") or self._resampler.orig_freq != sampling_rate: File /opt/conda/lib/python3.10/site-packages/torchaudio/backend/sox_io_backend.py:256, in load(filepath, frame_offset, num_frames, normalize, channels_first, format) 254 if ret is not None: 255 return ret --> 256 return _fallback_load(filepath, frame_offset, num_frames, normalize, channels_first, format) File /opt/conda/lib/python3.10/site-packages/torchaudio/backend/sox_io_backend.py:30, in _fail_load(filepath, frame_offset, num_frames, normalize, channels_first, format) 22 def _fail_load( 23 filepath: str, 24 frame_offset: int = 0, (...) 28 format: Optional[str] = None, 29 ) -> Tuple[torch.Tensor, int]: ---> 30 raise RuntimeError("Failed to load audio from {}".format(filepath)) RuntimeError: Failed to load audio from /kaggle/input/bengaliai-speech/train_mp3s/000005f3362c.mp3 ``` </details> ### Steps to reproduce the bug 1. - Create a custom dataset using Local files of type mp3. 3. - Try to read the first audio item. ### Expected behavior Expected output ```python audio_dataset[0]["audio"] {'array': array([ 0. , 0.00024414, -0.00024414, ..., -0.00024414, 0. , 0. ], dtype=float32), 'path': 'path/to/audio_1', 'sampling_rate': 16000} ``` ### Environment info N/A
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6239/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6239/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6238
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6238/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6238/comments
https://api.github.com/repos/huggingface/datasets/issues/6238/events
https://github.com/huggingface/datasets/issues/6238
1,895,207,828
I_kwDODunzps5w9pOU
6,238
`dataset.filter` ALWAYS removes the first item from the dataset when using batched=True
{ "avatar_url": "https://avatars.githubusercontent.com/u/1330693?v=4", "events_url": "https://api.github.com/users/Taytay/events{/privacy}", "followers_url": "https://api.github.com/users/Taytay/followers", "following_url": "https://api.github.com/users/Taytay/following{/other_user}", "gists_url": "https://api.github.com/users/Taytay/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Taytay", "id": 1330693, "login": "Taytay", "node_id": "MDQ6VXNlcjEzMzA2OTM=", "organizations_url": "https://api.github.com/users/Taytay/orgs", "received_events_url": "https://api.github.com/users/Taytay/received_events", "repos_url": "https://api.github.com/users/Taytay/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Taytay/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Taytay/subscriptions", "type": "User", "url": "https://api.github.com/users/Taytay" }
[]
closed
false
null
[]
null
2
"2023-09-13T20:20:37Z"
"2023-09-17T07:05:07Z"
"2023-09-17T07:05:07Z"
NONE
null
null
null
### Describe the bug If you call batched=True when calling `filter`, the first item is _always_ filtered out, regardless of the filter condition. ### Steps to reproduce the bug Here's a minimal example: ```python def filter_batch_always_true(batch, indices): print("First index being passed into this filter function: ", indices[0]) return indices # Keep all indices data = {"value": list(range(10))} dataset = Dataset.from_dict(data) filtered_dataset = dataset.filter(filter_batch_always_true, with_indices=True, batched=True) print("Length of original dataset: ", len(dataset)) print("Length of filtered_dataset: ", len(filtered_dataset)) print("Is equal to original? ", len(filtered_dataset) == len(dataset)) print("First item of filtered dataset: ", filtered_dataset[0]) print("Last item of filtered dataset: ", filtered_dataset[-1]) ``` prints: ``` First index being passed into this filter function: 0 Length of original dataset: 10 Length of filtered_dataset: 9 Is equal to original? False First item of filtered dataset: {'value': 1} Last item of filtered dataset: {'value': 9} ``` ### Expected behavior Filter should respect the filter condition. ### Environment info - `datasets` version: 2.14.4 - Platform: macOS-13.5-arm64-arm-64bit - Python version: 3.9.18 - Huggingface_hub version: 0.17.1 - PyArrow version: 10.0.1 - Pandas version: 2.0.2
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6238/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6238/timeline
null
completed
false