url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1B
node_id
stringlengths
18
32
number
int64
1
2.96k
title
stringlengths
1
268
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
sequence
created_at
int64
1,587B
1,632B
updated_at
int64
1,587B
1,632B
closed_at
int64
1,587B
1,632B
author_association
stringclasses
4 values
active_lock_reason
null
pull_request
dict
body
stringlengths
0
228k
timeline_url
stringlengths
67
70
performed_via_github_app
null
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/724
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/724/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/724/comments
https://api.github.com/repos/huggingface/datasets/issues/724/events
https://github.com/huggingface/datasets/issues/724
718,947,700
MDU6SXNzdWU3MTg5NDc3MDA=
724
need to redirect /nlp to /datasets and remove outdated info
{ "login": "stas00", "id": 10676103, "node_id": "MDQ6VXNlcjEwNjc2MTAz", "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stas00", "html_url": "https://github.com/stas00", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "organizations_url": "https://api.github.com/users/stas00/orgs", "repos_url": "https://api.github.com/users/stas00/repos", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "received_events_url": "https://api.github.com/users/stas00/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Should be fixed now: \r\n\r\n![image](https://user-images.githubusercontent.com/35882/95917301-040b0600-0d78-11eb-9655-c4ac0e788089.png)\r\n\r\nNot sure I understand what you mean by the second part?\r\n", "Thank you!\r\n\r\n> Not sure I understand what you mean by the second part?\r\n\r\nCompare the 2:\r\n* https://huggingface.co/datasets/wikihow\r\n* https://huggingface.co/nlp/viewer/?dataset=wikihow&config=all\r\nCan you see the difference? 2nd has formatting, 1st doesn't.\r\n", "For context, those are two different pages (not an old vs new one), one is from the dataset viewer (you can browse data inside the datasets) while the other is just a basic reference page displayed some metadata about the dataset.\r\n\r\nFor the second one, we'll move to markdown parsing soon, so it'll be formatted better.", "I understand. I was just flagging the lack of markup issue." ]
1,602,457,932,000
1,602,694,812,000
1,602,694,812,000
CONTRIBUTOR
null
null
It looks like the website still has all the `nlp` data, e.g.: https://huggingface.co/nlp/viewer/?dataset=wikihow&config=all should probably redirect to: https://huggingface.co/datasets/wikihow also for some reason the new information is slightly borked. If you look at the old one it was nicely formatted and had the links marked up, the new one is just a jumble of text in one chunk and no markup for links (i.e. not clickable).
https://api.github.com/repos/huggingface/datasets/issues/724/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/723
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/723/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/723/comments
https://api.github.com/repos/huggingface/datasets/issues/723/events
https://github.com/huggingface/datasets/issues/723
718,926,723
MDU6SXNzdWU3MTg5MjY3MjM=
723
Adding pseudo-labels to datasets
{ "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/sshleifer/followers", "following_url": "https://api.github.com/users/sshleifer/following{/other_user}", "gists_url": "https://api.github.com/users/sshleifer/gists{/gist_id}", "starred_url": "https://api.github.com/users/sshleifer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sshleifer/subscriptions", "organizations_url": "https://api.github.com/users/sshleifer/orgs", "repos_url": "https://api.github.com/users/sshleifer/repos", "events_url": "https://api.github.com/users/sshleifer/events{/privacy}", "received_events_url": "https://api.github.com/users/sshleifer/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/sshleifer/followers", "following_url": "https://api.github.com/users/sshleifer/following{/other_user}", "gists_url": "https://api.github.com/users/sshleifer/gists{/gist_id}", "starred_url": "https://api.github.com/users/sshleifer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sshleifer/subscriptions", "organizations_url": "https://api.github.com/users/sshleifer/orgs", "repos_url": "https://api.github.com/users/sshleifer/repos", "events_url": "https://api.github.com/users/sshleifer/events{/privacy}", "received_events_url": "https://api.github.com/users/sshleifer/received_events", "type": "User", "site_admin": false }
[ { "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/sshleifer/followers", "following_url": "https://api.github.com/users/sshleifer/following{/other_user}", "gists_url": "https://api.github.com/users/sshleifer/gists{/gist_id}", "starred_url": "https://api.github.com/users/sshleifer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sshleifer/subscriptions", "organizations_url": "https://api.github.com/users/sshleifer/orgs", "repos_url": "https://api.github.com/users/sshleifer/repos", "events_url": "https://api.github.com/users/sshleifer/events{/privacy}", "received_events_url": "https://api.github.com/users/sshleifer/received_events", "type": "User", "site_admin": false } ]
null
[ "Nice ! :)\r\nIt's indeed the first time we have such contributions so we'll have to figure out the appropriate way to integrate them.\r\nCould you add details on what they could be used for ?\r\n", "They can be used as training data for a smaller model.", "Sounds just like a regular dataset to me then, no?", "A new configuration for those datasets should do the job then.\r\nNote that until now datasets like xsum only had one configuration. It means that users didn't have to specify the configuration name when loading the dataset. If we add new configs, users that update the lib will have to update their code to specify the default/standard configuration name (not the one with pseudo labels).", "Could also be a `user-namespace` dataset maybe?", "Oh yes why not. I'm more in favor of this actually since pseudo labels are things that users (not dataset authors in general) can compute by themselves and share with the community", "![image](https://user-images.githubusercontent.com/6045025/96045248-b528a380-0e3f-11eb-9124-bd55afa031bb.png)\r\n\r\nI assume I should (for example) rename the xsum dir, change the URL, and put the modified dir somewhere in S3?", "You can use the `datasets-cli` to upload the folder with your version of xsum with the pseudo labels.\r\n\r\n```\r\ndatasets-cli upload_dataset path/to/xsum\r\n```" ]
1,602,450,345,000
1,627,967,511,000
1,627,967,511,000
MEMBER
null
null
I recently [uploaded pseudo-labels](https://github.com/huggingface/transformers/blob/master/examples/seq2seq/precomputed_pseudo_labels.md) for CNN/DM, XSUM and WMT16-en-ro to s3, and thom mentioned I should add them to this repo. Since pseudo-labels are just a large model's generations on an existing dataset, what is the right way to structure this contribution. I read https://huggingface.co/docs/datasets/add_dataset.html, but it doesn't really cover this type of contribution. I could, for example, make a new directory, `xsum_bart_pseudolabels` for each set of pseudolabels or add some sort of parametrization to `xsum.py`: https://github.com/huggingface/datasets/blob/5f4c6e830f603830117877b8990a0e65a2386aa6/datasets/xsum/xsum.py What do you think @lhoestq ?
https://api.github.com/repos/huggingface/datasets/issues/723/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/722
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/722/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/722/comments
https://api.github.com/repos/huggingface/datasets/issues/722/events
https://github.com/huggingface/datasets/pull/722
718,689,117
MDExOlB1bGxSZXF1ZXN0NTAxMDI3NjAw
722
datasets(RWTH-PHOENIX-Weather 2014 T): add initial loading script
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/followers", "following_url": "https://api.github.com/users/AmitMY/following{/other_user}", "gists_url": "https://api.github.com/users/AmitMY/gists{/gist_id}", "starred_url": "https://api.github.com/users/AmitMY/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AmitMY/subscriptions", "organizations_url": "https://api.github.com/users/AmitMY/orgs", "repos_url": "https://api.github.com/users/AmitMY/repos", "events_url": "https://api.github.com/users/AmitMY/events{/privacy}", "received_events_url": "https://api.github.com/users/AmitMY/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "This might be interesting to @kayoyin the author of https://github.com/kayoyin/transformer-slt – pinging you just in case :)", "Thanks Amit, this is a great idea! I'm thinking of porting the SLT models from my paper here as well, having this dataset would be perfect for that :)", "Thanks for your contribution, @AmitMY. Are you still interested in adding this dataset?\r\n\r\nWe are removing the dataset scripts from this GitHub repo and moving them to the Hugging Face Hub: https://huggingface.co/datasets\r\n\r\nWe would suggest you create this dataset there. Please, feel free to tell us if you need some help." ]
1,602,359,048,000
1,609,830,411,000
null
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/722", "html_url": "https://github.com/huggingface/datasets/pull/722", "diff_url": "https://github.com/huggingface/datasets/pull/722.diff", "patch_url": "https://github.com/huggingface/datasets/pull/722.patch" }
This is the first sign language dataset in this repo as far as I know. Following an old issue I opened https://github.com/huggingface/datasets/issues/302. I added the dataset official REAMDE file, but I see it's not very standard, so it can be removed.
https://api.github.com/repos/huggingface/datasets/issues/722/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/721
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/721/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/721/comments
https://api.github.com/repos/huggingface/datasets/issues/721/events
https://github.com/huggingface/datasets/issues/721
718,647,147
MDU6SXNzdWU3MTg2NDcxNDc=
721
feat(dl_manager): add support for ftp downloads
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/followers", "following_url": "https://api.github.com/users/AmitMY/following{/other_user}", "gists_url": "https://api.github.com/users/AmitMY/gists{/gist_id}", "starred_url": "https://api.github.com/users/AmitMY/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AmitMY/subscriptions", "organizations_url": "https://api.github.com/users/AmitMY/orgs", "repos_url": "https://api.github.com/users/AmitMY/repos", "events_url": "https://api.github.com/users/AmitMY/events{/privacy}", "received_events_url": "https://api.github.com/users/AmitMY/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "We only support http by default for downloading.\r\nIf you really need to use ftp, then feel free to use a library that allows to download through ftp in your dataset script (I see that you've started working on #722 , that's awesome !). The users will get a message to install the extra library when they load the dataset.\r\n\r\nTo make the download_manager work with a custom downloader, you can call `download_manager.download_custom` instead of `download_manager.download_and_extract`. The expected arguments are the following:\r\n```\r\nurl_or_urls: url or `list`/`dict` of urls to download and extract. Each\r\n url is a `str`.\r\ncustom_download: Callable with signature (src_url: str, dst_path: str) -> Any\r\n as for example `tf.io.gfile.copy`, that lets you download from google storage\r\n```\r\n", "Also maybe it coud be interesting to have a direct support of ftp inside the `datasets` library. Do you know any good libraries that we might consider adding as a (optional ?) dependency ?", "Downloading an `ftp` file is as simple as:\r\n```python\r\nimport urllib \r\nurllib.urlretrieve('ftp://server/path/to/file', 'file')\r\n```\r\n\r\nI believe this should be supported by the library, as its not using any dependency and is trivial amount of code.", "I know its unorthodox, but I added `ftp` download support to `file_utils` in the same PR https://github.com/huggingface/datasets/pull/722\r\nSo its possible to understand the interaction of the download component with the ftp download ability", "Awesome ! I'll take a look :)", "@AmitMY Can you now download the Phoenix2014 Dataset?", "@hoanganhpham1006 yes.\r\nSee pull request https://github.com/huggingface/datasets/pull/722 , it has a loader for this dataset, mostly ready.\r\nThere's one issue that delays it being merged - https://github.com/huggingface/datasets/issues/741 - regarding memory consumption.", "The problem which I have now is that this dataset seems does not allow to download? Can you share it with me pls", "The dataset loader is not yet ready, because of that issue.\r\nIf you want to just download the dataset the old-fashioned way, just go to: https://www-i6.informatik.rwth-aachen.de/ftp/pub/rwth-phoenix/2016/phoenix-2014-T.v3.tar.gz (the ftp link is now broken, and its available over https)", "Got it, thank you so much!", "FTP downloads are supported." ]
1,602,345,020,000
1,603,531,473,000
null
CONTRIBUTOR
null
null
I am working on a new dataset (#302) and encounter a problem downloading it. ```python # This is the official download link from https://www-i6.informatik.rwth-aachen.de/~koller/RWTH-PHOENIX-2014-T/ _URL = "ftp://wasserstoff.informatik.rwth-aachen.de/pub/rwth-phoenix/2016/phoenix-2014-T.v3.tar.gz" dl_manager.download_and_extract(_URL) ``` I get an error: > ValueError: unable to parse ftp://wasserstoff.informatik.rwth-aachen.de/pub/rwth-phoenix/2016/phoenix-2014-T.v3.tar.gz as a URL or as a local path I checked, and indeed you don't consider `ftp` as a remote file. https://github.com/huggingface/datasets/blob/4c2af707a6955cf4b45f83ac67990395327c5725/src/datasets/utils/file_utils.py#L188 Adding `ftp` to that list does not immediately solve the issue, so there probably needs to be some extra work.
https://api.github.com/repos/huggingface/datasets/issues/721/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/720
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/720/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/720/comments
https://api.github.com/repos/huggingface/datasets/issues/720/events
https://github.com/huggingface/datasets/issues/720
716,581,266
MDU6SXNzdWU3MTY1ODEyNjY=
720
OSError: Cannot find data file when not using the dummy dataset in RAG
{ "login": "josemlopez", "id": 4112135, "node_id": "MDQ6VXNlcjQxMTIxMzU=", "avatar_url": "https://avatars.githubusercontent.com/u/4112135?v=4", "gravatar_id": "", "url": "https://api.github.com/users/josemlopez", "html_url": "https://github.com/josemlopez", "followers_url": "https://api.github.com/users/josemlopez/followers", "following_url": "https://api.github.com/users/josemlopez/following{/other_user}", "gists_url": "https://api.github.com/users/josemlopez/gists{/gist_id}", "starred_url": "https://api.github.com/users/josemlopez/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/josemlopez/subscriptions", "organizations_url": "https://api.github.com/users/josemlopez/orgs", "repos_url": "https://api.github.com/users/josemlopez/repos", "events_url": "https://api.github.com/users/josemlopez/events{/privacy}", "received_events_url": "https://api.github.com/users/josemlopez/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false } ]
null
[ "Same issue here. I will be digging further, but it looks like the [script](https://github.com/huggingface/datasets/blob/master/datasets/wiki_dpr/wiki_dpr.py#L132) is attempting to open a file that is not downloaded yet. \r\n\r\n```\r\n99dcbca09109e58502e6b9271d4d3f3791b43f61f3161a76b25d2775ab1a4498.lock\r\n```\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nUnpicklingError Traceback (most recent call last)\r\n~/anaconda3/envs/eqa/lib/python3.7/site-packages/numpy/lib/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 446 try:\r\n--> 447 return pickle.load(fid, **pickle_kwargs)\r\n 448 except Exception:\r\n\r\nUnpicklingError: pickle data was truncated\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n~/src/datasets/src/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 559 \r\n--> 560 if verify_infos:\r\n 561 verify_splits(self.info.splits, split_dict)\r\n\r\n~/src/datasets/src/datasets/builder.py in _prepare_split(self, split_generator)\r\n 847 writer.write(example)\r\n--> 848 finally:\r\n 849 num_examples, num_bytes = writer.finalize()\r\n\r\n~/anaconda3/envs/eqa/lib/python3.7/site-packages/tqdm/notebook.py in __iter__(self, *args, **kwargs)\r\n 227 try:\r\n--> 228 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):\r\n 229 # return super(tqdm...) will not catch exception\r\n\r\n~/anaconda3/envs/eqa/lib/python3.7/site-packages/tqdm/std.py in __iter__(self)\r\n 1132 try:\r\n-> 1133 for obj in iterable:\r\n 1134 yield obj\r\n\r\n/hdd/rag/cache/huggingface/modules/datasets_modules/datasets/wiki_dpr/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2/wiki_dpr.py in _generate_examples(self, data_file, vectors_files)\r\n 131 break\r\n--> 132 vecs = np.load(open(vectors_files.pop(0), \"rb\"), allow_pickle=True)\r\n 133 vec_idx = 0\r\n\r\n~/anaconda3/envs/eqa/lib/python3.7/site-packages/numpy/lib/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding)\r\n 449 raise IOError(\r\n--> 450 \"Failed to interpret file %s as a pickle\" % repr(file))\r\n 451 \r\n\r\nOSError: Failed to interpret file <_io.BufferedReader name='/hdd/rag/downloads/99dcbca09109e58502e6b9271d4d3f3791b43f61f3161a76b25d2775ab1a4498'> as a pickle\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nOSError Traceback (most recent call last)\r\n<ipython-input-8-24351ff8ce44> in <module>\r\n 4 retriever = RagRetriever.from_pretrained(\"facebook/rag-sequence-nq\", \r\n 5 index_name=\"exact\",\r\n----> 6 use_dummy_dataset=False)\r\n\r\n~/src/transformers/src/transformers/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs)\r\n 321 generator_tokenizer = rag_tokenizer.generator\r\n 322 return cls(\r\n--> 323 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer\r\n 324 )\r\n 325 \r\n\r\n~/src/transformers/src/transformers/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer)\r\n 310 self.config = config\r\n 311 if self._init_retrieval:\r\n--> 312 self.init_retrieval()\r\n 313 \r\n 314 @classmethod\r\n\r\n~/src/transformers/src/transformers/retrieval_rag.py in init_retrieval(self)\r\n 338 \r\n 339 logger.info(\"initializing retrieval\")\r\n--> 340 self.index.init_index()\r\n 341 \r\n 342 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None):\r\n\r\n~/src/transformers/src/transformers/retrieval_rag.py in init_index(self)\r\n 248 split=self.dataset_split,\r\n 249 index_name=self.index_name,\r\n--> 250 dummy=self.use_dummy_dataset,\r\n 251 )\r\n 252 self.dataset.set_format(\"numpy\", columns=[\"embeddings\"], output_all_columns=True)\r\n\r\n~/src/datasets/src/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)\r\n 615 builder_instance.download_and_prepare(\r\n 616 download_config=download_config,\r\n--> 617 download_mode=download_mode,\r\n 618 ignore_verifications=ignore_verifications,\r\n 619 )\r\n\r\n~/src/datasets/src/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)\r\n 481 # Sync info\r\n 482 self.info.dataset_size = sum(split.num_bytes for split in self.info.splits.values())\r\n--> 483 self.info.download_checksums = dl_manager.get_recorded_sizes_checksums()\r\n 484 self.info.size_in_bytes = self.info.dataset_size + self.info.download_size\r\n 485 # Save info\r\n\r\n~/src/datasets/src/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)\r\n 560 if verify_infos:\r\n 561 verify_splits(self.info.splits, split_dict)\r\n--> 562 \r\n 563 # Update the info object with the splits.\r\n 564 self.info.splits = split_dict\r\n\r\nOSError: Cannot find data file.\r\n```\r\n\r\nThank you.", "An update on my end. This seems like a transient issue. Reran the script from scratch overnight with no errors. ", "Closing this one. Feel free to re-open if you have other questions about this issue" ]
1,602,080,833,000
1,608,732,271,000
1,608,732,271,000
NONE
null
null
## Environment info transformers version: 3.3.1 Platform: Linux-4.19 Python version: 3.7.7 PyTorch version (GPU?): 1.6.0 Tensorflow version (GPU?): No Using GPU in script?: Yes Using distributed or parallel set-up in script?: No ## To reproduce Steps to reproduce the behaviour: ``` import os os.environ['HF_DATASETS_CACHE'] = '/workspace/notebooks/POCs/cache' from transformers import RagTokenizer, RagRetriever, RagTokenForGeneration tokenizer = RagTokenizer.from_pretrained("facebook/rag-token-nq") retriever = RagRetriever.from_pretrained("facebook/rag-token-nq", index_name="exact", use_dummy_dataset=False) ``` Plese note that I'm using the whole dataset: **use_dummy_dataset=False** After around 4 hours (downloading and some other things) this is returned: ``` Downloading and preparing dataset wiki_dpr/psgs_w100.nq.exact (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /workspace/notebooks/POCs/cache/wiki_dpr/psgs_w100.nq.exact/0.0.0/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2... --------------------------------------------------------------------------- UnpicklingError Traceback (most recent call last) /opt/conda/lib/python3.7/site-packages/numpy/lib/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding) 459 try: --> 460 return pickle.load(fid, **pickle_kwargs) 461 except Exception: UnpicklingError: pickle data was truncated During handling of the above exception, another exception occurred: OSError Traceback (most recent call last) /opt/conda/lib/python3.7/site-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 552 # Prepare split will record examples associated to the split --> 553 self._prepare_split(split_generator, **prepare_split_kwargs) 554 except OSError: /opt/conda/lib/python3.7/site-packages/datasets/builder.py in _prepare_split(self, split_generator) 840 for key, record in utils.tqdm( --> 841 generator, unit=" examples", total=split_info.num_examples, leave=False, disable=not_verbose 842 ): /opt/conda/lib/python3.7/site-packages/tqdm/notebook.py in __iter__(self, *args, **kwargs) 217 try: --> 218 for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs): 219 # return super(tqdm...) will not catch exception /opt/conda/lib/python3.7/site-packages/tqdm/std.py in __iter__(self) 1128 try: -> 1129 for obj in iterable: 1130 yield obj ~/.cache/huggingface/modules/datasets_modules/datasets/wiki_dpr/14b973bf2a456087ff69c0fd34526684eed22e48e0dfce4338f9a22b965ce7c2/wiki_dpr.py in _generate_examples(self, data_file, vectors_files) 131 break --> 132 vecs = np.load(open(vectors_files.pop(0), "rb"), allow_pickle=True) 133 vec_idx = 0 /opt/conda/lib/python3.7/site-packages/numpy/lib/npyio.py in load(file, mmap_mode, allow_pickle, fix_imports, encoding) 462 raise IOError( --> 463 "Failed to interpret file %s as a pickle" % repr(file)) 464 finally: OSError: Failed to interpret file <_io.BufferedReader name='/workspace/notebooks/POCs/cache/downloads/f34d5f091294259b4ca90e813631e69a6ded660d71b6cbedf89ddba50df94448'> as a pickle During handling of the above exception, another exception occurred: OSError Traceback (most recent call last) <ipython-input-10-f28df370ac47> in <module> 1 # ln -s /workspace/notebooks/POCs/cache /root/.cache/huggingface/datasets ----> 2 retriever = RagRetriever.from_pretrained("facebook/rag-token-nq", index_name="exact", use_dummy_dataset=False) /opt/conda/lib/python3.7/site-packages/transformers/retrieval_rag.py in from_pretrained(cls, retriever_name_or_path, **kwargs) 307 generator_tokenizer = rag_tokenizer.generator 308 return cls( --> 309 config, question_encoder_tokenizer=question_encoder_tokenizer, generator_tokenizer=generator_tokenizer 310 ) 311 /opt/conda/lib/python3.7/site-packages/transformers/retrieval_rag.py in __init__(self, config, question_encoder_tokenizer, generator_tokenizer) 298 self.config = config 299 if self._init_retrieval: --> 300 self.init_retrieval() 301 302 @classmethod /opt/conda/lib/python3.7/site-packages/transformers/retrieval_rag.py in init_retrieval(self) 324 325 logger.info("initializing retrieval") --> 326 self.index.init_index() 327 328 def postprocess_docs(self, docs, input_strings, prefix, n_docs, return_tensors=None): /opt/conda/lib/python3.7/site-packages/transformers/retrieval_rag.py in init_index(self) 238 split=self.dataset_split, 239 index_name=self.index_name, --> 240 dummy=self.use_dummy_dataset, 241 ) 242 self.dataset.set_format("numpy", columns=["embeddings"], output_all_columns=True) /opt/conda/lib/python3.7/site-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs) 609 download_config=download_config, 610 download_mode=download_mode, --> 611 ignore_verifications=ignore_verifications, 612 ) 613 /opt/conda/lib/python3.7/site-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs) 474 if not downloaded_from_gcs: 475 self._download_and_prepare( --> 476 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 477 ) 478 # Sync info /opt/conda/lib/python3.7/site-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 553 self._prepare_split(split_generator, **prepare_split_kwargs) 554 except OSError: --> 555 raise OSError("Cannot find data file. " + (self.manual_download_instructions or "")) 556 557 if verify_infos: OSError: Cannot find data file. ``` Thanks
https://api.github.com/repos/huggingface/datasets/issues/720/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/719
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/719/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/719/comments
https://api.github.com/repos/huggingface/datasets/issues/719/events
https://github.com/huggingface/datasets/pull/719
716,492,263
MDExOlB1bGxSZXF1ZXN0NDk5MjE5Mjg2
719
Fix train_test_split output format
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,602,074,341,000
1,602,077,888,000
1,602,077,886,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/719", "html_url": "https://github.com/huggingface/datasets/pull/719", "diff_url": "https://github.com/huggingface/datasets/pull/719.diff", "patch_url": "https://github.com/huggingface/datasets/pull/719.patch" }
There was an issue in the `transmit_format` wrapper that returned bad formats when using train_test_split. This was due to `column_names` being handled as a List[str] instead of Dict[str, List[str]] when the dataset transform (train_test_split) returns a DatasetDict (one set of column names per split). This should fix @timothyjlaurent 's issue in #620 and fix #676 I added tests for `transmit_format` so that it doesn't happen again
https://api.github.com/repos/huggingface/datasets/issues/719/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/718
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/718/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/718/comments
https://api.github.com/repos/huggingface/datasets/issues/718/events
https://github.com/huggingface/datasets/pull/718
715,694,709
MDExOlB1bGxSZXF1ZXN0NDk4NTU5MDcw
718
Don't use tqdm 4.50.0
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,991,953,000
1,601,992,164,000
1,601,992,162,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/718", "html_url": "https://github.com/huggingface/datasets/pull/718", "diff_url": "https://github.com/huggingface/datasets/pull/718.diff", "patch_url": "https://github.com/huggingface/datasets/pull/718.patch" }
tqdm 4.50.0 introduced permission errors on windows see [here](https://app.circleci.com/pipelines/github/huggingface/datasets/235/workflows/cfb6a39f-68eb-4802-8b17-2cd5e8ea7369/jobs/1111) for the error details. For now I just added `<4.50.0` in the setup.py Hopefully we can find what's wrong with this version soon
https://api.github.com/repos/huggingface/datasets/issues/718/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/717
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/717/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/717/comments
https://api.github.com/repos/huggingface/datasets/issues/717/events
https://github.com/huggingface/datasets/pull/717
714,959,268
MDExOlB1bGxSZXF1ZXN0NDk3OTUwOTA2
717
Fixes #712 Error in the Overview.ipynb notebook
{ "login": "subhrm", "id": 850012, "node_id": "MDQ6VXNlcjg1MDAxMg==", "avatar_url": "https://avatars.githubusercontent.com/u/850012?v=4", "gravatar_id": "", "url": "https://api.github.com/users/subhrm", "html_url": "https://github.com/subhrm", "followers_url": "https://api.github.com/users/subhrm/followers", "following_url": "https://api.github.com/users/subhrm/following{/other_user}", "gists_url": "https://api.github.com/users/subhrm/gists{/gist_id}", "starred_url": "https://api.github.com/users/subhrm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/subhrm/subscriptions", "organizations_url": "https://api.github.com/users/subhrm/orgs", "repos_url": "https://api.github.com/users/subhrm/repos", "events_url": "https://api.github.com/users/subhrm/events{/privacy}", "received_events_url": "https://api.github.com/users/subhrm/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,913,041,000
1,601,965,903,000
1,601,915,141,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/717", "html_url": "https://github.com/huggingface/datasets/pull/717", "diff_url": "https://github.com/huggingface/datasets/pull/717.diff", "patch_url": "https://github.com/huggingface/datasets/pull/717.patch" }
Fixes #712 Error in the Overview.ipynb notebook by adding `with_details=True` parameter to `list_datasets` function in Cell 3 of **overview** notebook
https://api.github.com/repos/huggingface/datasets/issues/717/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/716
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/716/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/716/comments
https://api.github.com/repos/huggingface/datasets/issues/716/events
https://github.com/huggingface/datasets/pull/716
714,952,888
MDExOlB1bGxSZXF1ZXN0NDk3OTQ1ODAw
716
Fixes #712 Attribute error in cell 3 of the overview notebook
{ "login": "subhrm", "id": 850012, "node_id": "MDQ6VXNlcjg1MDAxMg==", "avatar_url": "https://avatars.githubusercontent.com/u/850012?v=4", "gravatar_id": "", "url": "https://api.github.com/users/subhrm", "html_url": "https://github.com/subhrm", "followers_url": "https://api.github.com/users/subhrm/followers", "following_url": "https://api.github.com/users/subhrm/following{/other_user}", "gists_url": "https://api.github.com/users/subhrm/gists{/gist_id}", "starred_url": "https://api.github.com/users/subhrm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/subhrm/subscriptions", "organizations_url": "https://api.github.com/users/subhrm/orgs", "repos_url": "https://api.github.com/users/subhrm/repos", "events_url": "https://api.github.com/users/subhrm/events{/privacy}", "received_events_url": "https://api.github.com/users/subhrm/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Referencing the wrong issue # in the commit message. Closing this to fix it again." ]
1,601,912,529,000
1,601,912,798,000
1,601,912,792,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/716", "html_url": "https://github.com/huggingface/datasets/pull/716", "diff_url": "https://github.com/huggingface/datasets/pull/716.diff", "patch_url": "https://github.com/huggingface/datasets/pull/716.patch" }
Fixes the Attribute error in cell 3 of the overview notebook
https://api.github.com/repos/huggingface/datasets/issues/716/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/715
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/715/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/715/comments
https://api.github.com/repos/huggingface/datasets/issues/715/events
https://github.com/huggingface/datasets/pull/715
714,690,192
MDExOlB1bGxSZXF1ZXN0NDk3NzMwMDQ2
715
Use python read for text dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "One thing though, could we try to read the files in parallel?", "We could but I'm not sure this would help a lot since the bottleneck is the drive IO if the files are big enough.\r\nIt could make sense for very small files.", "Looks like windows is not a big fan of this approach\r\nI'm working on a fix", "I remember issue https://github.com/huggingface/datasets/issues/546 where this was kinda requested (but maybe IO would bottleneck). What do you think?", "I think it's worth testing multiprocessing. It could also be something we add to our speed benchmarks", "> I remember issue #546 where this was kinda requested (but maybe IO would bottleneck). What do you think?\r\n\r\nIt still would be interesting I think, especially in scenarios where IO is less of an issue (SSDs particularly) and where there are many smaller files. Wrapping this function in a `pool.map` is perhaps an easy thing to try. ", "Merging this one for now for the patch release" ]
1,601,891,275,000
1,601,903,598,000
1,601,903,597,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/715", "html_url": "https://github.com/huggingface/datasets/pull/715", "diff_url": "https://github.com/huggingface/datasets/pull/715.diff", "patch_url": "https://github.com/huggingface/datasets/pull/715.patch" }
As mentioned in #622 the pandas reader used for text dataset doesn't work properly when there are \r characters in the text file. Instead I switched to pure python using `open` and `read`. From my benchmark on a 100MB text file, it's the same speed as the previous pandas reader.
https://api.github.com/repos/huggingface/datasets/issues/715/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/714
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/714/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/714/comments
https://api.github.com/repos/huggingface/datasets/issues/714/events
https://github.com/huggingface/datasets/pull/714
714,487,881
MDExOlB1bGxSZXF1ZXN0NDk3NTYzNjAx
714
Add the official dependabot implementation
{ "login": "ALazyMeme", "id": 12804673, "node_id": "MDQ6VXNlcjEyODA0Njcz", "avatar_url": "https://avatars.githubusercontent.com/u/12804673?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ALazyMeme", "html_url": "https://github.com/ALazyMeme", "followers_url": "https://api.github.com/users/ALazyMeme/followers", "following_url": "https://api.github.com/users/ALazyMeme/following{/other_user}", "gists_url": "https://api.github.com/users/ALazyMeme/gists{/gist_id}", "starred_url": "https://api.github.com/users/ALazyMeme/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ALazyMeme/subscriptions", "organizations_url": "https://api.github.com/users/ALazyMeme/orgs", "repos_url": "https://api.github.com/users/ALazyMeme/repos", "events_url": "https://api.github.com/users/ALazyMeme/events{/privacy}", "received_events_url": "https://api.github.com/users/ALazyMeme/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,869,785,000
1,602,503,361,000
1,602,503,361,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/714", "html_url": "https://github.com/huggingface/datasets/pull/714", "diff_url": "https://github.com/huggingface/datasets/pull/714.diff", "patch_url": "https://github.com/huggingface/datasets/pull/714.patch" }
This will keep dependencies up to date. This will require a pr label `dependencies` being created in order to function correctly.
https://api.github.com/repos/huggingface/datasets/issues/714/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/713
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/713/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/713/comments
https://api.github.com/repos/huggingface/datasets/issues/713/events
https://github.com/huggingface/datasets/pull/713
714,475,732
MDExOlB1bGxSZXF1ZXN0NDk3NTUzOTUy
713
Fix reading text files with carriage return symbols
{ "login": "mozharovsky", "id": 6762769, "node_id": "MDQ6VXNlcjY3NjI3Njk=", "avatar_url": "https://avatars.githubusercontent.com/u/6762769?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mozharovsky", "html_url": "https://github.com/mozharovsky", "followers_url": "https://api.github.com/users/mozharovsky/followers", "following_url": "https://api.github.com/users/mozharovsky/following{/other_user}", "gists_url": "https://api.github.com/users/mozharovsky/gists{/gist_id}", "starred_url": "https://api.github.com/users/mozharovsky/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mozharovsky/subscriptions", "organizations_url": "https://api.github.com/users/mozharovsky/orgs", "repos_url": "https://api.github.com/users/mozharovsky/repos", "events_url": "https://api.github.com/users/mozharovsky/events{/privacy}", "received_events_url": "https://api.github.com/users/mozharovsky/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Discussed in #622, fixed in #715. Closing the issue. Thanks @lhoestq, it works now! 👍 " ]
1,601,867,223,000
1,602,223,105,000
1,601,905,769,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/713", "html_url": "https://github.com/huggingface/datasets/pull/713", "diff_url": "https://github.com/huggingface/datasets/pull/713.diff", "patch_url": "https://github.com/huggingface/datasets/pull/713.patch" }
The new pandas-based text reader isn't able to work properly with files that contain carriage return symbols (`\r`). It fails with the following error message: ``` ... File "pandas/_libs/parsers.pyx", line 847, in pandas._libs.parsers.TextReader.read File "pandas/_libs/parsers.pyx", line 874, in pandas._libs.parsers.TextReader._read_low_memory File "pandas/_libs/parsers.pyx", line 918, in pandas._libs.parsers.TextReader._read_rows File "pandas/_libs/parsers.pyx", line 905, in pandas._libs.parsers.TextReader._tokenize_rows File "pandas/_libs/parsers.pyx", line 2042, in pandas._libs.parsers.raise_parser_error pandas.errors.ParserError: Error tokenizing data. C error: Buffer overflow caught - possible malformed input file. ``` ___ I figured out the pandas uses those symbols as line terminators and this eventually causes the error. Explicitly specifying the `lineterminator` fixes that issue and everything works fine. Please, consider this PR as it seems to be a common issue to solve.
https://api.github.com/repos/huggingface/datasets/issues/713/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/712
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/712/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/712/comments
https://api.github.com/repos/huggingface/datasets/issues/712/events
https://github.com/huggingface/datasets/issues/712
714,242,316
MDU6SXNzdWU3MTQyNDIzMTY=
712
Error in the notebooks/Overview.ipynb notebook
{ "login": "subhrm", "id": 850012, "node_id": "MDQ6VXNlcjg1MDAxMg==", "avatar_url": "https://avatars.githubusercontent.com/u/850012?v=4", "gravatar_id": "", "url": "https://api.github.com/users/subhrm", "html_url": "https://github.com/subhrm", "followers_url": "https://api.github.com/users/subhrm/followers", "following_url": "https://api.github.com/users/subhrm/following{/other_user}", "gists_url": "https://api.github.com/users/subhrm/gists{/gist_id}", "starred_url": "https://api.github.com/users/subhrm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/subhrm/subscriptions", "organizations_url": "https://api.github.com/users/subhrm/orgs", "repos_url": "https://api.github.com/users/subhrm/repos", "events_url": "https://api.github.com/users/subhrm/events{/privacy}", "received_events_url": "https://api.github.com/users/subhrm/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Do this:\r\n``` python\r\nsquad_dataset = list_datasets(with_details=True)[datasets.index('squad')]\r\npprint(squad_dataset.__dict__) # It's a simple python dataclass\r\n```", "Thanks! This worked. I have created a PR to fix this in the notebook. " ]
1,601,791,111,000
1,601,915,140,000
1,601,915,140,000
CONTRIBUTOR
null
null
Hi, I got the following error in **cell number 3** while exploring the **Overview.ipynb** notebook in google colab. I used the [link ](https://colab.research.google.com/github/huggingface/datasets/blob/master/notebooks/Overview.ipynb) provided in the main README file to open it in colab. ```python # You can access various attributes of the datasets before downloading them squad_dataset = list_datasets()[datasets.index('squad')] pprint(squad_dataset.__dict__) # It's a simple python dataclass ``` Error message ``` --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) <ipython-input-5-8dc805c4949c> in <module>() 2 squad_dataset = list_datasets()[datasets.index('squad')] 3 ----> 4 pprint(squad_dataset.__dict__) # It's a simple python dataclass AttributeError: 'str' object has no attribute '__dict__' ``` The object `squad_dataset` is a `str` not a `dataclass` .
https://api.github.com/repos/huggingface/datasets/issues/712/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/710
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/710/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/710/comments
https://api.github.com/repos/huggingface/datasets/issues/710/events
https://github.com/huggingface/datasets/pull/710
714,186,999
MDExOlB1bGxSZXF1ZXN0NDk3MzQ1NjQ0
710
fix README typos/ consistency
{ "login": "discdiver", "id": 7703961, "node_id": "MDQ6VXNlcjc3MDM5NjE=", "avatar_url": "https://avatars.githubusercontent.com/u/7703961?v=4", "gravatar_id": "", "url": "https://api.github.com/users/discdiver", "html_url": "https://github.com/discdiver", "followers_url": "https://api.github.com/users/discdiver/followers", "following_url": "https://api.github.com/users/discdiver/following{/other_user}", "gists_url": "https://api.github.com/users/discdiver/gists{/gist_id}", "starred_url": "https://api.github.com/users/discdiver/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/discdiver/subscriptions", "organizations_url": "https://api.github.com/users/discdiver/orgs", "repos_url": "https://api.github.com/users/discdiver/repos", "events_url": "https://api.github.com/users/discdiver/events{/privacy}", "received_events_url": "https://api.github.com/users/discdiver/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,763,656,000
1,602,928,365,000
1,602,928,365,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/710", "html_url": "https://github.com/huggingface/datasets/pull/710", "diff_url": "https://github.com/huggingface/datasets/pull/710.diff", "patch_url": "https://github.com/huggingface/datasets/pull/710.patch" }
https://api.github.com/repos/huggingface/datasets/issues/710/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/709
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/709/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/709/comments
https://api.github.com/repos/huggingface/datasets/issues/709/events
https://github.com/huggingface/datasets/issues/709
714,067,902
MDU6SXNzdWU3MTQwNjc5MDI=
709
How to use similarity settings other then "BM25" in Elasticsearch index ?
{ "login": "nsankar", "id": 431890, "node_id": "MDQ6VXNlcjQzMTg5MA==", "avatar_url": "https://avatars.githubusercontent.com/u/431890?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nsankar", "html_url": "https://github.com/nsankar", "followers_url": "https://api.github.com/users/nsankar/followers", "following_url": "https://api.github.com/users/nsankar/following{/other_user}", "gists_url": "https://api.github.com/users/nsankar/gists{/gist_id}", "starred_url": "https://api.github.com/users/nsankar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nsankar/subscriptions", "organizations_url": "https://api.github.com/users/nsankar/orgs", "repos_url": "https://api.github.com/users/nsankar/repos", "events_url": "https://api.github.com/users/nsankar/events{/privacy}", "received_events_url": "https://api.github.com/users/nsankar/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Datasets does not use elasticsearch API to define custom similarity. If you want to use a custom similarity, the best would be to run a curl request directly to your elasticsearch instance (see sample hereafter, directly from ES documentation), then you should be able to use `my_similarity` in your configuration passed to datasets\r\n\r\n```\r\ncurl -X PUT \"localhost:9200/index?pretty\" -H 'Content-Type: application/json' -d'\r\n{\r\n \"settings\": {\r\n \"index\": {\r\n \"similarity\": {\r\n \"my_similarity\": {\r\n \"type\": \"DFR\",\r\n \"basic_model\": \"g\",\r\n \"after_effect\": \"l\",\r\n \"normalization\": \"h2\",\r\n \"normalization.h2.c\": \"3.0\"\r\n }\r\n }\r\n }\r\n }\r\n}\r\n'\r\n\r\n```" ]
1,601,723,929,000
1,626,634,975,000
null
NONE
null
null
**QUESTION : How should we use other similarity algorithms supported by Elasticsearch other than "BM25" ?** **ES Reference** https://www.elastic.co/guide/en/elasticsearch/reference/current/index-modules-similarity.html **HF doc reference:** https://huggingface.co/docs/datasets/faiss_and_ea.html **context :** ======== I used the latest Elasticsearch server version 7.9.2 When I set DFR which is one of the other similarity algorithms supported by elasticsearch in the mapping, I get an error For example DFR that I had tried in the first instance in mappings as below., `"mappings": {"properties": {"text": {"type": "text", "analyzer": "standard", "similarity": "DFR"}}},` I get the following error RequestError: RequestError(400, 'mapper_parsing_exception', 'Unknown Similarity type [DFR] for field [text]') The other thing as another option I had tried was to declare "similarity": "my_similarity" within settings and then assigning "my_similarity" inside the mappings as below `es_config = { "settings": { "number_of_shards": 1, **"similarity": "my_similarity"**: { "type": "DFR", "basic_model": "g", "after_effect": "l", "normalization": "h2", "normalization.h2.c": "3.0" } , "analysis": {"analyzer": {"stop_standard": {"type": "standard", " stopwords": "_english_"}}}, }, "mappings": {"properties": {"text": {"type": "text", "analyzer": "standard", "similarity": "my_similarity"}}}, }` For this , I got the following error RequestError: RequestError(400, 'illegal_argument_exception', 'unknown setting [index.similarity] please check that any required plugins are installed, or check the breaking changes documentation for removed settings')
https://api.github.com/repos/huggingface/datasets/issues/709/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/708
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/708/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/708/comments
https://api.github.com/repos/huggingface/datasets/issues/708/events
https://github.com/huggingface/datasets/issues/708
714,020,953
MDU6SXNzdWU3MTQwMjA5NTM=
708
Datasets performance slow? - 6.4x slower than in memory dataset
{ "login": "eugeneware", "id": 38154, "node_id": "MDQ6VXNlcjM4MTU0", "avatar_url": "https://avatars.githubusercontent.com/u/38154?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eugeneware", "html_url": "https://github.com/eugeneware", "followers_url": "https://api.github.com/users/eugeneware/followers", "following_url": "https://api.github.com/users/eugeneware/following{/other_user}", "gists_url": "https://api.github.com/users/eugeneware/gists{/gist_id}", "starred_url": "https://api.github.com/users/eugeneware/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eugeneware/subscriptions", "organizations_url": "https://api.github.com/users/eugeneware/orgs", "repos_url": "https://api.github.com/users/eugeneware/repos", "events_url": "https://api.github.com/users/eugeneware/events{/privacy}", "received_events_url": "https://api.github.com/users/eugeneware/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Facing a similar issue here. My model using SQuAD dataset takes about 1h to process with in memory data and more than 2h with datasets directly.", "And if you use in-memory-data with datasets with `load_dataset(..., keep_in_memory=True)`?", "Thanks for the tip @thomwolf ! I did not see that flag in the docs. I'll try with that.", "We should add it indeed and also maybe a specific section with all the tips for maximal speed. What do you think @lhoestq @SBrandeis @yjernite ?", "By default the datasets loaded with `load_dataset` live on disk.\r\nIt's possible to load them in memory by using some transforms like `.map(..., keep_in_memory=True)`.\r\n\r\nSmall correction to @thomwolf 's comment above: currently we don't have the `keep_in_memory` parameter for `load_dataset` AFAIK but it would be nice to add it indeed :)", "Yes indeed we should add it!", "Great! Thanks a lot.\r\n\r\nI did a test using `map(..., keep_in_memory=True)` and also a test using in-memory only data.\r\n\r\n```python\r\nfeatures = dataset.map(tokenize, batched=True, remove_columns=dataset['train'].column_names)\r\nfeatures.set_format(type='torch', columns=['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\nfeatures_in_memory = dataset.map(tokenize, batched=True, keep_in_memory=True, remove_columns=dataset['train'].column_names)\r\nfeatures_in_memory.set_format(type='torch', columns=['input_ids', 'token_type_ids', 'attention_mask'])\r\n\r\nin_memory = [features['train'][i] for i in range(len(features['train']))]\r\n```\r\n\r\nFor using the features without any tweak, I got **1min17s** for copying the entire DataLoader to CUDA:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(features['train'], batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nFor using the features mapped with `keep_in_memory=True`, I also got **1min17s** for copying the entire DataLoader to CUDA:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(features_in_memory['train'], batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nAnd for the case using every element in memory, converted from the original dataset, I got **12.5s**:\r\n\r\n```\r\n%%time\r\n\r\nfor i, batch in enumerate(DataLoader(in_memory, batch_size=16, num_workers=4)):\r\n batch['input_ids'].to(device)\r\n```\r\n\r\nTaking a closer look in my SQuAD code, using a profiler, I see a lot of calls to `posix read` api. It seems that it is really reliying on disk, which results in a very high train time.", "I am having the same issue here. When loading from memory I can get the GPU up to 70% util but when loading after mapping I can only get 40%.\r\n\r\nIn disk:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='/home/ad/Desktop/bookcorpus', split='train[:20%]')\r\nbook_corpus = book_corpus.map(encode, batched=True, num_proc=20, load_from_cache_file=True, batch_size=2500)\r\nbook_corpus.set_format(type='torch', columns=['text', \"input_ids\", \"attention_mask\", \"token_type_ids\"])\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\"./mobile_bert_big\",\r\n overwrite_output_dir=True,\r\n num_train_epochs=1,\r\n per_device_train_batch_size=32,\r\n per_device_eval_batch_size=16,\r\n save_steps=50,\r\n save_total_limit=2,\r\n logging_first_step=True,\r\n warmup_steps=100,\r\n logging_steps=50,\r\n eval_steps=100,\r\n no_cuda=False,\r\n gradient_accumulation_steps=16,\r\n fp16=True)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=book_corpus,\r\n tokenizer=tokenizer)\r\n```\r\n\r\nIn disk I can only get 0,17 it/s:\r\n`[ 13/28907 01:03 < 46:03:27, 0.17 it/s, Epoch 0.00/1] `\r\n\r\nIf I load it with torch.utils.data.Dataset()\r\n```\r\nclass BCorpusDataset(torch.utils.data.Dataset):\r\n def __init__(self, encodings):\r\n self.encodings = encodings\r\n\r\n def __getitem__(self, idx):\r\n item = [torch.tensor(val[idx]) for key, val in self.encodings.items()][0]\r\n return item\r\n\r\n def __len__(self):\r\n length = [len(val) for key, val in self.encodings.items()][0]\r\n return length\r\n\r\n**book_corpus = book_corpus.select([i for i in range(16*2000)])** # filtering to not have 20% of BC in memory...\r\nbook_corpus = book_corpus(book_corpus)\r\n```\r\nI can get:\r\n` [ 5/62 00:09 < 03:03, 0.31 it/s, Epoch 0.06/1]`\r\n\r\nBut obviously I can not get BookCorpus in memory xD\r\n\r\nEDIT: it is something weird. If i load in disk 1% of bookcorpus:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='/home/ad/Desktop/bookcorpus', split='train[:1%]')\r\n```\r\n\r\nI can get 0.28 it/s, (the same that in memory) but if I load 20% of bookcorpus:\r\n```\r\nbook_corpus = load_dataset('bookcorpus', 'plain_text', cache_dir='/home/ad/Desktop/bookcorpus', split='train[:20%]')\r\n```\r\nI get again 0.17 it/s. \r\n\r\nI am missing something? I think it is something related to size, and not disk or in-memory.", "There is a way to increase the batches read from memory? or multiprocessed it? I think that one of two or it is reading with just 1 core o it is reading very small chunks from disk and left my GPU at 0 between batches", "My fault! I had not seen the `dataloader_num_workers` in `TrainingArguments` ! Now I can parallelize and go fast! Sorry, and thanks." ]
1,601,707,447,000
1,613,139,208,000
1,613,139,208,000
NONE
null
null
I've been very excited about this amazing datasets project. However, I've noticed that the performance can be substantially slower than using an in-memory dataset. Now, this is expected I guess, due to memory mapping data using arrow files, and you don't get anything for free. But I was surprised at how much slower. For example, in the `yelp_polarity` dataset (560000 datapoints, or 17500 batches of 32), it was taking me 3:31 to just get process the data and get it on the GPU (no model involved). Whereas, the equivalent in-memory dataset would finish in just 0:33. Is this expected? Given that one of the goals of this project is also accelerate dataset processing, this seems a bit slower than I would expect. I understand the advantages of being able to work on datasets that exceed memory, and that's very exciting to me, but thought I'd open this issue to discuss. For reference I'm running a AMD Ryzen Threadripper 1900X 8-Core Processor CPU, with 128 GB of RAM and an NVME SSD Samsung 960 EVO. I'm running with an RTX Titan 24GB GPU. I can see with `iotop` that the dataset gets quickly loaded into the system read buffers, and thus doesn't incur any additional IO reads. Thus in theory, all the data *should* be in RAM, but in my benchmark code below it's still 6.4 times slower. What am I doing wrong? And is there a way to force the datasets to completely load into memory instead of being memory mapped in cases where you want maximum performance? At 3:31 for 17500 batches, that's 12ms per batch. Does this 12ms just become insignificant as a proportion of forward and backward passes in practice, and thus it's not worth worrying about this in practice? In any case, here's my code `benchmark.py`. If you run it with an argument of `memory` it will copy the data into memory before executing the same test. ``` py import sys from datasets import load_dataset from transformers import DataCollatorWithPadding, BertTokenizerFast from torch.utils.data import DataLoader from tqdm import tqdm if __name__ == '__main__': tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased') collate_fn = DataCollatorWithPadding(tokenizer, padding=True) ds = load_dataset('yelp_polarity') def do_tokenize(x): return tokenizer(x['text'], truncation=True) ds = ds.map(do_tokenize, batched=True) ds.set_format('torch', ['input_ids', 'token_type_ids', 'attention_mask']) if len(sys.argv) == 2 and sys.argv[1] == 'memory': # copy to memory - probably a faster way to do this - but demonstrates the point # approximately 530 batches per second - 17500 batches in 0:33 print('using memory') _ds = [data for data in tqdm(ds['train'])] else: # approximately 83 batches per second - 17500 batches in 3:31 print('using datasets') _ds = ds['train'] dl = DataLoader(_ds, shuffle=True, collate_fn=collate_fn, batch_size=32, num_workers=4) for data in tqdm(dl): for k, v in data.items(): data[k] = v.to('cuda') ``` For reference, my conda environment is [here](https://gist.github.com/05b6101518ff70ed42a858b302a0405d) Once again, I'm very excited about this library, and how easy it is to load datasets, and to do so without worrying about system memory constraints. Thanks for all your great work.
https://api.github.com/repos/huggingface/datasets/issues/708/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/707
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/707/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/707/comments
https://api.github.com/repos/huggingface/datasets/issues/707/events
https://github.com/huggingface/datasets/issues/707
713,954,666
MDU6SXNzdWU3MTM5NTQ2NjY=
707
Requirements should specify pyarrow<1
{ "login": "mathcass", "id": 918541, "node_id": "MDQ6VXNlcjkxODU0MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/918541?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mathcass", "html_url": "https://github.com/mathcass", "followers_url": "https://api.github.com/users/mathcass/followers", "following_url": "https://api.github.com/users/mathcass/following{/other_user}", "gists_url": "https://api.github.com/users/mathcass/gists{/gist_id}", "starred_url": "https://api.github.com/users/mathcass/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mathcass/subscriptions", "organizations_url": "https://api.github.com/users/mathcass/orgs", "repos_url": "https://api.github.com/users/mathcass/repos", "events_url": "https://api.github.com/users/mathcass/events{/privacy}", "received_events_url": "https://api.github.com/users/mathcass/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hello @mathcass I would want to work on this issue. May I do the same? ", "@punitaojha, certainly. Feel free to work on this. Let me know if you need any help or clarity.", "Hello @mathcass \r\n1. I did fork the repository and clone the same on my local system. \r\n\r\n2. Then learnt about how we can publish our package on pypi.org. Also, found some instructions on same in setup.py documentation.\r\n\r\n3. Then I Perplexity document link that you shared above. I created a colab link from there keep both tensorflow and pytorch means a mixed option and tried to run it in colab but I encountered no errors at a point where you mentioned. Can you help me to figure out the issue. \r\n\r\n4.Here is the link of the colab file with my saved responses. \r\nhttps://colab.research.google.com/drive/1hfYz8Ira39FnREbxgwa_goZWpOojp2NH?usp=sharing", "Also, please share some links which made you conclude that pyarrow < 1 would help. ", "Access granted for the colab link. ", "Thanks for looking at this @punitaojha and thanks for sharing the notebook. \r\n\r\nI just tried to reproduce this on my own (based on the environment where I had this issue) and I can't reproduce it somehow. If I run into this again, I'll include some steps to reproduce it. I'll close this as invalid. \r\n\r\nThanks again. ", "I am sorry for hijacking this closed issue, but I believe I was able to reproduce this very issue. Strangely enough, it also turned out that running `pip install \"pyarrow<1\" --upgrade` did indeed fix the issue (PyArrow was installed in version `0.14.1` in my case).\r\n\r\nPlease see the Colab below:\r\n\r\nhttps://colab.research.google.com/drive/15QQS3xWjlKW2aK0J74eEcRFuhXUddUST\r\n\r\nThanks!" ]
1,601,681,979,000
1,607,070,159,000
1,601,844,628,000
NONE
null
null
I was looking at the docs on [Perplexity](https://huggingface.co/transformers/perplexity.html) via GPT2. When you load datasets and try to load Wikitext, you get the error, ``` module 'pyarrow' has no attribute 'PyExtensionType' ``` I traced it back to datasets having installed PyArrow 1.0.1 but there's not pinning in the setup file. https://github.com/huggingface/datasets/blob/e86a2a8f869b91654e782c9133d810bb82783200/setup.py#L68 Downgrading by installing `pip install "pyarrow<1"` resolved the issue.
https://api.github.com/repos/huggingface/datasets/issues/707/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/706
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/706/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/706/comments
https://api.github.com/repos/huggingface/datasets/issues/706/events
https://github.com/huggingface/datasets/pull/706
713,721,959
MDExOlB1bGxSZXF1ZXN0NDk2OTkwMDA0
706
Fix config creation for data files with NamedSplit
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,653,609,000
1,601,885,700,000
1,601,885,699,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/706", "html_url": "https://github.com/huggingface/datasets/pull/706", "diff_url": "https://github.com/huggingface/datasets/pull/706.diff", "patch_url": "https://github.com/huggingface/datasets/pull/706.patch" }
During config creation, we need to iterate through the data files of all the splits to compute a hash. To make sure the hash is unique given a certain combination of files/splits, we sort the split names. However the `NamedSplit` objects can't be passed to `sorted` and currently it raises an error: we need to sort the string of their names instead. Fix #705
https://api.github.com/repos/huggingface/datasets/issues/706/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/705
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/705/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/705/comments
https://api.github.com/repos/huggingface/datasets/issues/705/events
https://github.com/huggingface/datasets/issues/705
713,709,100
MDU6SXNzdWU3MTM3MDkxMDA=
705
TypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit'
{ "login": "pvcastro", "id": 12713359, "node_id": "MDQ6VXNlcjEyNzEzMzU5", "avatar_url": "https://avatars.githubusercontent.com/u/12713359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pvcastro", "html_url": "https://github.com/pvcastro", "followers_url": "https://api.github.com/users/pvcastro/followers", "following_url": "https://api.github.com/users/pvcastro/following{/other_user}", "gists_url": "https://api.github.com/users/pvcastro/gists{/gist_id}", "starred_url": "https://api.github.com/users/pvcastro/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pvcastro/subscriptions", "organizations_url": "https://api.github.com/users/pvcastro/orgs", "repos_url": "https://api.github.com/users/pvcastro/repos", "events_url": "https://api.github.com/users/pvcastro/events{/privacy}", "received_events_url": "https://api.github.com/users/pvcastro/received_events", "type": "User", "site_admin": false }
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false } ]
null
[ "Hi !\r\nThanks for reporting :) \r\nIndeed this is an issue on the `datasets` side.\r\nI'm creating a PR", "Thanks @lhoestq !" ]
1,601,652,475,000
1,601,885,699,000
1,601,885,699,000
NONE
null
null
## Environment info <!-- You can run the command `transformers-cli env` and copy-and-paste its output below. Don't forget to fill out the missing fields in that output! --> - `transformers` version: 3.3.1 (installed from master) - `datasets` version: 1.0.2 (installed as a dependency from transformers) - Platform: Linux-4.15.0-118-generic-x86_64-with-debian-stretch-sid - Python version: 3.7.9 I'm testing my own text classification dataset using [this example](https://github.com/huggingface/transformers/tree/master/examples/text-classification#run-generic-text-classification-script-in-tensorflow) from transformers. The dataset is split into train / dev / test, and in csv format, containing just a text and a label columns, using comma as sep. Here's a sample: ``` text,label "Registra-se a presença do acadêmico <name> . <REL_SEP> Ao me deparar com a descrição de dois autores no polo ativo da ação junto ao PJe , margem esquerda foi informado pela procuradora do reclamante que se trata de uma reclamação trabalhista individual . <REL_SEP> Diante disso , face a ausência injustificada do autor <name> , determina-se o ARQUIVAMENTO do presente processo , com relação a este , nos termos do [[ art . 844 da CLT ]] . <REL_SEP> CUSTAS AUTOR - DISPENSADO <REL_SEP> Custas pelo autor no importe de R $326,82 , calculadas sobre R $16.341,03 , dispensadas na forma da lei , em virtude da concessão dos benefícios da Justiça Gratuita , ora deferida . <REL_SEP> Cientes os presentes . <REL_SEP> Audiência encerrada às 8h42min . <REL_SEP> <name> <REL_SEP> Juíza do Trabalho <REL_SEP> Ata redigida por << <name> >> , Secretário de Audiência .",NO_RELATION ``` However, @Santosh-Gupta reported in #7351 that he had the exact same problem using the ChemProt dataset. His colab notebook is referenced in the following section. ## To reproduce Steps to reproduce the behavior: 1. Created a new conda environment using conda env -n transformers python=3.7 2. Cloned transformers master, `cd` into it and installed using pip install --editable . -r examples/requirements.txt 3. Installed tensorflow with `pip install tensorflow` 3. Ran `run_tf_text_classification.py` with the following parameters: ``` --train_file <DATASET_PATH>/train.csv \ --dev_file <DATASET_PATH>/dev.csv \ --test_file <DATASET_PATH>/test.csv \ --label_column_id 1 \ --model_name_or_path neuralmind/bert-base-portuguese-cased \ --output_dir <OUTPUT_PATH> \ --num_train_epochs 4 \ --per_device_train_batch_size 4 \ --per_device_eval_batch_size 4 \ --do_train \ --do_eval \ --do_predict \ --logging_steps 1000 \ --evaluate_during_training \ --save_steps 1000 \ --overwrite_output_dir \ --overwrite_cache ``` I have also copied [@Santosh-Gupta 's colab notebook](https://colab.research.google.com/drive/11APei6GjphCZbH5wD9yVlfGvpIkh8pwr?usp=sharing) as a reference. <!-- If you have code snippets, error messages, stack traces please provide them here as well. Important! Use code tags to correctly format your code. See https://help.github.com/en/github/writing-on-github/creating-and-highlighting-code-blocks#syntax-highlighting Do not use screenshots, as they are hard to read and (more importantly) don't allow others to copy-and-paste your code.--> Here is the stack trace: ``` 2020-10-02 07:33:41.622011: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 /media/discoD/repositorios/transformers_pedro/src/transformers/training_args.py:333: FutureWarning: The `evaluate_during_training` argument is deprecated in favor of `evaluation_strategy` (which has more options) FutureWarning, 2020-10-02 07:33:43.471648: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcuda.so.1 2020-10-02 07:33:43.471791: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:43.472664: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1716] Found device 0 with properties: pciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1 coreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB/s 2020-10-02 07:33:43.472684: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 2020-10-02 07:33:43.472765: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10 2020-10-02 07:33:43.472809: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10 2020-10-02 07:33:43.472848: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10 2020-10-02 07:33:43.474209: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10 2020-10-02 07:33:43.474276: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10 2020-10-02 07:33:43.561219: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7 2020-10-02 07:33:43.561397: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:43.562345: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:43.563219: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1858] Adding visible gpu devices: 0 2020-10-02 07:33:43.563595: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations: AVX2 FMA To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags. 2020-10-02 07:33:43.570091: I tensorflow/core/platform/profile_utils/cpu_utils.cc:104] CPU Frequency: 3591830000 Hz 2020-10-02 07:33:43.570494: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x560842432400 initialized for platform Host (this does not guarantee that XLA will be used). Devices: 2020-10-02 07:33:43.570511: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): Host, Default Version 2020-10-02 07:33:43.570702: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:43.571599: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1716] Found device 0 with properties: pciBusID: 0000:01:00.0 name: GeForce GTX 1070 computeCapability: 6.1 coreClock: 1.7085GHz coreCount: 15 deviceMemorySize: 7.92GiB deviceMemoryBandwidth: 238.66GiB/s 2020-10-02 07:33:43.571633: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 2020-10-02 07:33:43.571645: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10 2020-10-02 07:33:43.571654: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10 2020-10-02 07:33:43.571664: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10 2020-10-02 07:33:43.571691: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10 2020-10-02 07:33:43.571704: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10 2020-10-02 07:33:43.571718: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7 2020-10-02 07:33:43.571770: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:43.572641: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:43.573475: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1858] Adding visible gpu devices: 0 2020-10-02 07:33:47.139227: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1257] Device interconnect StreamExecutor with strength 1 edge matrix: 2020-10-02 07:33:47.139265: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1263] 0 2020-10-02 07:33:47.139272: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1276] 0: N 2020-10-02 07:33:47.140323: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:47.141248: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:47.142085: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-10-02 07:33:47.142854: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1402] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 5371 MB memory) -> physical GPU (device: 0, name: GeForce GTX 1070, pci bus id: 0000:01:00.0, compute capability: 6.1) 2020-10-02 07:33:47.146317: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x5608b95dc5c0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices: 2020-10-02 07:33:47.146336: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): GeForce GTX 1070, Compute Capability 6.1 10/02/2020 07:33:47 - INFO - __main__ - n_replicas: 1, distributed training: False, 16-bits training: False 10/02/2020 07:33:47 - INFO - __main__ - Training/evaluation parameters TFTrainingArguments(output_dir='/media/discoD/models/datalawyer/pedidos/transformers_tf', overwrite_output_dir=True, do_train=True, do_eval=True, do_predict=True, evaluate_during_training=True, evaluation_strategy=<EvaluationStrategy.STEPS: 'steps'>, prediction_loss_only=False, per_device_train_batch_size=4, per_device_eval_batch_size=4, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=1, learning_rate=5e-05, weight_decay=0.0, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=4.0, max_steps=-1, warmup_steps=0, logging_dir='runs/Oct02_07-33-43_user-XPS-8700', logging_first_step=False, logging_steps=1000, save_steps=1000, save_total_limit=None, no_cuda=False, seed=42, fp16=False, fp16_opt_level='O1', local_rank=-1, tpu_num_cores=None, tpu_metrics_debug=False, debug=False, dataloader_drop_last=False, eval_steps=1000, dataloader_num_workers=0, past_index=-1, run_name='/media/discoD/models/datalawyer/pedidos/transformers_tf', disable_tqdm=False, remove_unused_columns=True, label_names=None, load_best_model_at_end=False, metric_for_best_model=None, greater_is_better=False, tpu_name=None, xla=False) 10/02/2020 07:33:53 - INFO - filelock - Lock 140407857405776 acquired on /home/user/.cache/huggingface/datasets/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock 10/02/2020 07:33:53 - INFO - filelock - Lock 140407857405776 released on /home/user/.cache/huggingface/datasets/e0f1e9ed46db1e2429189f06b479cbd4075c0976104c1aacf8f77d9a53d2ad87.03756fef6da334f50a7ff73608e21b5018229944ca250416ce7352e25d84a552.py.lock Using custom data configuration default Traceback (most recent call last): File "run_tf_text_classification.py", line 283, in <module> main() File "run_tf_text_classification.py", line 222, in main max_seq_length=data_args.max_seq_length, File "run_tf_text_classification.py", line 43, in get_tfds ds = datasets.load_dataset("csv", data_files=files) File "/media/discoD/anaconda3/envs/transformers/lib/python3.7/site-packages/datasets/load.py", line 604, in load_dataset **config_kwargs, File "/media/discoD/anaconda3/envs/transformers/lib/python3.7/site-packages/datasets/builder.py", line 158, in __init__ **config_kwargs, File "/media/discoD/anaconda3/envs/transformers/lib/python3.7/site-packages/datasets/builder.py", line 269, in _create_builder_config for key in sorted(data_files.keys()): TypeError: '<' not supported between instances of 'NamedSplit' and 'NamedSplit' ``` ## Expected behavior Should be able to run the text-classification example as described in [https://github.com/huggingface/transformers/tree/master/examples/text-classification#run-generic-text-classification-script-in-tensorflow](https://github.com/huggingface/transformers/tree/master/examples/text-classification#run-generic-text-classification-script-in-tensorflow) Originally opened this issue at transformers' repository: [https://github.com/huggingface/transformers/issues/7535](https://github.com/huggingface/transformers/issues/7535). @jplu instructed me to open here, since according to [this](https://github.com/huggingface/transformers/issues/7535#issuecomment-702778885) evidence, the problem is from datasets. Thanks!
https://api.github.com/repos/huggingface/datasets/issues/705/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/704
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/704/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/704/comments
https://api.github.com/repos/huggingface/datasets/issues/704/events
https://github.com/huggingface/datasets/pull/704
713,572,556
MDExOlB1bGxSZXF1ZXN0NDk2ODY2NTQ0
704
Fix remote tests for new datasets
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,640,484,000
1,601,640,722,000
1,601,640,721,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/704", "html_url": "https://github.com/huggingface/datasets/pull/704", "diff_url": "https://github.com/huggingface/datasets/pull/704.diff", "patch_url": "https://github.com/huggingface/datasets/pull/704.patch" }
When adding a new dataset, the remote tests fail because they try to get the new dataset from the master branch (i.e., where the dataset doesn't exist yet) To fix that I reverted to the use of the HF API that fetch the available datasets on S3 that is synced with the master branch
https://api.github.com/repos/huggingface/datasets/issues/704/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/703
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/703/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/703/comments
https://api.github.com/repos/huggingface/datasets/issues/703/events
https://github.com/huggingface/datasets/pull/703
713,559,718
MDExOlB1bGxSZXF1ZXN0NDk2ODU1OTQ5
703
Add hotpot QA
{ "login": "ghomasHudson", "id": 13795113, "node_id": "MDQ6VXNlcjEzNzk1MTEz", "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghomasHudson", "html_url": "https://github.com/ghomasHudson", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gists_url": "https://api.github.com/users/ghomasHudson/gists{/gist_id}", "starred_url": "https://api.github.com/users/ghomasHudson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ghomasHudson/subscriptions", "organizations_url": "https://api.github.com/users/ghomasHudson/orgs", "repos_url": "https://api.github.com/users/ghomasHudson/repos", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "received_events_url": "https://api.github.com/users/ghomasHudson/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Awesome :) \r\n\r\nDon't pay attention to the RemoteDatasetTest error, I'm fixing it right now", "You can rebase from master to fix the CI test :)", "If we're lucky we can even include this dataset in today's release", "Just thinking since `type` can only be `comparison` or `bridge` and `level` can only be `easy`, `medium`, `hard` should they be `ClassLabel`?", "> Just thinking since `type` can only be `comparison` or `bridge` and `level` can only be `easy`, `medium`, `hard` should they be `ClassLabel`?\r\n\r\nI think it's more a tag than a label. I guess a string is fine\r\n" ]
1,601,639,068,000
1,601,643,281,000
1,601,643,281,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/703", "html_url": "https://github.com/huggingface/datasets/pull/703", "diff_url": "https://github.com/huggingface/datasets/pull/703.diff", "patch_url": "https://github.com/huggingface/datasets/pull/703.patch" }
Added the [HotpotQA](https://github.com/hotpotqa/hotpot) multi-hop question answering dataset.
https://api.github.com/repos/huggingface/datasets/issues/703/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/702
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/702/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/702/comments
https://api.github.com/repos/huggingface/datasets/issues/702/events
https://github.com/huggingface/datasets/pull/702
713,499,628
MDExOlB1bGxSZXF1ZXN0NDk2ODA3Mjg4
702
Complete rouge kwargs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,632,741,000
1,601,633,464,000
1,601,633,463,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/702", "html_url": "https://github.com/huggingface/datasets/pull/702", "diff_url": "https://github.com/huggingface/datasets/pull/702.diff", "patch_url": "https://github.com/huggingface/datasets/pull/702.patch" }
In #701 we noticed that some kwargs were missing for rouge
https://api.github.com/repos/huggingface/datasets/issues/702/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/701
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/701/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/701/comments
https://api.github.com/repos/huggingface/datasets/issues/701/events
https://github.com/huggingface/datasets/pull/701
713,485,757
MDExOlB1bGxSZXF1ZXN0NDk2Nzk2MTQ1
701
Add rouge 2 and rouge Lsum to rouge metric outputs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Oups too late, sorry" ]
1,601,631,346,000
1,601,632,514,000
1,601,632,338,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/701", "html_url": "https://github.com/huggingface/datasets/pull/701", "diff_url": "https://github.com/huggingface/datasets/pull/701.diff", "patch_url": "https://github.com/huggingface/datasets/pull/701.patch" }
Continuation of #700 Rouge 2 and Rouge Lsum were missing in Rouge's outputs. Rouge Lsum is also useful to evaluate Rouge L for sentences with `\n` Fix #617
https://api.github.com/repos/huggingface/datasets/issues/701/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/700
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/700/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/700/comments
https://api.github.com/repos/huggingface/datasets/issues/700/events
https://github.com/huggingface/datasets/pull/700
713,450,295
MDExOlB1bGxSZXF1ZXN0NDk2NzY3MTMz
700
Add rouge-2 in rouge_types for metric calculation
{ "login": "Shashi456", "id": 18056781, "node_id": "MDQ6VXNlcjE4MDU2Nzgx", "avatar_url": "https://avatars.githubusercontent.com/u/18056781?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Shashi456", "html_url": "https://github.com/Shashi456", "followers_url": "https://api.github.com/users/Shashi456/followers", "following_url": "https://api.github.com/users/Shashi456/following{/other_user}", "gists_url": "https://api.github.com/users/Shashi456/gists{/gist_id}", "starred_url": "https://api.github.com/users/Shashi456/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Shashi456/subscriptions", "organizations_url": "https://api.github.com/users/Shashi456/orgs", "repos_url": "https://api.github.com/users/Shashi456/repos", "events_url": "https://api.github.com/users/Shashi456/events{/privacy}", "received_events_url": "https://api.github.com/users/Shashi456/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Indeed there's currently a mismatch between the description and what it rouge actually returns.\r\nThanks for proposing this fix :) \r\n\r\nI think it's better to return rouge 1-2-L.\r\nWas there a reason to only include rouge 1 and rouge L @thomwolf ? ", "rougeLsum is also missing, could you add it ?", "Adding `RougeLSum` would fix https://github.com/huggingface/datasets/issues/617", "I am opening a PR with both of them right now actually :)", "Also the format of the output isn't exactly ideal, It's usually only the F-1 score that is cared about. \r\n\r\nFormatting the output to reflect how `ROUGE-1-5-5` (the perl version thats usually used and pyrouge is a wrapper over it), would be better.\r\n\r\n", "I'll close this since you seem to have already added it in another PR. Sorry for the delay in responding to you @lhoestq.", "What do you mean by \"Formatting the output to reflect how ROUGE-1-5-5\" @Shashi456 ?", "I like the idea of returning all the scores for two reason:\r\n- Rouge's aggregator does sampling and therefore it returns \"low\" \"mid\" and \"high\" scores\r\n- It is interesting to have the precision and recall to see how the F1 score was computed\r\nBut I understand your point that returning only the F1 score makes sense since it's the one that's always used ", "@thomwolf the scores now returned look like this:\r\n```\r\n{'rouge1': AggregateScore(low=Score(precision=0.16620308156871524, recall=0.18219819615984395, fmeasure=0.16226017699359463), mid=Score(precision=0.17274338501705871, recall=0.1890957812369246, fmeasure=0.16823877588620403), high=Score(precision=0.17934569582981455, recall=0.1965626706042028, fmeasure=0.17491509794856058)), \r\n'rouge2': AggregateScore(low=Score(precision=0.12478835737689957, recall=0.1362113231755514, fmeasure=0.12055941950062395), mid=Score(precision=0.1303967602691664, recall=0.1423747229852964, fmeasure=0.1258363976151122), high=Score(precision=0.13654527560789362, recall=0.1488071465116122, fmeasure=0.13184989406704056)), \r\n'rougeL': AggregateScore(low=Score(precision=0.16568068818352072, recall=0.1811919016674486, fmeasure=0.1614784523482225), mid=Score(precision=0.17156684723552357, recall=0.1879777628247058, fmeasure=0.16720699286250762), high=Score(precision=0.17788847350584547, recall=0.1948899838530898, fmeasure=0.17316501523379826))}\r\n```\r\n\r\nWhile when computed through the perl rouge script, it looks like:\r\n```\r\nROUGE-1 Average_R: 0.34775 (95%-conf.int. 0.34546 - 0.35025)\r\nROUGE-1 Average_P: 0.19381 (95%-conf.int. 0.19246 - 0.19538)\r\nROUGE-1 Average_F: 0.24070 (95%-conf.int. 0.23925 - 0.24230)\r\n---------------------------------------------\r\nROUGE-2 Average_R: 0.07160 (95%-conf.int. 0.07010 - 0.07298)\r\nROUGE-2 Average_F: 0.04845 (95%-conf.int. 0.04741 - 0.04942)\r\n---------------------------------------------\r\nROUGE-L Average_R: 0.26404 (95%-conf.int. 0.26215 - 0.26598)\r\nROUGE-L Average_P: 0.14696 (95%-conf.int. 0.14576 - 0.14815)\r\nROUGE-L Average_F: 0.18245 (95%-conf.int. 0.18120 - 0.18367)\r\n```\r\nwhile the wrapper returns the much more readable:\r\n```\r\n[2020-07-30 18:13:38,556 INFO] Rouges at step 13000 \r\n>> ROUGE-F(1/2/3/l): 43.43/20.42/39.78 \r\nROUGE-R(1/2/3/l): 53.91/25.34/49.32\r\n```\r\n\r\nThe formatting allows for easy reading, and although \"low\", \"mid\", \"high\" make sense, this is more concise and effective. \r\n\r\nOne way of changing this might be to return a dictionary that returns values like `rouge_1_precision`, `rouge_1_F1`, `rouge_1_recall`, and maybe also having the ability to get the values you are interested in and keeping `recall` and `F1` as default.", "cc: @lhoestq ", "Ok I see.\r\nI think it's also important to follow one of the existing output format (there are already too many different formats, let's try not to add another different one)\r\nI'd still stick with the current format and not transform the output of the python implementation of rouge since it's already widely used.\r\nWhat do you think ?", "Maybe we could convert the dataclasses in dictionnaries, would that help @Shashi456 ?", "@thomwolf yeah I think that would help. I initially didn't understand the high low mid categories. Dictionaries could help in this case I guess, and if we allow the user to choose what they want i.e F1 and precision or recall." ]
1,601,627,805,000
1,601,636,929,000
1,601,632,745,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/700", "html_url": "https://github.com/huggingface/datasets/pull/700", "diff_url": "https://github.com/huggingface/datasets/pull/700.diff", "patch_url": "https://github.com/huggingface/datasets/pull/700.patch" }
The description of the ROUGE metric says, ``` _KWARGS_DESCRIPTION = """ Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each predictions should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. Returns: rouge1: rouge_1 f1, rouge2: rouge_2 f1, rougeL: rouge_l f1, rougeLsum: rouge_l precision """ ``` but the `rouge_types` argument defaults to `rouge_types = ["rouge1", "rougeL"]`, this PR updates and add `rouge2` to the list so as to reflect the description card.
https://api.github.com/repos/huggingface/datasets/issues/700/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/699
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/699/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/699/comments
https://api.github.com/repos/huggingface/datasets/issues/699/events
https://github.com/huggingface/datasets/issues/699
713,395,642
MDU6SXNzdWU3MTMzOTU2NDI=
699
XNLI dataset is not loading
{ "login": "imadarsh1001", "id": 14936525, "node_id": "MDQ6VXNlcjE0OTM2NTI1", "avatar_url": "https://avatars.githubusercontent.com/u/14936525?v=4", "gravatar_id": "", "url": "https://api.github.com/users/imadarsh1001", "html_url": "https://github.com/imadarsh1001", "followers_url": "https://api.github.com/users/imadarsh1001/followers", "following_url": "https://api.github.com/users/imadarsh1001/following{/other_user}", "gists_url": "https://api.github.com/users/imadarsh1001/gists{/gist_id}", "starred_url": "https://api.github.com/users/imadarsh1001/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/imadarsh1001/subscriptions", "organizations_url": "https://api.github.com/users/imadarsh1001/orgs", "repos_url": "https://api.github.com/users/imadarsh1001/repos", "events_url": "https://api.github.com/users/imadarsh1001/events{/privacy}", "received_events_url": "https://api.github.com/users/imadarsh1001/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "also i tried below code to solve checksum error \r\n`datasets-cli test ./datasets/xnli --save_infos --all_configs`\r\n\r\nand it shows \r\n\r\n```\r\n2020-10-02 07:06:16.588760: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1\r\nTraceback (most recent call last):\r\n File \"/opt/conda/lib/python3.7/site-packages/datasets/load.py\", line 268, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"/opt/conda/lib/python3.7/site-packages/datasets/utils/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"/opt/conda/lib/python3.7/site-packages/datasets/utils/file_utils.py\", line 474, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/1.0.2/datasets/./datasets/xnli/xnli.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"/opt/conda/lib/python3.7/site-packages/datasets/load.py\", line 279, in prepare_module\r\n local_path = cached_path(file_path, download_config=download_config)\r\n File \"/opt/conda/lib/python3.7/site-packages/datasets/utils/file_utils.py\", line 308, in cached_path\r\n use_etag=download_config.use_etag,\r\n File \"/opt/conda/lib/python3.7/site-packages/datasets/utils/file_utils.py\", line 474, in get_from_cache\r\n raise FileNotFoundError(\"Couldn't find file at {}\".format(url))\r\nFileNotFoundError: Couldn't find file at https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/./datasets/xnli/xnli.py\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"/opt/conda/bin/datasets-cli\", line 36, in <module>\r\n service.run()\r\n File \"/opt/conda/lib/python3.7/site-packages/datasets/commands/test.py\", line 76, in run\r\n module_path, hash = prepare_module(path)\r\n File \"/opt/conda/lib/python3.7/site-packages/datasets/load.py\", line 283, in prepare_module\r\n combined_path, github_file_path, file_path\r\nFileNotFoundError: Couldn't find file locally at ./datasets/xnli/xnli.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.0.2/datasets/./datasets/xnli/xnli.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/./datasets/xnli/xnli.py\r\n```\r\n\r\n", "Hi !\r\nYes the download url changed.\r\nIt's updated on the master branch. I'm doing a release today to fix that :)", "the issue is fixed with latest release \r\n\r\n" ]
1,601,621,596,000
1,601,747,152,000
1,601,747,017,000
NONE
null
null
`dataset = datasets.load_dataset(path='xnli')` showing below error ``` /opt/conda/lib/python3.7/site-packages/nlp/utils/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name) 36 if len(bad_urls) > 0: 37 error_msg = "Checksums didn't match" + for_verification_name + ":\n" ---> 38 raise NonMatchingChecksumError(error_msg + str(bad_urls)) 39 logger.info("All the checksums matched successfully" + for_verification_name) 40 NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://www.nyu.edu/projects/bowman/xnli/XNLI-1.0.zip'] ``` I think URL is now changed to "https://cims.nyu.edu/~sbowman/xnli/XNLI-MT-1.0.zip"
https://api.github.com/repos/huggingface/datasets/issues/699/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/697
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/697/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/697/comments
https://api.github.com/repos/huggingface/datasets/issues/697/events
https://github.com/huggingface/datasets/pull/697
712,979,029
MDExOlB1bGxSZXF1ZXN0NDk2MzczNDU5
697
Update README.md
{ "login": "bishug", "id": 71011306, "node_id": "MDQ6VXNlcjcxMDExMzA2", "avatar_url": "https://avatars.githubusercontent.com/u/71011306?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bishug", "html_url": "https://github.com/bishug", "followers_url": "https://api.github.com/users/bishug/followers", "following_url": "https://api.github.com/users/bishug/following{/other_user}", "gists_url": "https://api.github.com/users/bishug/gists{/gist_id}", "starred_url": "https://api.github.com/users/bishug/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bishug/subscriptions", "organizations_url": "https://api.github.com/users/bishug/orgs", "repos_url": "https://api.github.com/users/bishug/repos", "events_url": "https://api.github.com/users/bishug/events{/privacy}", "received_events_url": "https://api.github.com/users/bishug/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,568,162,000
1,601,568,720,000
1,601,568,720,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/697", "html_url": "https://github.com/huggingface/datasets/pull/697", "diff_url": "https://github.com/huggingface/datasets/pull/697.diff", "patch_url": "https://github.com/huggingface/datasets/pull/697.patch" }
Hey I was just telling my subscribers to check out your repositories Thank you
https://api.github.com/repos/huggingface/datasets/issues/697/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/696
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/696/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/696/comments
https://api.github.com/repos/huggingface/datasets/issues/696/events
https://github.com/huggingface/datasets/pull/696
712,942,977
MDExOlB1bGxSZXF1ZXN0NDk2MzQzMjEy
696
Elasticsearch index docs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,565,538,000
1,601,624,899,000
1,601,624,898,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/696", "html_url": "https://github.com/huggingface/datasets/pull/696", "diff_url": "https://github.com/huggingface/datasets/pull/696.diff", "patch_url": "https://github.com/huggingface/datasets/pull/696.patch" }
I added the docs for ES indexes. I also added a `load_elasticsearch_index` method to load an index that has already been built. I checked the tests for the ES index and we have tests that mock ElasticSearch. I think this is good for now but at some point it would be cool to have an end-to-end test with a real ES running.
https://api.github.com/repos/huggingface/datasets/issues/696/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/695
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/695/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/695/comments
https://api.github.com/repos/huggingface/datasets/issues/695/events
https://github.com/huggingface/datasets/pull/695
712,843,949
MDExOlB1bGxSZXF1ZXN0NDk2MjU5NTM0
695
Update XNLI download link
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,558,842,000
1,601,560,875,000
1,601,560,874,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/695", "html_url": "https://github.com/huggingface/datasets/pull/695", "diff_url": "https://github.com/huggingface/datasets/pull/695.diff", "patch_url": "https://github.com/huggingface/datasets/pull/695.patch" }
The old link isn't working anymore. I updated it with the new official link. Fix #690
https://api.github.com/repos/huggingface/datasets/issues/695/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/694
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/694/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/694/comments
https://api.github.com/repos/huggingface/datasets/issues/694/events
https://github.com/huggingface/datasets/pull/694
712,827,751
MDExOlB1bGxSZXF1ZXN0NDk2MjQ1NzU0
694
Use GitHub instead of aws in remote dataset tests
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,557,670,000
1,601,624,848,000
1,601,624,847,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/694", "html_url": "https://github.com/huggingface/datasets/pull/694", "diff_url": "https://github.com/huggingface/datasets/pull/694.diff", "patch_url": "https://github.com/huggingface/datasets/pull/694.patch" }
Recently we switched from aws s3 to github to download dataset scripts. However in the tests, the dummy data were still downloaded from s3. So I changed that to download them from github instead, in the MockDownloadManager. Moreover I noticed that `anli`'s dummy data were quite heavy (18MB compressed, i.e. the entire dataset) so I replaced them with dummy data with few examples.
https://api.github.com/repos/huggingface/datasets/issues/694/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/693
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/693/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/693/comments
https://api.github.com/repos/huggingface/datasets/issues/693/events
https://github.com/huggingface/datasets/pull/693
712,822,200
MDExOlB1bGxSZXF1ZXN0NDk2MjQxMjUw
693
Rachel ker add dataset/mlsum
{ "login": "pdhg", "id": 32742136, "node_id": "MDQ6VXNlcjMyNzQyMTM2", "avatar_url": "https://avatars.githubusercontent.com/u/32742136?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdhg", "html_url": "https://github.com/pdhg", "followers_url": "https://api.github.com/users/pdhg/followers", "following_url": "https://api.github.com/users/pdhg/following{/other_user}", "gists_url": "https://api.github.com/users/pdhg/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdhg/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdhg/subscriptions", "organizations_url": "https://api.github.com/users/pdhg/orgs", "repos_url": "https://api.github.com/users/pdhg/repos", "events_url": "https://api.github.com/users/pdhg/events{/privacy}", "received_events_url": "https://api.github.com/users/pdhg/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "It looks like an outdated PR (we've already added mlsum). Closing it" ]
1,601,557,270,000
1,601,571,673,000
1,601,571,673,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/693", "html_url": "https://github.com/huggingface/datasets/pull/693", "diff_url": "https://github.com/huggingface/datasets/pull/693.diff", "patch_url": "https://github.com/huggingface/datasets/pull/693.patch" }
.
https://api.github.com/repos/huggingface/datasets/issues/693/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/692
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/692/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/692/comments
https://api.github.com/repos/huggingface/datasets/issues/692/events
https://github.com/huggingface/datasets/pull/692
712,818,968
MDExOlB1bGxSZXF1ZXN0NDk2MjM4NzIw
692
Update README.md
{ "login": "mayank1897", "id": 62796466, "node_id": "MDQ6VXNlcjYyNzk2NDY2", "avatar_url": "https://avatars.githubusercontent.com/u/62796466?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mayank1897", "html_url": "https://github.com/mayank1897", "followers_url": "https://api.github.com/users/mayank1897/followers", "following_url": "https://api.github.com/users/mayank1897/following{/other_user}", "gists_url": "https://api.github.com/users/mayank1897/gists{/gist_id}", "starred_url": "https://api.github.com/users/mayank1897/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mayank1897/subscriptions", "organizations_url": "https://api.github.com/users/mayank1897/orgs", "repos_url": "https://api.github.com/users/mayank1897/repos", "events_url": "https://api.github.com/users/mayank1897/events{/privacy}", "received_events_url": "https://api.github.com/users/mayank1897/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hacktoberfest spam", "To enhance its readability.....not Hacktoberfest spam", "How is adding a punctuation to the end of a sentence justified as \"To enhance its readability\". \r\nConsidering that this is not your first \"README enhancement '' please don't spam the open source community with useless PR to get a free T-Shirt it just hurts the maintainers.\r\n\r\n//Joey", "closed as spam" ]
1,601,557,042,000
1,601,636,519,000
1,601,636,519,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/692", "html_url": "https://github.com/huggingface/datasets/pull/692", "diff_url": "https://github.com/huggingface/datasets/pull/692.diff", "patch_url": "https://github.com/huggingface/datasets/pull/692.patch" }
https://api.github.com/repos/huggingface/datasets/issues/692/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/691
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/691/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/691/comments
https://api.github.com/repos/huggingface/datasets/issues/691/events
https://github.com/huggingface/datasets/issues/691
712,389,499
MDU6SXNzdWU3MTIzODk0OTk=
691
Add UI filter to filter datasets based on task
{ "login": "praateekmahajan", "id": 7589415, "node_id": "MDQ6VXNlcjc1ODk0MTU=", "avatar_url": "https://avatars.githubusercontent.com/u/7589415?v=4", "gravatar_id": "", "url": "https://api.github.com/users/praateekmahajan", "html_url": "https://github.com/praateekmahajan", "followers_url": "https://api.github.com/users/praateekmahajan/followers", "following_url": "https://api.github.com/users/praateekmahajan/following{/other_user}", "gists_url": "https://api.github.com/users/praateekmahajan/gists{/gist_id}", "starred_url": "https://api.github.com/users/praateekmahajan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/praateekmahajan/subscriptions", "organizations_url": "https://api.github.com/users/praateekmahajan/orgs", "repos_url": "https://api.github.com/users/praateekmahajan/repos", "events_url": "https://api.github.com/users/praateekmahajan/events{/privacy}", "received_events_url": "https://api.github.com/users/praateekmahajan/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Already supported." ]
1,601,513,778,000
1,603,812,270,000
null
NONE
null
null
This is great work, so huge shoutout to contributors and huggingface. The [/nlp/viewer](https://huggingface.co/nlp/viewer/) is great and the [/datasets](https://huggingface.co/datasets) page is great. I was wondering if in both or either places we can have a filter that selects if a dataset is good for the following tasks (non exhaustive list) - Classification - Multi label - Multi class - Q&A - Summarization - Translation I believe this feature might have some value, for folks trying to find datasets for a particular task, and then testing their model capabilities. Thank you :)
https://api.github.com/repos/huggingface/datasets/issues/691/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/690
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/690/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/690/comments
https://api.github.com/repos/huggingface/datasets/issues/690/events
https://github.com/huggingface/datasets/issues/690
712,150,321
MDU6SXNzdWU3MTIxNTAzMjE=
690
XNLI dataset: NonMatchingChecksumError
{ "login": "xiey1", "id": 13307358, "node_id": "MDQ6VXNlcjEzMzA3MzU4", "avatar_url": "https://avatars.githubusercontent.com/u/13307358?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xiey1", "html_url": "https://github.com/xiey1", "followers_url": "https://api.github.com/users/xiey1/followers", "following_url": "https://api.github.com/users/xiey1/following{/other_user}", "gists_url": "https://api.github.com/users/xiey1/gists{/gist_id}", "starred_url": "https://api.github.com/users/xiey1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xiey1/subscriptions", "organizations_url": "https://api.github.com/users/xiey1/orgs", "repos_url": "https://api.github.com/users/xiey1/repos", "events_url": "https://api.github.com/users/xiey1/events{/privacy}", "received_events_url": "https://api.github.com/users/xiey1/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Thanks for reporting.\r\nThe data file must have been updated by the host.\r\nI'll update the checksum with the new one.", "Well actually it looks like the link isn't working anymore :(", "The new link is https://cims.nyu.edu/~sbowman/xnli/XNLI-1.0.zip\r\nI'll update the dataset script", "I'll do a release in the next few days to make the fix available for everyone.\r\nIn the meantime you can load `xnli` with\r\n```\r\nxnli = load_dataset('xnli', script_version=\"master\")\r\n```\r\nThis will use the latest version of the xnli script (available on master branch), instead of the old one.", "That's awesome! Thanks a lot!" ]
1,601,488,203,000
1,601,572,508,000
1,601,560,874,000
NONE
null
null
Hi, I tried to download "xnli" dataset in colab using `xnli = load_dataset(path='xnli')` but got 'NonMatchingChecksumError' error `NonMatchingChecksumError Traceback (most recent call last) <ipython-input-27-a87bedc82eeb> in <module>() ----> 1 xnli = load_dataset(path='xnli') 3 frames /usr/local/lib/python3.6/dist-packages/datasets/utils/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name) 37 if len(bad_urls) > 0: 38 error_msg = "Checksums didn't match" + for_verification_name + ":\n" ---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls)) 40 logger.info("All the checksums matched successfully" + for_verification_name) 41 NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://www.nyu.edu/projects/bowman/xnli/XNLI-1.0.zip']` The same code worked well several days ago in colab but stopped working now. Thanks!
https://api.github.com/repos/huggingface/datasets/issues/690/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/689
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/689/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/689/comments
https://api.github.com/repos/huggingface/datasets/issues/689/events
https://github.com/huggingface/datasets/pull/689
712,095,262
MDExOlB1bGxSZXF1ZXN0NDk1NjMzNjMy
689
Switch to pandas reader for text dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "If the windows tests in the CI pass, today will be a happy day" ]
1,601,483,292,000
1,601,484,332,000
1,601,484,331,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/689", "html_url": "https://github.com/huggingface/datasets/pull/689", "diff_url": "https://github.com/huggingface/datasets/pull/689.diff", "patch_url": "https://github.com/huggingface/datasets/pull/689.patch" }
Following the discussion in #622 , it appears that there's no appropriate ways to use the payrrow csv reader to read text files because of the separator. In this PR I switched to pandas to read the file. Moreover pandas allows to read the file by chunk, which means that you can build the arrow dataset from a text file that is bigger than RAM (we used to have to shard text files an mentioned in https://github.com/huggingface/datasets/issues/610#issuecomment-691672919) From a test that I did locally on a 1GB text file, the pyarrow reader used to run in 150ms while the new one takes 650ms (multithreading off for pyarrow). This is probably due to chunking since I am having the same speed difference by calling `read()` and calling `read(chunksize)` + `readline()` to read the text file.
https://api.github.com/repos/huggingface/datasets/issues/689/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/688
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/688/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/688/comments
https://api.github.com/repos/huggingface/datasets/issues/688/events
https://github.com/huggingface/datasets/pull/688
711,804,828
MDExOlB1bGxSZXF1ZXN0NDk1MzkwMTc1
688
Disable tokenizers parallelism in multiprocessed map
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,459,614,000
1,601,541,946,000
1,601,541,945,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/688", "html_url": "https://github.com/huggingface/datasets/pull/688", "diff_url": "https://github.com/huggingface/datasets/pull/688.diff", "patch_url": "https://github.com/huggingface/datasets/pull/688.patch" }
It was reported in #620 that using multiprocessing with a tokenizers shows this message: ``` The current process just got forked. Disabling parallelism to avoid deadlocks... To disable this warning, please explicitly set TOKENIZERS_PARALLELISM=(true | false) ``` This message is shown when TOKENIZERS_PARALLELISM is unset. Moreover if it is set to `true`, then the program just hangs. To hide the message (if TOKENIZERS_PARALLELISM is unset) and avoid hanging (if TOKENIZERS_PARALLELISM is `true`), then I set TOKENIZERS_PARALLELISM to `false` when forking the process. After forking is gets back to its original value. Also I added a warning if TOKENIZERS_PARALLELISM was `true` and is set to `false`: ``` Setting TOKENIZERS_PARALLELISM=false for forked processes. ``` cc @n1t0
https://api.github.com/repos/huggingface/datasets/issues/688/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/687
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/687/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/687/comments
https://api.github.com/repos/huggingface/datasets/issues/687/events
https://github.com/huggingface/datasets/issues/687
711,664,810
MDU6SXNzdWU3MTE2NjQ4MTA=
687
`ArrowInvalid` occurs while running `Dataset.map()` function
{ "login": "peinan", "id": 5601012, "node_id": "MDQ6VXNlcjU2MDEwMTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5601012?v=4", "gravatar_id": "", "url": "https://api.github.com/users/peinan", "html_url": "https://github.com/peinan", "followers_url": "https://api.github.com/users/peinan/followers", "following_url": "https://api.github.com/users/peinan/following{/other_user}", "gists_url": "https://api.github.com/users/peinan/gists{/gist_id}", "starred_url": "https://api.github.com/users/peinan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/peinan/subscriptions", "organizations_url": "https://api.github.com/users/peinan/orgs", "repos_url": "https://api.github.com/users/peinan/repos", "events_url": "https://api.github.com/users/peinan/events{/privacy}", "received_events_url": "https://api.github.com/users/peinan/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi !\r\n\r\nThis is because `encode` expects one single text as input (str), or one tokenized text (List[str]).\r\nI believe that you actually wanted to use `encode_batch` which expects a batch of texts.\r\nHowever this method is only available for our \"fast\" tokenizers (ex: BertTokenizerFast).\r\nBertJapanese is not one of them unfortunately and I don't think it will be added for now (see https://github.com/huggingface/transformers/pull/7141)...\r\ncc @thomwolf for confirmation.\r\n\r\nTherefore what I'd suggest for now is disable batching and process one text at a time using `encode`.\r\nNote that you can make it faster by using multiprocessing:\r\n\r\n```python\r\nnum_proc = None # Specify here the number of processes if you want to use multiprocessing. ex: num_proc = 4\r\nencoded = train_ds.map(\r\n lambda example: {'tokens': t.encode(example['title'], max_length=1000)}, num_proc=num_proc\r\n)\r\n```\r\n", "Thank you very much for the kind and precise suggestion!\r\nI'm looking forward to seeing BertJapaneseTokenizer built into the \"fast\" tokenizers.\r\n\r\nI tried `map` with multiprocessing as follows, and it worked!\r\n\r\n```python\r\n# There was a Pickle problem if I use `lambda` for multiprocessing\r\ndef encode(examples):\r\n return {'tokens': t.encode(examples['title'], max_length=1000)}\r\n\r\nnum_proc = 8\r\nencoded = train_ds.map(encode, num_proc=num_proc)\r\n```\r\n\r\nThank you again!" ]
1,601,446,610,000
1,601,459,583,000
1,601,459,583,000
NONE
null
null
It seems to fail to process the final batch. This [colab](https://colab.research.google.com/drive/1_byLZRHwGP13PHMkJWo62Wp50S_Z2HMD?usp=sharing) can reproduce the error. Code: ```python # train_ds = Dataset(features: { # 'title': Value(dtype='string', id=None), # 'score': Value(dtype='float64', id=None) # }, num_rows: 99999) # suggested in #665 class PicklableTokenizer(BertJapaneseTokenizer): def __getstate__(self): state = dict(self.__dict__) state['do_lower_case'] = self.word_tokenizer.do_lower_case state['never_split'] = self.word_tokenizer.never_split del state['word_tokenizer'] return state def __setstate(self): do_lower_case = state.pop('do_lower_case') never_split = state.pop('never_split') self.__dict__ = state self.word_tokenizer = MecabTokenizer( do_lower_case=do_lower_case, never_split=never_split ) t = PicklableTokenizer.from_pretrained('bert-base-japanese-whole-word-masking') encoded = train_ds.map( lambda examples: {'tokens': t.encode(examples['title'], max_length=1000)}, batched=True, batch_size=1000 ) ``` Error Message: ``` 99% 99/100 [00:22<00:00, 39.07ba/s] --------------------------------------------------------------------------- ArrowInvalid Traceback (most recent call last) <timed exec> in <module> /usr/local/lib/python3.6/site-packages/datasets/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint) 1242 fn_kwargs=fn_kwargs, 1243 new_fingerprint=new_fingerprint, -> 1244 update_data=update_data, 1245 ) 1246 else: /usr/local/lib/python3.6/site-packages/datasets/arrow_dataset.py in wrapper(*args, **kwargs) 151 "output_all_columns": self._output_all_columns, 152 } --> 153 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 154 if new_format["columns"] is not None: 155 new_format["columns"] = list(set(new_format["columns"]) & set(out.column_names)) /usr/local/lib/python3.6/site-packages/datasets/fingerprint.py in wrapper(*args, **kwargs) 161 # Call actual function 162 --> 163 out = func(self, *args, **kwargs) 164 165 # Update fingerprint of in-place transforms + update in-place history of transforms /usr/local/lib/python3.6/site-packages/datasets/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, update_data) 1496 if update_data: 1497 batch = cast_to_python_objects(batch) -> 1498 writer.write_batch(batch) 1499 if update_data: 1500 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file /usr/local/lib/python3.6/site-packages/datasets/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size) 271 typed_sequence = TypedSequence(batch_examples[col], type=col_type, try_type=col_try_type) 272 typed_sequence_examples[col] = typed_sequence --> 273 pa_table = pa.Table.from_pydict(typed_sequence_examples) 274 self.write_table(pa_table) 275 /usr/local/lib/python3.6/site-packages/pyarrow/table.pxi in pyarrow.lib.Table.from_pydict() /usr/local/lib/python3.6/site-packages/pyarrow/table.pxi in pyarrow.lib.Table.from_arrays() /usr/local/lib/python3.6/site-packages/pyarrow/table.pxi in pyarrow.lib.Table.validate() /usr/local/lib/python3.6/site-packages/pyarrow/error.pxi in pyarrow.lib.check_status() ArrowInvalid: Column 4 named tokens expected length 999 but got length 1000 ```
https://api.github.com/repos/huggingface/datasets/issues/687/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/686
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/686/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/686/comments
https://api.github.com/repos/huggingface/datasets/issues/686/events
https://github.com/huggingface/datasets/issues/686
711,385,739
MDU6SXNzdWU3MTEzODU3Mzk=
686
Dataset browser url is still https://huggingface.co/nlp/viewer/
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists_url": "https://api.github.com/users/jarednielsen/gists{/gist_id}", "starred_url": "https://api.github.com/users/jarednielsen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jarednielsen/subscriptions", "organizations_url": "https://api.github.com/users/jarednielsen/orgs", "repos_url": "https://api.github.com/users/jarednielsen/repos", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "received_events_url": "https://api.github.com/users/jarednielsen/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Yes! might do it with @srush one of these days. Hopefully it won't break too many links (we can always redirect from old url to new)", "This was fixed but forgot to close the issue. cc @lhoestq @yjernite \r\n\r\nThanks @jarednielsen!" ]
1,601,407,312,000
1,610,130,566,000
1,610,130,566,000
CONTRIBUTOR
null
null
Might be worth updating to https://huggingface.co/datasets/viewer/
https://api.github.com/repos/huggingface/datasets/issues/686/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/685
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/685/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/685/comments
https://api.github.com/repos/huggingface/datasets/issues/685/events
https://github.com/huggingface/datasets/pull/685
711,182,185
MDExOlB1bGxSZXF1ZXN0NDk0ODg1NjIz
685
Add features parameter to CSV
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,390,616,000
1,601,455,196,000
1,601,455,194,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/685", "html_url": "https://github.com/huggingface/datasets/pull/685", "diff_url": "https://github.com/huggingface/datasets/pull/685.diff", "patch_url": "https://github.com/huggingface/datasets/pull/685.patch" }
Add support for the `features` parameter when loading a csv dataset: ```python from datasets import load_dataset, Features features = Features({...}) csv_dataset = load_dataset("csv", data_files=["path/to/my/file.csv"], features=features) ``` I added tests to make sure that it is also compatible with the caching system Fix #623
https://api.github.com/repos/huggingface/datasets/issues/685/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/684
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/684/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/684/comments
https://api.github.com/repos/huggingface/datasets/issues/684/events
https://github.com/huggingface/datasets/pull/684
711,080,947
MDExOlB1bGxSZXF1ZXN0NDk0ODA2NjE1
684
Fix column order issue in cast
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,383,753,000
1,601,395,006,000
1,601,395,005,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/684", "html_url": "https://github.com/huggingface/datasets/pull/684", "diff_url": "https://github.com/huggingface/datasets/pull/684.diff", "patch_url": "https://github.com/huggingface/datasets/pull/684.patch" }
Previously, the order of the columns in the features passes to `cast_` mattered. However even though features passed to `cast_` had the same order as the dataset features, it could fail because the schema that was built was always in alphabetical order. This issue was reported by @lewtun in #623 To fix that I fixed the schema to follow the order of the arrow table columns. I also added the possibility to give features that are not ordered the same way as the dataset features.
https://api.github.com/repos/huggingface/datasets/issues/684/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/683
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/683/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/683/comments
https://api.github.com/repos/huggingface/datasets/issues/683/events
https://github.com/huggingface/datasets/pull/683
710,942,704
MDExOlB1bGxSZXF1ZXN0NDk0NzAwNzY1
683
Fix wrong delimiter in text dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,372,604,000
1,620,239,071,000
1,601,372,646,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/683", "html_url": "https://github.com/huggingface/datasets/pull/683", "diff_url": "https://github.com/huggingface/datasets/pull/683.diff", "patch_url": "https://github.com/huggingface/datasets/pull/683.patch" }
The delimiter is set to the bell character as it is used nowhere is text files usually. However in the text dataset the delimiter was set to `\b` which is backspace in python, while the bell character is `\a`. I replace \b by \a Hopefully it fixes issues mentioned by some users in #622
https://api.github.com/repos/huggingface/datasets/issues/683/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/682
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/682/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/682/comments
https://api.github.com/repos/huggingface/datasets/issues/682/events
https://github.com/huggingface/datasets/pull/682
710,325,399
MDExOlB1bGxSZXF1ZXN0NDk0MTkzMzEw
682
Update navbar chapter titles color
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,303,717,000
1,601,314,213,000
1,601,314,212,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/682", "html_url": "https://github.com/huggingface/datasets/pull/682", "diff_url": "https://github.com/huggingface/datasets/pull/682.diff", "patch_url": "https://github.com/huggingface/datasets/pull/682.patch" }
Consistency with the color change that was done in transformers at https://github.com/huggingface/transformers/pull/7423 It makes the background-color of the chapter titles in the docs navbar darker, to differentiate them from the inner sections. see changes [here](https://691-250213286-gh.circle-artifacts.com/0/docs/_build/html/index.html)
https://api.github.com/repos/huggingface/datasets/issues/682/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/681
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/681/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/681/comments
https://api.github.com/repos/huggingface/datasets/issues/681/events
https://github.com/huggingface/datasets/pull/681
710,075,721
MDExOlB1bGxSZXF1ZXN0NDkzOTkwMjEz
681
Adding missing @property (+2 small flake8 fixes).
{ "login": "Narsil", "id": 204321, "node_id": "MDQ6VXNlcjIwNDMyMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Narsil", "html_url": "https://github.com/Narsil", "followers_url": "https://api.github.com/users/Narsil/followers", "following_url": "https://api.github.com/users/Narsil/following{/other_user}", "gists_url": "https://api.github.com/users/Narsil/gists{/gist_id}", "starred_url": "https://api.github.com/users/Narsil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Narsil/subscriptions", "organizations_url": "https://api.github.com/users/Narsil/orgs", "repos_url": "https://api.github.com/users/Narsil/repos", "events_url": "https://api.github.com/users/Narsil/events{/privacy}", "received_events_url": "https://api.github.com/users/Narsil/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,283,233,000
1,601,288,773,000
1,601,288,769,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/681", "html_url": "https://github.com/huggingface/datasets/pull/681", "diff_url": "https://github.com/huggingface/datasets/pull/681.diff", "patch_url": "https://github.com/huggingface/datasets/pull/681.patch" }
Fixes #678
https://api.github.com/repos/huggingface/datasets/issues/681/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/680
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/680/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/680/comments
https://api.github.com/repos/huggingface/datasets/issues/680/events
https://github.com/huggingface/datasets/pull/680
710,066,138
MDExOlB1bGxSZXF1ZXN0NDkzOTgyMjY4
680
Fix bug related to boolean in GAP dataset.
{ "login": "otakumesi", "id": 14996977, "node_id": "MDQ6VXNlcjE0OTk2OTc3", "avatar_url": "https://avatars.githubusercontent.com/u/14996977?v=4", "gravatar_id": "", "url": "https://api.github.com/users/otakumesi", "html_url": "https://github.com/otakumesi", "followers_url": "https://api.github.com/users/otakumesi/followers", "following_url": "https://api.github.com/users/otakumesi/following{/other_user}", "gists_url": "https://api.github.com/users/otakumesi/gists{/gist_id}", "starred_url": "https://api.github.com/users/otakumesi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/otakumesi/subscriptions", "organizations_url": "https://api.github.com/users/otakumesi/orgs", "repos_url": "https://api.github.com/users/otakumesi/repos", "events_url": "https://api.github.com/users/otakumesi/events{/privacy}", "received_events_url": "https://api.github.com/users/otakumesi/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi !\r\n\r\nGood catch, thanks for creating this PR :)\r\n\r\nCould you also regenerate the metadata for this dataset using \r\n```\r\ndatasets-cli test ./datasets/gap --save_infos --all_configs\r\n```\r\n\r\nThat'd be awesome", "@lhoestq Thank you for your revieing!!!\r\n\r\nI've performed it and have read CONTRIBUTING.md now!" ]
1,601,282,379,000
1,601,394,887,000
1,601,394,887,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/680", "html_url": "https://github.com/huggingface/datasets/pull/680", "diff_url": "https://github.com/huggingface/datasets/pull/680.diff", "patch_url": "https://github.com/huggingface/datasets/pull/680.patch" }
### Why I did The value in `row["A-coref"]` and `row["B-coref"]` is `'TRUE'` or `'FALSE'`. This type is `string`, then `bool('FALSE')` is equal to `True` in Python. So, both rows are transformed into `True` now. So, I modified this problem. ### What I did I modified `bool(row["A-coref"])` and `bool(row["B-coref"])` to `row["A-coref"] == "TRUE"` and `row["B-coref"] == "TRUE"`. Thank you!
https://api.github.com/repos/huggingface/datasets/issues/680/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/679
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/679/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/679/comments
https://api.github.com/repos/huggingface/datasets/issues/679/events
https://github.com/huggingface/datasets/pull/679
710,065,838
MDExOlB1bGxSZXF1ZXN0NDkzOTgyMDMx
679
Fix negative ids when slicing with an array
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,282,348,000
1,601,304,140,000
1,601,304,139,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/679", "html_url": "https://github.com/huggingface/datasets/pull/679", "diff_url": "https://github.com/huggingface/datasets/pull/679.diff", "patch_url": "https://github.com/huggingface/datasets/pull/679.patch" }
```python from datasets import Dataset d = ds.Dataset.from_dict({"a": range(10)}) print(d[[0, -1]]) # OverflowError ``` raises an error because of the negative id. This PR fixes that. Fix #668
https://api.github.com/repos/huggingface/datasets/issues/679/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/678
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/678/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/678/comments
https://api.github.com/repos/huggingface/datasets/issues/678/events
https://github.com/huggingface/datasets/issues/678
710,060,497
MDU6SXNzdWU3MTAwNjA0OTc=
678
The download instructions for c4 datasets are not contained in the error message
{ "login": "Narsil", "id": 204321, "node_id": "MDQ6VXNlcjIwNDMyMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Narsil", "html_url": "https://github.com/Narsil", "followers_url": "https://api.github.com/users/Narsil/followers", "following_url": "https://api.github.com/users/Narsil/following{/other_user}", "gists_url": "https://api.github.com/users/Narsil/gists{/gist_id}", "starred_url": "https://api.github.com/users/Narsil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Narsil/subscriptions", "organizations_url": "https://api.github.com/users/Narsil/orgs", "repos_url": "https://api.github.com/users/Narsil/repos", "events_url": "https://api.github.com/users/Narsil/events{/privacy}", "received_events_url": "https://api.github.com/users/Narsil/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Good catch !\r\nIndeed the `@property` is missing.\r\n\r\nFeel free to open a PR :)", "Also not that C4 is a dataset that needs an Apache Beam runtime to be generated.\r\nFor example Dataflow, Spark, Flink etc.\r\n\r\nUsually we generate the dataset on our side once and for all, but we haven't done it for C4 yet.\r\nMore info about beam datasets [here](https://huggingface.co/docs/datasets/beam_dataset.html)\r\n\r\nLet me know if you have any questions" ]
1,601,281,854,000
1,601,288,769,000
1,601,288,769,000
CONTRIBUTOR
null
null
The manual download instructions are not clear ```The dataset c4 with config en requires manual data. Please follow the manual download instructions: <bound method C4.manual_download_instructions of <datasets_modules.datasets.c4.830b0c218bd41fed439812c8dd19dbd4767d2a3faa385eb695cf8666c982b1b3.c4.C4 object at 0x7ff8c5969760>>. Manual data can be loaded with `datasets.load_dataset(c4, data_dir='<path/to/manual/data>') ``` Either `@property` could be added to C4.manual_download_instrcutions (or make it a real property), or the manual_download_instructions function needs to be called I think. Let me know if you want a PR for this, but I'm not sure which possible fix is the correct one.
https://api.github.com/repos/huggingface/datasets/issues/678/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/677
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/677/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/677/comments
https://api.github.com/repos/huggingface/datasets/issues/677/events
https://github.com/huggingface/datasets/pull/677
710,055,239
MDExOlB1bGxSZXF1ZXN0NDkzOTczNDE3
677
Move cache dir root creation in builder's init
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,281,366,000
1,601,304,163,000
1,601,304,162,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/677", "html_url": "https://github.com/huggingface/datasets/pull/677", "diff_url": "https://github.com/huggingface/datasets/pull/677.diff", "patch_url": "https://github.com/huggingface/datasets/pull/677.patch" }
We use lock files in the builder initialization but sometimes the cache directory where they're supposed to be was not created. To fix that I moved the builder's cache dir root creation in the builder's init. Fix #671
https://api.github.com/repos/huggingface/datasets/issues/677/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/676
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/676/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/676/comments
https://api.github.com/repos/huggingface/datasets/issues/676/events
https://github.com/huggingface/datasets/issues/676
710,014,319
MDU6SXNzdWU3MTAwMTQzMTk=
676
train_test_split returns empty dataset item
{ "login": "mojave-pku", "id": 26648528, "node_id": "MDQ6VXNlcjI2NjQ4NTI4", "avatar_url": "https://avatars.githubusercontent.com/u/26648528?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mojave-pku", "html_url": "https://github.com/mojave-pku", "followers_url": "https://api.github.com/users/mojave-pku/followers", "following_url": "https://api.github.com/users/mojave-pku/following{/other_user}", "gists_url": "https://api.github.com/users/mojave-pku/gists{/gist_id}", "starred_url": "https://api.github.com/users/mojave-pku/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mojave-pku/subscriptions", "organizations_url": "https://api.github.com/users/mojave-pku/orgs", "repos_url": "https://api.github.com/users/mojave-pku/repos", "events_url": "https://api.github.com/users/mojave-pku/events{/privacy}", "received_events_url": "https://api.github.com/users/mojave-pku/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "The problem still exists after removing the cache files.", "Can you reproduce this example in a Colab so we can investigate? (or give more information on your software/hardware config)", "Thanks for reporting.\r\nI just found the issue, I'm creating a PR", "We'll do a release pretty soon to include the fix :)\r\nIn the meantime you can install the lib from source if you want to " ]
1,601,277,573,000
1,602,078,393,000
1,602,077,886,000
NONE
null
null
I try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty. The codes: ``` yelp_data = datasets.load_from_disk('/home/ssd4/huanglianzhe/test_yelp') print(yelp_data[0]) yelp_data = yelp_data.train_test_split(test_size=0.1) print(yelp_data) print(yelp_data['test']) print(yelp_data['test'][0]) ``` The outputs: ``` {'stars': 2.0, 'text': 'xxxx'} Loading cached split indices for dataset at /home/ssd4/huanglianzhe/test_yelp/cache-f9b22d8b9d5a7346.arrow and /home/ssd4/huanglianzhe/test_yelp/cache-4aa26fa4005059d1.arrow DatasetDict({'train': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 7219009), 'test': Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113)}) Dataset(features: {'stars': Value(dtype='float64', id=None), 'text': Value(dtype='string', id=None)}, num_rows: 802113) {} # yelp_data['test'][0] is empty ```
https://api.github.com/repos/huggingface/datasets/issues/676/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/675
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/675/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/675/comments
https://api.github.com/repos/huggingface/datasets/issues/675/events
https://github.com/huggingface/datasets/issues/675
709,818,725
MDU6SXNzdWU3MDk4MTg3MjU=
675
Add custom dataset to NLP?
{ "login": "timpal0l", "id": 6556710, "node_id": "MDQ6VXNlcjY1NTY3MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/6556710?v=4", "gravatar_id": "", "url": "https://api.github.com/users/timpal0l", "html_url": "https://github.com/timpal0l", "followers_url": "https://api.github.com/users/timpal0l/followers", "following_url": "https://api.github.com/users/timpal0l/following{/other_user}", "gists_url": "https://api.github.com/users/timpal0l/gists{/gist_id}", "starred_url": "https://api.github.com/users/timpal0l/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/timpal0l/subscriptions", "organizations_url": "https://api.github.com/users/timpal0l/orgs", "repos_url": "https://api.github.com/users/timpal0l/repos", "events_url": "https://api.github.com/users/timpal0l/events{/privacy}", "received_events_url": "https://api.github.com/users/timpal0l/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Yes you can have a look here: https://huggingface.co/docs/datasets/loading_datasets.html#csv-files", "No activity, closing" ]
1,601,241,770,000
1,603,184,929,000
1,603,184,929,000
CONTRIBUTOR
null
null
Is it possible to add a custom dataset such as a .csv to the NLP library? Thanks.
https://api.github.com/repos/huggingface/datasets/issues/675/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/674
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/674/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/674/comments
https://api.github.com/repos/huggingface/datasets/issues/674/events
https://github.com/huggingface/datasets/issues/674
709,661,006
MDU6SXNzdWU3MDk2NjEwMDY=
674
load_dataset() won't download in Windows
{ "login": "ThisDavehead", "id": 34422661, "node_id": "MDQ6VXNlcjM0NDIyNjYx", "avatar_url": "https://avatars.githubusercontent.com/u/34422661?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ThisDavehead", "html_url": "https://github.com/ThisDavehead", "followers_url": "https://api.github.com/users/ThisDavehead/followers", "following_url": "https://api.github.com/users/ThisDavehead/following{/other_user}", "gists_url": "https://api.github.com/users/ThisDavehead/gists{/gist_id}", "starred_url": "https://api.github.com/users/ThisDavehead/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ThisDavehead/subscriptions", "organizations_url": "https://api.github.com/users/ThisDavehead/orgs", "repos_url": "https://api.github.com/users/ThisDavehead/repos", "events_url": "https://api.github.com/users/ThisDavehead/events{/privacy}", "received_events_url": "https://api.github.com/users/ThisDavehead/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I have the same issue. Tried to download a few of them and not a single one is downloaded successfully.\r\n\r\nThis is the output:\r\n```\r\n>>> dataset = load_dataset('blended_skill_talk', split='train')\r\nUsing custom data configuration default <-- This step never ends\r\n```", "This was fixed in #644 \r\nI'll do a new release soon :)\r\n\r\nIn the meantime you can run it by installing from source", "Closing since version 1.1.0 got released with Windows support :) \r\nLet me know if it works for you now" ]
1,601,178,985,000
1,601,886,498,000
1,601,886,498,000
NONE
null
null
I don't know if this is just me or Windows. Maybe other Windows users can chime in if they don't have this problem. I've been trying to get some of the tutorials working on Windows, but when I use the load_dataset() function, it just stalls and the script keeps running indefinitely without downloading anything. I've waited upwards of 18 hours to download the 'multi-news' dataset (which isn't very big), and still nothing. I've tried running it through different IDE's and the command line, but it had the same behavior. I've also tried it with all virus and malware protection turned off. I've made sure python and all IDE's are exceptions to the firewall and all the requisite permissions are enabled. Additionally, I checked to see if other packages could download content such as an nltk corpus, and they could. I've also run the same script using Ubuntu and it downloaded fine (and quickly). When I copied the downloaded datasets from my Ubuntu drive to my Windows .cache folder it worked fine by reusing the already-downloaded dataset, but it's cumbersome to do that for every dataset I want to try in my Windows environment. Could this be a bug, or is there something I'm doing wrong or not thinking of? Thanks.
https://api.github.com/repos/huggingface/datasets/issues/674/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/673
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/673/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/673/comments
https://api.github.com/repos/huggingface/datasets/issues/673/events
https://github.com/huggingface/datasets/issues/673
709,603,989
MDU6SXNzdWU3MDk2MDM5ODk=
673
blog_authorship_corpus crashed
{ "login": "Moshiii", "id": 7553188, "node_id": "MDQ6VXNlcjc1NTMxODg=", "avatar_url": "https://avatars.githubusercontent.com/u/7553188?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Moshiii", "html_url": "https://github.com/Moshiii", "followers_url": "https://api.github.com/users/Moshiii/followers", "following_url": "https://api.github.com/users/Moshiii/following{/other_user}", "gists_url": "https://api.github.com/users/Moshiii/gists{/gist_id}", "starred_url": "https://api.github.com/users/Moshiii/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Moshiii/subscriptions", "organizations_url": "https://api.github.com/users/Moshiii/orgs", "repos_url": "https://api.github.com/users/Moshiii/repos", "events_url": "https://api.github.com/users/Moshiii/events{/privacy}", "received_events_url": "https://api.github.com/users/Moshiii/received_events", "type": "User", "site_admin": false }
[ { "id": 2107841032, "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer", "name": "nlp-viewer", "color": "94203D", "default": false, "description": "" } ]
open
false
null
[]
null
[ "Thanks for reporting !\r\nWe'll free some memory" ]
1,601,151,328,000
1,601,280,290,000
null
NONE
null
null
This is just to report that When I pick blog_authorship_corpus in https://huggingface.co/nlp/viewer/?dataset=blog_authorship_corpus I get this: ![image](https://user-images.githubusercontent.com/7553188/94349542-4364f300-0013-11eb-897d-b25660a449f0.png)
https://api.github.com/repos/huggingface/datasets/issues/673/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/672
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/672/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/672/comments
https://api.github.com/repos/huggingface/datasets/issues/672/events
https://github.com/huggingface/datasets/issues/672
709,575,527
MDU6SXNzdWU3MDk1NzU1Mjc=
672
Questions about XSUM
{ "login": "danyaljj", "id": 2441454, "node_id": "MDQ6VXNlcjI0NDE0NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2441454?v=4", "gravatar_id": "", "url": "https://api.github.com/users/danyaljj", "html_url": "https://github.com/danyaljj", "followers_url": "https://api.github.com/users/danyaljj/followers", "following_url": "https://api.github.com/users/danyaljj/following{/other_user}", "gists_url": "https://api.github.com/users/danyaljj/gists{/gist_id}", "starred_url": "https://api.github.com/users/danyaljj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/danyaljj/subscriptions", "organizations_url": "https://api.github.com/users/danyaljj/orgs", "repos_url": "https://api.github.com/users/danyaljj/repos", "events_url": "https://api.github.com/users/danyaljj/events{/privacy}", "received_events_url": "https://api.github.com/users/danyaljj/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "We should try to regenerate the data using the official script.\r\nBut iirc that's what we used in the first place, so not sure why it didn't match in the first place.\r\n\r\nI'll let you know when the dataset is updated", "Thanks, looking forward to hearing your update on this thread. \r\n\r\nThis is a blocking issue for us; would appreciate any progress on this front. We can also help with the fix, if you deem it appropriately. ", "I just started the generation on my side, I'll let you know how it goes :) ", "Hmm after a first run I'm still missing 136668/226711 urls.\r\nI'll relaunch it tomorrow to try to get the remaining ones.", "Update: I'm missing 36/226711 urls but I haven't managed to download them yet", "Thanks! That sounds like a reasonable number! ", "So I managed to download them all but when parsing only 226,181/226,711 worked.\r\nNot sure if it's worth digging and debugging parsing at this point :/ ", "Maybe @sshleifer can help, I think he's already played with xsum at one point", "Thanks @lhoestq\r\nIt would be great to improve coverage, but IDs are the really crucial part for us. We'd really appreciate an update to the dataset with IDs either way!", "I gave up at an even earlier point. The dataset I use has 204,017 train examples.", "@lhoestq @sshleifer like @jbragg said earlier, the main issue for us is that the current XSUM dataset (in your package) does not have IDs suggested by the original dataset ([here is the file](https://raw.githubusercontent.com/EdinburghNLP/XSum/master/XSum-Dataset/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json).) Would appreciate if you update the XSUM dataset to include the instance IDs. \r\n\r\nThe missing instances is also a problem, but likely not worth pursuing given its relatively small scale. ", ">So I managed to download them all but when parsing only 226,181/226,711 worked.\r\n\r\n@lhoestq any chance we could update the HF-hosted dataset with the IDs in your new version? Happy to help if there's something I can do.", "Well I couldn't parse what I downloaded.\r\nUnfortunately I think I won't be able to take a look at it this week.\r\nI can try to send you what I got if you want to give it a shot @jbragg \r\nOtherwise feel free to re-run the xsum download script, maybe you'll be luckier than me", "Resolved via #754" ]
1,601,140,584,000
1,603,185,367,000
null
CONTRIBUTOR
null
null
Hi there ✋ I'm looking into your `xsum` dataset and I have several questions on that. So here is how I loaded the data: ``` >>> data = datasets.load_dataset('xsum', version='1.0.1') >>> data['train'] Dataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 204017) >>> data['test'] Dataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, num_rows: 11333) ``` The first issue is, the instance counts don’t match what I see on [the dataset's website](https://github.com/EdinburghNLP/XSum/tree/master/XSum-Dataset#what-builds-the-xsum-dataset) (11,333 vs 11,334 for test set; 204,017 vs 204,045 for training set) ``` … training (90%, 204,045), validation (5%, 11,332), and test (5%, 11,334) set. ``` Any thoughts why? Perhaps @mariamabarham could help here, since she recently had a PR on this dataaset https://github.com/huggingface/datasets/pull/289 (reviewed by @patrickvonplaten) Another issue is that the instances don't seem to have IDs. The original datasets provides IDs for the instances: https://github.com/EdinburghNLP/XSum/blob/master/XSum-Dataset/XSum-TRAINING-DEV-TEST-SPLIT-90-5-5.json but to be able to use them, the dataset sizes need to match. CC @jbragg
https://api.github.com/repos/huggingface/datasets/issues/672/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/671
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/671/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/671/comments
https://api.github.com/repos/huggingface/datasets/issues/671/events
https://github.com/huggingface/datasets/issues/671
709,093,151
MDU6SXNzdWU3MDkwOTMxNTE=
671
[BUG] No such file or directory
{ "login": "jbragg", "id": 2238344, "node_id": "MDQ6VXNlcjIyMzgzNDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2238344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jbragg", "html_url": "https://github.com/jbragg", "followers_url": "https://api.github.com/users/jbragg/followers", "following_url": "https://api.github.com/users/jbragg/following{/other_user}", "gists_url": "https://api.github.com/users/jbragg/gists{/gist_id}", "starred_url": "https://api.github.com/users/jbragg/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jbragg/subscriptions", "organizations_url": "https://api.github.com/users/jbragg/orgs", "repos_url": "https://api.github.com/users/jbragg/repos", "events_url": "https://api.github.com/users/jbragg/events{/privacy}", "received_events_url": "https://api.github.com/users/jbragg/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,051,934,000
1,601,304,162,000
1,601,304,162,000
CONTRIBUTOR
null
null
This happens when both 1. Huggingface datasets cache dir does not exist 2. Try to load a local dataset script builder.py throws an error when trying to create a filelock in a directory (cache/datasets) that does not exist https://github.com/huggingface/datasets/blob/master/src/datasets/builder.py#L177 Tested on v1.0.2 @lhoestq
https://api.github.com/repos/huggingface/datasets/issues/671/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/670
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/670/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/670/comments
https://api.github.com/repos/huggingface/datasets/issues/670/events
https://github.com/huggingface/datasets/pull/670
709,061,231
MDExOlB1bGxSZXF1ZXN0NDkzMTc4OTQw
670
Fix SQuAD metric kwargs description
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,601,050,137,000
1,601,395,059,000
1,601,395,058,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/670", "html_url": "https://github.com/huggingface/datasets/pull/670", "diff_url": "https://github.com/huggingface/datasets/pull/670.diff", "patch_url": "https://github.com/huggingface/datasets/pull/670.patch" }
The `answer_start` field was missing in the kwargs docstring. This should fix #657 FYI another fix was proposed by @tshrjn in #658 and suggests to remove this field. However IMO `answer_start` is useful to match the squad dataset format for consistency, even though it is not used in the metric computation. I think it's better to keep it this way, so that you can just give references=squad["answers"] to .compute(). Let me know what sounds the best for you
https://api.github.com/repos/huggingface/datasets/issues/670/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/669
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/669/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/669/comments
https://api.github.com/repos/huggingface/datasets/issues/669/events
https://github.com/huggingface/datasets/issues/669
708,857,595
MDU6SXNzdWU3MDg4NTc1OTU=
669
How to skip a example when running dataset.map
{ "login": "xixiaoyao", "id": 24541791, "node_id": "MDQ6VXNlcjI0NTQxNzkx", "avatar_url": "https://avatars.githubusercontent.com/u/24541791?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xixiaoyao", "html_url": "https://github.com/xixiaoyao", "followers_url": "https://api.github.com/users/xixiaoyao/followers", "following_url": "https://api.github.com/users/xixiaoyao/following{/other_user}", "gists_url": "https://api.github.com/users/xixiaoyao/gists{/gist_id}", "starred_url": "https://api.github.com/users/xixiaoyao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xixiaoyao/subscriptions", "organizations_url": "https://api.github.com/users/xixiaoyao/orgs", "repos_url": "https://api.github.com/users/xixiaoyao/repos", "events_url": "https://api.github.com/users/xixiaoyao/events{/privacy}", "received_events_url": "https://api.github.com/users/xixiaoyao/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi @xixiaoyao,\r\nDepending on what you want to do you can:\r\n- use a first step of `filter` to filter out the invalid examples: https://huggingface.co/docs/datasets/processing.html#filtering-rows-select-and-filter\r\n- or directly detect the invalid examples inside the callable used with `map` and return them unchanged or even remove them at the same time if you are using `map` in batched mode. Here is an example where we use `map` in batched mode to add new rows on the fly but you can also use it to remove examples on the fly (that's what `filter` actually do under-the-hood): https://huggingface.co/docs/datasets/processing.html#augmenting-the-dataset", "Closing this one.\r\nFeel free to re-open if you have other questions", "Letting finders-of-this-thread know that the new link is: https://huggingface.co/docs/datasets/process#data-augmentation\r\n" ]
1,601,032,673,000
1,601,915,293,000
1,601,915,293,000
NONE
null
null
in processing func, I process examples and detect some invalid examples, which I did not want it to be added into train dataset. However I did not find how to skip this recognized invalid example when doing dataset.map.
https://api.github.com/repos/huggingface/datasets/issues/669/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/668
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/668/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/668/comments
https://api.github.com/repos/huggingface/datasets/issues/668/events
https://github.com/huggingface/datasets/issues/668
708,310,956
MDU6SXNzdWU3MDgzMTA5NTY=
668
OverflowError when slicing with an array containing negative ids
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,964,834,000
1,601,304,139,000
1,601,304,139,000
MEMBER
null
null
```python from datasets import Dataset d = ds.Dataset.from_dict({"a": range(10)}) print(d[0]) # {'a': 0} print(d[-1]) # {'a': 9} print(d[[0, -1]]) # OverflowError ``` results in ``` --------------------------------------------------------------------------- OverflowError Traceback (most recent call last) <ipython-input-5-863dc3555598> in <module> ----> 1 d[[0, -1]] ~/Desktop/hf/nlp/src/datasets/arrow_dataset.py in __getitem__(self, key) 1070 format_columns=self._format_columns, 1071 output_all_columns=self._output_all_columns, -> 1072 format_kwargs=self._format_kwargs, 1073 ) 1074 ~/Desktop/hf/nlp/src/datasets/arrow_dataset.py in _getitem(self, key, format_type, format_columns, output_all_columns, format_kwargs) 1025 indices = key 1026 -> 1027 indices_array = pa.array([int(i) for i in indices], type=pa.uint64()) 1028 1029 # Check if we need to convert indices ~/.virtualenvs/hf-datasets/lib/python3.7/site-packages/pyarrow/array.pxi in pyarrow.lib.array() ~/.virtualenvs/hf-datasets/lib/python3.7/site-packages/pyarrow/array.pxi in pyarrow.lib._sequence_to_array() OverflowError: can't convert negative value to unsigned int ```
https://api.github.com/repos/huggingface/datasets/issues/668/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/667
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/667/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/667/comments
https://api.github.com/repos/huggingface/datasets/issues/667/events
https://github.com/huggingface/datasets/issues/667
708,258,392
MDU6SXNzdWU3MDgyNTgzOTI=
667
Loss not decrease with Datasets and Transformers
{ "login": "wangcongcong123", "id": 23032865, "node_id": "MDQ6VXNlcjIzMDMyODY1", "avatar_url": "https://avatars.githubusercontent.com/u/23032865?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wangcongcong123", "html_url": "https://github.com/wangcongcong123", "followers_url": "https://api.github.com/users/wangcongcong123/followers", "following_url": "https://api.github.com/users/wangcongcong123/following{/other_user}", "gists_url": "https://api.github.com/users/wangcongcong123/gists{/gist_id}", "starred_url": "https://api.github.com/users/wangcongcong123/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wangcongcong123/subscriptions", "organizations_url": "https://api.github.com/users/wangcongcong123/orgs", "repos_url": "https://api.github.com/users/wangcongcong123/repos", "events_url": "https://api.github.com/users/wangcongcong123/events{/privacy}", "received_events_url": "https://api.github.com/users/wangcongcong123/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "And I tested it on T5ForConditionalGeneration, that works no problem.", "Hi did you manage to fix your issue ?\r\n\r\nIf so feel free to share your fix and close this thread" ]
1,600,960,483,000
1,609,531,285,000
1,609,531,285,000
NONE
null
null
HI, The following script is used to fine-tune a BertForSequenceClassification model on SST2. The script is adapted from [this colab](https://colab.research.google.com/github/huggingface/datasets/blob/master/notebooks/Overview.ipynb) that presents an example of fine-tuning BertForQuestionAnswering using squad dataset. In that colab, loss works fine. When I adapt it to SST2, the loss fails to decrease as it should. I attach the adapted script below and appreciate anyone pointing out what I miss? ```python import torch from datasets import load_dataset from transformers import BertForSequenceClassification from transformers import BertTokenizerFast # Load our training dataset and tokenizer dataset = load_dataset("glue", 'sst2') tokenizer = BertTokenizerFast.from_pretrained('bert-base-cased') del dataset["test"] # let's remove it in this demo # Tokenize our training dataset def convert_to_features(example_batch): encodings = tokenizer(example_batch["sentence"]) encodings.update({"labels": example_batch["label"]}) return encodings encoded_dataset = dataset.map(convert_to_features, batched=True) # Format our dataset to outputs torch.Tensor to train a pytorch model columns = ['input_ids', 'token_type_ids', 'attention_mask', 'labels'] encoded_dataset.set_format(type='torch', columns=columns) # Instantiate a PyTorch Dataloader around our dataset # Let's do dynamic batching (pad on the fly with our own collate_fn) def collate_fn(examples): return tokenizer.pad(examples, return_tensors='pt') dataloader = torch.utils.data.DataLoader(encoded_dataset['train'], collate_fn=collate_fn, batch_size=8) # Now let's train our model device = 'cuda' if torch.cuda.is_available() else 'cpu' # Let's load a pretrained Bert model and a simple optimizer model = BertForSequenceClassification.from_pretrained('bert-base-cased', return_dict=True) optimizer = torch.optim.Adam(model.parameters(), lr=1e-5) model.train().to(device) for i, batch in enumerate(dataloader): batch.to(device) outputs = model(**batch) loss = outputs.loss loss.backward() optimizer.step() model.zero_grad() print(f'Step {i} - loss: {loss:.3}') ``` In case needed. - datasets == 1.0.2 - transformers == 3.2.0
https://api.github.com/repos/huggingface/datasets/issues/667/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/666
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/666/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/666/comments
https://api.github.com/repos/huggingface/datasets/issues/666/events
https://github.com/huggingface/datasets/issues/666
707,608,578
MDU6SXNzdWU3MDc2MDg1Nzg=
666
Does both 'bookcorpus' and 'wikipedia' belong to the same datasets which Google used for pretraining BERT?
{ "login": "wahab4114", "id": 31090427, "node_id": "MDQ6VXNlcjMxMDkwNDI3", "avatar_url": "https://avatars.githubusercontent.com/u/31090427?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wahab4114", "html_url": "https://github.com/wahab4114", "followers_url": "https://api.github.com/users/wahab4114/followers", "following_url": "https://api.github.com/users/wahab4114/following{/other_user}", "gists_url": "https://api.github.com/users/wahab4114/gists{/gist_id}", "starred_url": "https://api.github.com/users/wahab4114/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wahab4114/subscriptions", "organizations_url": "https://api.github.com/users/wahab4114/orgs", "repos_url": "https://api.github.com/users/wahab4114/repos", "events_url": "https://api.github.com/users/wahab4114/events{/privacy}", "received_events_url": "https://api.github.com/users/wahab4114/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "No they are other similar copies but they are not provided by the official Bert models authors." ]
1,600,887,745,000
1,603,811,965,000
1,603,811,965,000
NONE
null
null
https://api.github.com/repos/huggingface/datasets/issues/666/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/665
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/665/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/665/comments
https://api.github.com/repos/huggingface/datasets/issues/665/events
https://github.com/huggingface/datasets/issues/665
707,037,738
MDU6SXNzdWU3MDcwMzc3Mzg=
665
runing dataset.map, it raises TypeError: can't pickle Tokenizer objects
{ "login": "xixiaoyao", "id": 24541791, "node_id": "MDQ6VXNlcjI0NTQxNzkx", "avatar_url": "https://avatars.githubusercontent.com/u/24541791?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xixiaoyao", "html_url": "https://github.com/xixiaoyao", "followers_url": "https://api.github.com/users/xixiaoyao/followers", "following_url": "https://api.github.com/users/xixiaoyao/following{/other_user}", "gists_url": "https://api.github.com/users/xixiaoyao/gists{/gist_id}", "starred_url": "https://api.github.com/users/xixiaoyao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xixiaoyao/subscriptions", "organizations_url": "https://api.github.com/users/xixiaoyao/orgs", "repos_url": "https://api.github.com/users/xixiaoyao/repos", "events_url": "https://api.github.com/users/xixiaoyao/events{/privacy}", "received_events_url": "https://api.github.com/users/xixiaoyao/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi !\r\nIt works on my side with both the LongFormerTokenizer and the LongFormerTokenizerFast.\r\n\r\nWhich version of transformers/datasets are you using ?", "transformers and datasets are both the latest", "Then I guess you need to give us more informations on your setup (OS, python, GPU, etc) or a Google Colab reproducing the error for us to be able to debug this error.", "And your version of `dill` if possible :)", "I have the same issue with `transformers/BertJapaneseTokenizer`.\r\n\r\n\r\n\r\n```python\r\n# train_ds = Dataset(features: {\r\n# 'title': Value(dtype='string', id=None), \r\n# 'score': Value(dtype='float64', id=None)\r\n# }, num_rows: 99999)\r\n\r\nt = BertJapaneseTokenizer.from_pretrained('bert-base-japanese-whole-word-masking')\r\nencoded = train_ds.map(lambda examples: {'tokens': t.encode(examples['title'])}, batched=True)\r\n```\r\n\r\n<details><summary>Error Message</summary>\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n<ipython-input-35-2b7d66b291c1> in <module>\r\n 2 \r\n 3 encoded = train_ds.map(lambda examples:\r\n----> 4 {'tokens': t.encode(examples['title'])}, batched=True)\r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)\r\n 1242 fn_kwargs=fn_kwargs,\r\n 1243 new_fingerprint=new_fingerprint,\r\n-> 1244 update_data=update_data,\r\n 1245 )\r\n 1246 else:\r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/arrow_dataset.py in wrapper(*args, **kwargs)\r\n 151 \"output_all_columns\": self._output_all_columns,\r\n 152 }\r\n--> 153 out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n 154 if new_format[\"columns\"] is not None:\r\n 155 new_format[\"columns\"] = list(set(new_format[\"columns\"]) & set(out.column_names))\r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/fingerprint.py in wrapper(*args, **kwargs)\r\n 156 kwargs_for_fingerprint[\"fingerprint_name\"] = fingerprint_name\r\n 157 kwargs[fingerprint_name] = update_fingerprint(\r\n--> 158 self._fingerprint, transform, kwargs_for_fingerprint\r\n 159 )\r\n 160 \r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args)\r\n 103 for key in sorted(transform_args):\r\n 104 hasher.update(key)\r\n--> 105 hasher.update(transform_args[key])\r\n 106 return hasher.hexdigest()\r\n 107 \r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/fingerprint.py in update(self, value)\r\n 55 def update(self, value):\r\n 56 self.m.update(f\"=={type(value)}==\".encode(\"utf8\"))\r\n---> 57 self.m.update(self.hash(value).encode(\"utf-8\"))\r\n 58 \r\n 59 def hexdigest(self):\r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/fingerprint.py in hash(cls, value)\r\n 51 return cls.dispatch[type(value)](cls, value)\r\n 52 else:\r\n---> 53 return cls.hash_default(value)\r\n 54 \r\n 55 def update(self, value):\r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/fingerprint.py in hash_default(cls, value)\r\n 44 @classmethod\r\n 45 def hash_default(cls, value):\r\n---> 46 return cls.hash_bytes(dumps(value))\r\n 47 \r\n 48 @classmethod\r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/utils/py_utils.py in dumps(obj)\r\n 365 file = StringIO()\r\n 366 with _no_cache_fields(obj):\r\n--> 367 dump(obj, file)\r\n 368 return file.getvalue()\r\n 369 \r\n\r\n/usr/local/lib/python3.6/site-packages/datasets/utils/py_utils.py in dump(obj, file)\r\n 337 def dump(obj, file):\r\n 338 \"\"\"pickle an object to a file\"\"\"\r\n--> 339 Pickler(file, recurse=True).dump(obj)\r\n 340 return\r\n 341 \r\n\r\n/usr/local/lib/python3.6/site-packages/dill/_dill.py in dump(self, obj)\r\n 444 raise PicklingError(msg)\r\n 445 else:\r\n--> 446 StockPickler.dump(self, obj)\r\n 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects\r\n 448 return\r\n\r\n/usr/local/lib/python3.6/pickle.py in dump(self, obj)\r\n 407 if self.proto >= 4:\r\n 408 self.framer.start_framing()\r\n--> 409 self.save(obj)\r\n 410 self.write(STOP)\r\n 411 self.framer.end_framing()\r\n\r\n/usr/local/lib/python3.6/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n/usr/local/lib/python3.6/site-packages/dill/_dill.py in save_function(pickler, obj)\r\n 1436 globs, obj.__name__,\r\n 1437 obj.__defaults__, obj.__closure__,\r\n-> 1438 obj.__dict__, fkwdefaults), obj=obj)\r\n 1439 else:\r\n 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False)\r\n\r\n/usr/local/lib/python3.6/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 608 else:\r\n 609 save(func)\r\n--> 610 save(args)\r\n 611 write(REDUCE)\r\n 612 \r\n\r\n/usr/local/lib/python3.6/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n/usr/local/lib/python3.6/pickle.py in save_tuple(self, obj)\r\n 749 write(MARK)\r\n 750 for element in obj:\r\n--> 751 save(element)\r\n 752 \r\n 753 if id(obj) in memo:\r\n\r\n/usr/local/lib/python3.6/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n/usr/local/lib/python3.6/site-packages/dill/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n/usr/local/lib/python3.6/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n/usr/local/lib/python3.6/pickle.py in _batch_setitems(self, items)\r\n 850 k, v = tmp[0]\r\n 851 save(k)\r\n--> 852 save(v)\r\n 853 write(SETITEM)\r\n 854 # else tmp is empty, and we're done\r\n\r\n/usr/local/lib/python3.6/pickle.py in save(self, obj, save_persistent_id)\r\n 519 \r\n 520 # Save the reduce() output and finally memoize the object\r\n--> 521 self.save_reduce(obj=obj, *rv)\r\n 522 \r\n 523 def persistent_id(self, obj):\r\n\r\n/usr/local/lib/python3.6/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 632 \r\n 633 if state is not None:\r\n--> 634 save(state)\r\n 635 write(BUILD)\r\n 636 \r\n\r\n/usr/local/lib/python3.6/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n/usr/local/lib/python3.6/site-packages/dill/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n/usr/local/lib/python3.6/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n/usr/local/lib/python3.6/pickle.py in _batch_setitems(self, items)\r\n 845 for k, v in tmp:\r\n 846 save(k)\r\n--> 847 save(v)\r\n 848 write(SETITEMS)\r\n 849 elif n:\r\n\r\n/usr/local/lib/python3.6/pickle.py in save(self, obj, save_persistent_id)\r\n 519 \r\n 520 # Save the reduce() output and finally memoize the object\r\n--> 521 self.save_reduce(obj=obj, *rv)\r\n 522 \r\n 523 def persistent_id(self, obj):\r\n\r\n/usr/local/lib/python3.6/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj)\r\n 632 \r\n 633 if state is not None:\r\n--> 634 save(state)\r\n 635 write(BUILD)\r\n 636 \r\n\r\n/usr/local/lib/python3.6/pickle.py in save(self, obj, save_persistent_id)\r\n 474 f = self.dispatch.get(t)\r\n 475 if f is not None:\r\n--> 476 f(self, obj) # Call unbound method with explicit self\r\n 477 return\r\n 478 \r\n\r\n/usr/local/lib/python3.6/site-packages/dill/_dill.py in save_module_dict(pickler, obj)\r\n 931 # we only care about session the first pass thru\r\n 932 pickler._session = False\r\n--> 933 StockPickler.save_dict(pickler, obj)\r\n 934 log.info(\"# D2\")\r\n 935 return\r\n\r\n/usr/local/lib/python3.6/pickle.py in save_dict(self, obj)\r\n 819 \r\n 820 self.memoize(obj)\r\n--> 821 self._batch_setitems(obj.items())\r\n 822 \r\n 823 dispatch[dict] = save_dict\r\n\r\n/usr/local/lib/python3.6/pickle.py in _batch_setitems(self, items)\r\n 845 for k, v in tmp:\r\n 846 save(k)\r\n--> 847 save(v)\r\n 848 write(SETITEMS)\r\n 849 elif n:\r\n\r\n/usr/local/lib/python3.6/pickle.py in save(self, obj, save_persistent_id)\r\n 494 reduce = getattr(obj, \"__reduce_ex__\", None)\r\n 495 if reduce is not None:\r\n--> 496 rv = reduce(self.proto)\r\n 497 else:\r\n 498 reduce = getattr(obj, \"__reduce__\", None)\r\n\r\nTypeError: can't pickle Tagger objects\r\n```\r\n\r\n</details>\r\n\r\ntrainsformers: 2.10.0\r\ndatasets: 1.0.2\r\ndill: 0.3.2\r\npython: 3.6.8\r\n\r\nOS: ubuntu 16.04 (Docker Image) on [Deep Learning VM](https://console.cloud.google.com/marketplace/details/click-to-deploy-images/deeplearning) (GCP)\r\nGPU: Tesla P100 (CUDA 10)\r\n", "> I have the same issue with `transformers/BertJapaneseTokenizer`.\r\n\r\nIt looks like it this tokenizer is not supported unfortunately.\r\nThis is because `t.word_tokenizer.mecab` is a `fugashi.fugashi.GenericTagger` which is not compatible with pickle nor dill.\r\n\r\nWe need objects passes to `map` to be picklable for our caching system to work properly.\r\nHere it crashes because the caching system is not able to pickle the GenericTagger.\r\n\r\n\\> Maybe you can create an issue on [fugashi](https://github.com/polm/fugashi/issues) 's repo and ask to make `fugashi.fugashi.GenericTagger` compatible with pickle ?\r\n\r\nWhat you can do in the meantime is use a picklable wrapper of the tokenizer:\r\n\r\n\r\n```python\r\nfrom transformers import BertJapaneseTokenizer, MecabTokenizer\r\n\r\nclass PicklableTokenizer(BertJapaneseTokenizer):\r\n\r\n def __getstate__(self):\r\n state = dict(self.__dict__)\r\n state[\"do_lower_case\"] = self.word_tokenizer.do_lower_case\r\n state[\"never_split\"] = self.word_tokenizer.never_split \r\n del state[\"word_tokenizer\"]\r\n return state\r\n\r\n def __setstate__(self, state):\r\n do_lower_case = state.pop(\"do_lower_case\")\r\n never_split = state.pop(\"never_split\")\r\n self.__dict__ = state\r\n self.word_tokenizer = MecabTokenizer(\r\n do_lower_case=do_lower_case, never_split=never_split)\r\n )\r\n\r\nt = PicklableTokenizer.from_pretrained(\"cl-tohoku/bert-base-japanese-whole-word-masking\")\r\nencoded = train_ds.map(lambda examples: {'tokens': t.encode(examples['title'])}, batched=True) # it works\r\n```", "We can also update the `BertJapaneseTokenizer` in `transformers` as you just shown @lhoestq to make it compatible with pickle. It will be faster than asking on fugashi 's repo and good for the other users of `transformers` as well.\r\n\r\nI'm currently working on `transformers` I'll include it in the https://github.com/huggingface/transformers/pull/7141 PR and the next release of `transformers`.", "Thank you for the rapid and polite response!\r\n\r\n@lhoestq Thanks for the suggestion! I've passed the pickle phase, but another `ArrowInvalid` problem occored. I created another issue #687 .\r\n\r\n@thomwolf Wow, really fast work. I'm looking forward to the next release 🤗" ]
1,600,835,294,000
1,602,149,536,000
1,602,149,536,000
NONE
null
null
I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`. ``` def convert_to_features(example): # Tokenize contexts and questions (as pairs of inputs) input_pairs = [example['question'], example['context']] encodings = tokenizer.encode_plus(input_pairs, pad_to_max_length=True, max_length=512) context_encodings = tokenizer.encode_plus(example['context']) # Compute start and end tokens for labels using Transformers's fast tokenizers alignement methodes. # this will give us the position of answer span in the context text start_idx, end_idx = get_correct_alignement(example['context'], example['answers']) start_positions_context = context_encodings.char_to_token(start_idx) end_positions_context = context_encodings.char_to_token(end_idx-1) # here we will compute the start and end position of the answer in the whole example # as the example is encoded like this <s> question</s></s> context</s> # and we know the postion of the answer in the context # we can just find out the index of the sep token and then add that to position + 1 (+1 because there are two sep tokens) # this will give us the position of the answer span in whole example sep_idx = encodings['input_ids'].index(tokenizer.sep_token_id) start_positions = start_positions_context + sep_idx + 1 end_positions = end_positions_context + sep_idx + 1 if end_positions > 512: start_positions, end_positions = 0, 0 encodings.update({'start_positions': start_positions, 'end_positions': end_positions, 'attention_mask': encodings['attention_mask']}) return encodings ``` Then I run `dataset.map(convert_to_features)`, it raise ``` In [59]: a.map(convert_to_features) --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-59-c453b508761d> in <module> ----> 1 a.map(convert_to_features) /opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint) 1242 fn_kwargs=fn_kwargs, 1243 new_fingerprint=new_fingerprint, -> 1244 update_data=update_data, 1245 ) 1246 else: /opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py in wrapper(*args, **kwargs) 151 "output_all_columns": self._output_all_columns, 152 } --> 153 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 154 if new_format["columns"] is not None: 155 new_format["columns"] = list(set(new_format["columns"]) & set(out.column_names)) /opt/conda/lib/python3.7/site-packages/datasets/fingerprint.py in wrapper(*args, **kwargs) 156 kwargs_for_fingerprint["fingerprint_name"] = fingerprint_name 157 kwargs[fingerprint_name] = update_fingerprint( --> 158 self._fingerprint, transform, kwargs_for_fingerprint 159 ) 160 /opt/conda/lib/python3.7/site-packages/datasets/fingerprint.py in update_fingerprint(fingerprint, transform, transform_args) 103 for key in sorted(transform_args): 104 hasher.update(key) --> 105 hasher.update(transform_args[key]) 106 return hasher.hexdigest() 107 /opt/conda/lib/python3.7/site-packages/datasets/fingerprint.py in update(self, value) 55 def update(self, value): 56 self.m.update(f"=={type(value)}==".encode("utf8")) ---> 57 self.m.update(self.hash(value).encode("utf-8")) 58 59 def hexdigest(self): /opt/conda/lib/python3.7/site-packages/datasets/fingerprint.py in hash(cls, value) 51 return cls.dispatch[type(value)](cls, value) 52 else: ---> 53 return cls.hash_default(value) 54 55 def update(self, value): /opt/conda/lib/python3.7/site-packages/datasets/fingerprint.py in hash_default(cls, value) 44 @classmethod 45 def hash_default(cls, value): ---> 46 return cls.hash_bytes(dumps(value)) 47 48 @classmethod /opt/conda/lib/python3.7/site-packages/datasets/utils/py_utils.py in dumps(obj) 365 file = StringIO() 366 with _no_cache_fields(obj): --> 367 dump(obj, file) 368 return file.getvalue() 369 /opt/conda/lib/python3.7/site-packages/datasets/utils/py_utils.py in dump(obj, file) 337 def dump(obj, file): 338 """pickle an object to a file""" --> 339 Pickler(file, recurse=True).dump(obj) 340 return 341 /opt/conda/lib/python3.7/site-packages/dill/_dill.py in dump(self, obj) 444 raise PicklingError(msg) 445 else: --> 446 StockPickler.dump(self, obj) 447 stack.clear() # clear record of 'recursion-sensitive' pickled objects 448 return /opt/conda/lib/python3.7/pickle.py in dump(self, obj) 435 if self.proto >= 4: 436 self.framer.start_framing() --> 437 self.save(obj) 438 self.write(STOP) 439 self.framer.end_framing() /opt/conda/lib/python3.7/pickle.py in save(self, obj, save_persistent_id) 502 f = self.dispatch.get(t) 503 if f is not None: --> 504 f(self, obj) # Call unbound method with explicit self 505 return 506 /opt/conda/lib/python3.7/site-packages/dill/_dill.py in save_function(pickler, obj) 1436 globs, obj.__name__, 1437 obj.__defaults__, obj.__closure__, -> 1438 obj.__dict__, fkwdefaults), obj=obj) 1439 else: 1440 _super = ('super' in getattr(obj.func_code,'co_names',())) and (_byref is not None) and getattr(pickler, '_recurse', False) /opt/conda/lib/python3.7/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj) 636 else: 637 save(func) --> 638 save(args) 639 write(REDUCE) 640 /opt/conda/lib/python3.7/pickle.py in save(self, obj, save_persistent_id) 502 f = self.dispatch.get(t) 503 if f is not None: --> 504 f(self, obj) # Call unbound method with explicit self 505 return 506 /opt/conda/lib/python3.7/pickle.py in save_tuple(self, obj) 787 write(MARK) 788 for element in obj: --> 789 save(element) 790 791 if id(obj) in memo: /opt/conda/lib/python3.7/pickle.py in save(self, obj, save_persistent_id) 502 f = self.dispatch.get(t) 503 if f is not None: --> 504 f(self, obj) # Call unbound method with explicit self 505 return 506 /opt/conda/lib/python3.7/site-packages/dill/_dill.py in save_module_dict(pickler, obj) 931 # we only care about session the first pass thru 932 pickler._session = False --> 933 StockPickler.save_dict(pickler, obj) 934 log.info("# D2") 935 return /opt/conda/lib/python3.7/pickle.py in save_dict(self, obj) 857 858 self.memoize(obj) --> 859 self._batch_setitems(obj.items()) 860 861 dispatch[dict] = save_dict /opt/conda/lib/python3.7/pickle.py in _batch_setitems(self, items) 883 for k, v in tmp: 884 save(k) --> 885 save(v) 886 write(SETITEMS) 887 elif n: /opt/conda/lib/python3.7/pickle.py in save(self, obj, save_persistent_id) 547 548 # Save the reduce() output and finally memoize the object --> 549 self.save_reduce(obj=obj, *rv) 550 551 def persistent_id(self, obj): /opt/conda/lib/python3.7/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj) 660 661 if state is not None: --> 662 save(state) 663 write(BUILD) 664 /opt/conda/lib/python3.7/pickle.py in save(self, obj, save_persistent_id) 502 f = self.dispatch.get(t) 503 if f is not None: --> 504 f(self, obj) # Call unbound method with explicit self 505 return 506 /opt/conda/lib/python3.7/site-packages/dill/_dill.py in save_module_dict(pickler, obj) 931 # we only care about session the first pass thru 932 pickler._session = False --> 933 StockPickler.save_dict(pickler, obj) 934 log.info("# D2") 935 return /opt/conda/lib/python3.7/pickle.py in save_dict(self, obj) 857 858 self.memoize(obj) --> 859 self._batch_setitems(obj.items()) 860 861 dispatch[dict] = save_dict /opt/conda/lib/python3.7/pickle.py in _batch_setitems(self, items) 883 for k, v in tmp: 884 save(k) --> 885 save(v) 886 write(SETITEMS) 887 elif n: /opt/conda/lib/python3.7/pickle.py in save(self, obj, save_persistent_id) 547 548 # Save the reduce() output and finally memoize the object --> 549 self.save_reduce(obj=obj, *rv) 550 551 def persistent_id(self, obj): /opt/conda/lib/python3.7/pickle.py in save_reduce(self, func, args, state, listitems, dictitems, obj) 660 661 if state is not None: --> 662 save(state) 663 write(BUILD) 664 /opt/conda/lib/python3.7/pickle.py in save(self, obj, save_persistent_id) 502 f = self.dispatch.get(t) 503 if f is not None: --> 504 f(self, obj) # Call unbound method with explicit self 505 return 506 /opt/conda/lib/python3.7/site-packages/dill/_dill.py in save_module_dict(pickler, obj) 931 # we only care about session the first pass thru 932 pickler._session = False --> 933 StockPickler.save_dict(pickler, obj) 934 log.info("# D2") 935 return /opt/conda/lib/python3.7/pickle.py in save_dict(self, obj) 857 858 self.memoize(obj) --> 859 self._batch_setitems(obj.items()) 860 861 dispatch[dict] = save_dict /opt/conda/lib/python3.7/pickle.py in _batch_setitems(self, items) 883 for k, v in tmp: 884 save(k) --> 885 save(v) 886 write(SETITEMS) 887 elif n: /opt/conda/lib/python3.7/pickle.py in save(self, obj, save_persistent_id) 522 reduce = getattr(obj, "__reduce_ex__", None) 523 if reduce is not None: --> 524 rv = reduce(self.proto) 525 else: 526 reduce = getattr(obj, "__reduce__", None) TypeError: can't pickle Tokenizer objects ```
https://api.github.com/repos/huggingface/datasets/issues/665/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/664
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/664/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/664/comments
https://api.github.com/repos/huggingface/datasets/issues/664/events
https://github.com/huggingface/datasets/issues/664
707,017,791
MDU6SXNzdWU3MDcwMTc3OTE=
664
load_dataset from local squad.py, raise error: TypeError: 'NoneType' object is not callable
{ "login": "xixiaoyao", "id": 24541791, "node_id": "MDQ6VXNlcjI0NTQxNzkx", "avatar_url": "https://avatars.githubusercontent.com/u/24541791?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xixiaoyao", "html_url": "https://github.com/xixiaoyao", "followers_url": "https://api.github.com/users/xixiaoyao/followers", "following_url": "https://api.github.com/users/xixiaoyao/following{/other_user}", "gists_url": "https://api.github.com/users/xixiaoyao/gists{/gist_id}", "starred_url": "https://api.github.com/users/xixiaoyao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xixiaoyao/subscriptions", "organizations_url": "https://api.github.com/users/xixiaoyao/orgs", "repos_url": "https://api.github.com/users/xixiaoyao/repos", "events_url": "https://api.github.com/users/xixiaoyao/events{/privacy}", "received_events_url": "https://api.github.com/users/xixiaoyao/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi !\r\nThanks for reporting.\r\nIt looks like no object inherits from `datasets.GeneratorBasedBuilder` (or more generally from `datasets.DatasetBuilder`) in your script.\r\n\r\nCould you check that there exist at least one dataset builder class ?", "Hi @xixiaoyao did you manage to fix your issue ?", "No activity, closing", "It happened when try to change the old project which use 'nlp' to new project which use 'datasets'. You should check you old 'my_squad.py' file, change the inherit class from `nlp.xxx` to `datasets.xxx`. Otherwise datasets - load.py - import_main_class() `if inspect.isclass(obj) and issubclass(obj, main_cls_type):` can not find the main_cls." ]
1,600,833,216,000
1,603,184,773,000
1,603,184,773,000
NONE
null
null
version: 1.0.2 ``` train_dataset = datasets.load_dataset('squad') ``` The above code can works. However, when I download the squad.py from your server, and saved as `my_squad.py` to local. I run followings raise errors. ``` train_dataset = datasets.load_dataset('./my_squad.py') ``` --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-28-25a84b4d1581> in <module> ----> 1 train_dataset = nlp.load_dataset('./my_squad.py') /opt/conda/lib/python3.7/site-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs) 602 hash=hash, 603 features=features, --> 604 **config_kwargs, 605 ) 606 TypeError: 'NoneType' object is not callable
https://api.github.com/repos/huggingface/datasets/issues/664/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/663
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/663/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/663/comments
https://api.github.com/repos/huggingface/datasets/issues/663/events
https://github.com/huggingface/datasets/pull/663
706,732,636
MDExOlB1bGxSZXF1ZXN0NDkxMjI3NzUz
663
Created dataset card snli.md
{ "login": "mcmillanmajora", "id": 26722925, "node_id": "MDQ6VXNlcjI2NzIyOTI1", "avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mcmillanmajora", "html_url": "https://github.com/mcmillanmajora", "followers_url": "https://api.github.com/users/mcmillanmajora/followers", "following_url": "https://api.github.com/users/mcmillanmajora/following{/other_user}", "gists_url": "https://api.github.com/users/mcmillanmajora/gists{/gist_id}", "starred_url": "https://api.github.com/users/mcmillanmajora/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mcmillanmajora/subscriptions", "organizations_url": "https://api.github.com/users/mcmillanmajora/orgs", "repos_url": "https://api.github.com/users/mcmillanmajora/repos", "events_url": "https://api.github.com/users/mcmillanmajora/events{/privacy}", "received_events_url": "https://api.github.com/users/mcmillanmajora/received_events", "type": "User", "site_admin": false }
[ { "id": 2067401494, "node_id": "MDU6TGFiZWwyMDY3NDAxNDk0", "url": "https://api.github.com/repos/huggingface/datasets/labels/Dataset%20discussion", "name": "Dataset discussion", "color": "72f99f", "default": false, "description": "Discussions on the datasets" } ]
closed
false
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}", "starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yjernite/subscriptions", "organizations_url": "https://api.github.com/users/yjernite/orgs", "repos_url": "https://api.github.com/users/yjernite/repos", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "received_events_url": "https://api.github.com/users/yjernite/received_events", "type": "User", "site_admin": false }
[ { "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}", "starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yjernite/subscriptions", "organizations_url": "https://api.github.com/users/yjernite/orgs", "repos_url": "https://api.github.com/users/yjernite/repos", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "received_events_url": "https://api.github.com/users/yjernite/received_events", "type": "User", "site_admin": false } ]
null
[ "Adding a direct link to the rendered markdown:\r\nhttps://github.com/mcmillanmajora/datasets/blob/add_dataset_documentation/datasets/snli/README.md\r\n", "It would be amazing if we ended up with this much information on all of our datasets :) \r\n\r\nI don't think there's too much repetition, everything that is in here is relevant. The main challenge will be to figure out how to structure the sheet so that all of the information can be presented without overwhelming the reader. We'll also want to have as much of it as possible in structured form so it can be easily navigated.", "@mcmillanmajora for now can you remove the prompts / quoted blocks so we can see what the datasheet would look like on its own?\r\n\r\nWould also love to hear if @sgugger has some first impressions", "I removed the prompts. It's definitely a little easier to read without them!", "Should we name the file `README.md` for consistency with models?", "Asked @sleepinyourhat for some insights too :) ", "Thank you for taking the time to look through the card and for all your comments @sleepinyourhat ! I've incorporated them in the latest update. ", "Be careful to keep the ‘sa’ term in the license. It’s something we\ninherited from the Flickr captions.\n\nOn Thu, Oct 1, 2020 at 10:09 AM Julien Chaumond <[email protected]>\nwrote:\n\n> *@julien-c* commented on this pull request.\n> ------------------------------\n>\n> In datasets/snli/README.md\n> <https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_huggingface_datasets_pull_663-23discussion-5Fr498273172&d=DwMFaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=sCzLyHdE8zgQwk2-sKwA1w&m=PHPCew9Xj3CBQrudcaii70ln-wpRtbngE_tj3Ioy3NI&s=WbEkKXCbL6j5Ui3sox_WqvzrbShbJn2WW-51SENL2ZQ&e=>\n> :\n>\n> > +---\n> +language:\n> +- en\n> +task:\n> +- text-classification\n> +purpose:\n> +- NLI\n> +size:\n> +- \">100k\"\n> +language producers:\n> +- crowdsourced\n> +annotation:\n> +- crowdsourced\n> +tags:\n> +- extended-from-other-datasets\n> +license: \"CC BY-SA 4.0\"\n>\n> ⬇️ Suggested change\n>\n> -license: \"CC BY-SA 4.0\"\n> +license: cc-by-4.0\n>\n> For models (documented at\n> https://huggingface.co/docs#what-metadata-can-i-add-to-my-model-card\n> <https://urldefense.proofpoint.com/v2/url?u=https-3A__huggingface.co_docs-23what-2Dmetadata-2Dcan-2Di-2Dadd-2Dto-2Dmy-2Dmodel-2Dcard&d=DwMFaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=sCzLyHdE8zgQwk2-sKwA1w&m=PHPCew9Xj3CBQrudcaii70ln-wpRtbngE_tj3Ioy3NI&s=ck3x8c_ujrwKReDTSGuWWgD9W6REHEPbZaO7S4GFRd4&e=>)\n> we use the License keywords listed by GitHub at\n> https://docs.github.com/en/free-pro-team@latest/github/creating-cloning-and-archiving-repositories/licensing-a-repository#searching-github-by-license-type\n> <https://urldefense.proofpoint.com/v2/url?u=https-3A__docs.github.com_en_free-2Dpro-2Dteam-40latest_github_creating-2Dcloning-2Dand-2Darchiving-2Drepositories_licensing-2Da-2Drepository-23searching-2Dgithub-2Dby-2Dlicense-2Dtype&d=DwMFaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=sCzLyHdE8zgQwk2-sKwA1w&m=PHPCew9Xj3CBQrudcaii70ln-wpRtbngE_tj3Ioy3NI&s=dWBP-ZvtMErD-egoBiBTCKA4500mjDXVSk03oW1g16U&e=>\n>\n> (Hopefully we'll plug some sort of form validation for users at some point)\n>\n> —\n> You are receiving this because you were mentioned.\n> Reply to this email directly, view it on GitHub\n> <https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_huggingface_datasets_pull_663-23pullrequestreview-2D500386385&d=DwMFaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=sCzLyHdE8zgQwk2-sKwA1w&m=PHPCew9Xj3CBQrudcaii70ln-wpRtbngE_tj3Ioy3NI&s=HU2Hwi7HH9W2NtMoCIiQlhXxxEULLi8L9gnWU5PBAPY&e=>,\n> or unsubscribe\n> <https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_notifications_unsubscribe-2Dauth_AAJZSWL63W2LB7SBICA2GMTSISEPZANCNFSM4RWKAZRA&d=DwMFaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=sCzLyHdE8zgQwk2-sKwA1w&m=PHPCew9Xj3CBQrudcaii70ln-wpRtbngE_tj3Ioy3NI&s=086__lKQLxTanHfjE8kOIpaJbaWPzBB9gGIt_prWeH8&e=>\n> .\n>\n", "@sleepinyourhat You're right, wrong copy/paste", "Question: Where does this standard come from? It looks similar to both\n'Data Statements' and 'Datasheets for Datasets', but it doesn't look quite\nlike either.\n\nOn Mon, Oct 12, 2020 at 4:27 PM Yacine Jernite <[email protected]>\nwrote:\n\n> Merged #663\n> <https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_huggingface_datasets_pull_663&d=DwMCaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=sCzLyHdE8zgQwk2-sKwA1w&m=D34WbiHBTYHOdXsI9JV9wJqSieP6zAPGqGKDziM5uKU&s=s4_X-BSEnTKgGg9rPLBt3cyVptyMX_iWD5Ql3UMBi-I&e=>\n> into master.\n>\n> —\n> You are receiving this because you were mentioned.\n> Reply to this email directly, view it on GitHub\n> <https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_huggingface_datasets_pull_663-23event-2D3868180429&d=DwMCaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=sCzLyHdE8zgQwk2-sKwA1w&m=D34WbiHBTYHOdXsI9JV9wJqSieP6zAPGqGKDziM5uKU&s=elcM4umqReQfIrgHhpey9W_wPaq5QRgq7xNlubM47QI&e=>,\n> or unsubscribe\n> <https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_notifications_unsubscribe-2Dauth_AAJZSWJVGQRCR4OTTV27VTTSKNRBXANCNFSM4RWKAZRA&d=DwMCaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=sCzLyHdE8zgQwk2-sKwA1w&m=D34WbiHBTYHOdXsI9JV9wJqSieP6zAPGqGKDziM5uKU&s=NB6nEROnTPgwNyF3ZklOmHnvP7kOkOm7sEa740KbVCs&e=>\n> .\n>\n", "@sleepinyourhat The schema is definitely drawing from Data Statements and Datasheets for Datasets but we also wanted to include some more general information to introduce the dataset to new users. If you have any suggestions for changes to the schema itself, please let us know!" ]
1,600,813,777,000
1,602,608,720,000
1,602,534,412,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/663", "html_url": "https://github.com/huggingface/datasets/pull/663", "diff_url": "https://github.com/huggingface/datasets/pull/663.diff", "patch_url": "https://github.com/huggingface/datasets/pull/663.patch" }
First draft of a dataset card using the SNLI corpus as an example. This is mostly based on the [Google Doc draft](https://docs.google.com/document/d/1dKPGP-dA2W0QoTRGfqQ5eBp0CeSsTy7g2yM8RseHtos/edit), but I added a few sections and moved some things around. - I moved **Who Was Involved** to follow **Language**, both because I thought the authors should be presented more towards the front and because I think it makes sense to present the speakers close to the language so it doesn't have to be repeated. - I created a section I called **Data Characteristics** by pulling some things out of the other sections. I was thinking that this would be more about the language use in context of the specific task construction. That name isn't very descriptive though and could probably be improved. -- Domain and language type out of **Language**. I particularly wanted to keep the Language section as simple and as abstracted from the task as possible. -- 'How was the data collected' out of **Who Was Involved** -- Normalization out of **Features/Dataset Structure** -- I also added an annotation process section. - I kept the **Features** section mostly the same as the Google Doc, but I renamed it **Dataset Structure** to more clearly separate it from the language use, and added some links to the documentation pages. - I also kept **Tasks Supported**, **Known Limitations**, and **Licensing Information** mostly the same. Looking at it again though, maybe **Tasks Supported** should come before **Data Characteristics**? The trickiest part about writing a dataset card for the SNLI corpus specifically is that it's built on datasets which are themselves built on datasets so I had to dig in a lot of places to find information. I think this will be easier with other datasets and once there is more uptake of dataset cards so they can just link to each other. (Maybe that needs to be an added section?) I also made an effort not to repeat information across the sections or to refer to a previous section if the information was relevant in a later one. Is there too much repetition still?
https://api.github.com/repos/huggingface/datasets/issues/663/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/662
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/662/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/662/comments
https://api.github.com/repos/huggingface/datasets/issues/662/events
https://github.com/huggingface/datasets/pull/662
706,689,866
MDExOlB1bGxSZXF1ZXN0NDkxMTkyNTM3
662
Created dataset card snli.md
{ "login": "mcmillanmajora", "id": 26722925, "node_id": "MDQ6VXNlcjI2NzIyOTI1", "avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mcmillanmajora", "html_url": "https://github.com/mcmillanmajora", "followers_url": "https://api.github.com/users/mcmillanmajora/followers", "following_url": "https://api.github.com/users/mcmillanmajora/following{/other_user}", "gists_url": "https://api.github.com/users/mcmillanmajora/gists{/gist_id}", "starred_url": "https://api.github.com/users/mcmillanmajora/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mcmillanmajora/subscriptions", "organizations_url": "https://api.github.com/users/mcmillanmajora/orgs", "repos_url": "https://api.github.com/users/mcmillanmajora/repos", "events_url": "https://api.github.com/users/mcmillanmajora/events{/privacy}", "received_events_url": "https://api.github.com/users/mcmillanmajora/received_events", "type": "User", "site_admin": false }
[ { "id": 2067401494, "node_id": "MDU6TGFiZWwyMDY3NDAxNDk0", "url": "https://api.github.com/repos/huggingface/datasets/labels/Dataset%20discussion", "name": "Dataset discussion", "color": "72f99f", "default": false, "description": "Discussions on the datasets" } ]
closed
false
null
[]
null
[ "Resubmitting on a new fork" ]
1,600,808,417,000
1,600,809,981,000
1,600,809,981,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/662", "html_url": "https://github.com/huggingface/datasets/pull/662", "diff_url": "https://github.com/huggingface/datasets/pull/662.diff", "patch_url": "https://github.com/huggingface/datasets/pull/662.patch" }
First draft of a dataset card using the SNLI corpus as an example
https://api.github.com/repos/huggingface/datasets/issues/662/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/661
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/661/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/661/comments
https://api.github.com/repos/huggingface/datasets/issues/661/events
https://github.com/huggingface/datasets/pull/661
706,465,936
MDExOlB1bGxSZXF1ZXN0NDkxMDA3NjEw
661
Replace pa.OSFile by open
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,787,159,000
1,620,239,076,000
1,600,787,725,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/661", "html_url": "https://github.com/huggingface/datasets/pull/661", "diff_url": "https://github.com/huggingface/datasets/pull/661.diff", "patch_url": "https://github.com/huggingface/datasets/pull/661.patch" }
It should fix #643
https://api.github.com/repos/huggingface/datasets/issues/661/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/660
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/660/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/660/comments
https://api.github.com/repos/huggingface/datasets/issues/660/events
https://github.com/huggingface/datasets/pull/660
706,324,032
MDExOlB1bGxSZXF1ZXN0NDkwODkyMjQ0
660
add openwebtext
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gists_url": "https://api.github.com/users/richarddwang/gists{/gist_id}", "starred_url": "https://api.github.com/users/richarddwang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/richarddwang/subscriptions", "organizations_url": "https://api.github.com/users/richarddwang/orgs", "repos_url": "https://api.github.com/users/richarddwang/repos", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "received_events_url": "https://api.github.com/users/richarddwang/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "BTW, is there a one-line command to make our building scripts pass flake8 test? (included code quality test), I got like trailing space or mixed space and tab warning and error, and fixed them manually.", "> BTW, is there a one-line command to make our building scripts pass flake8 test? (included code quality test), I got like trailing space or mixed space and tab warning and error, and fixed them manually.\r\n\r\nI don't think so.\r\nWe have a command for black and isort but not flake8 as far as I know.", "Thanks for your awesome work too.\r\nBTW a little reminder, this solves #132 " ]
1,600,776,322,000
1,601,976,010,000
1,601,284,046,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/660", "html_url": "https://github.com/huggingface/datasets/pull/660", "diff_url": "https://github.com/huggingface/datasets/pull/660.diff", "patch_url": "https://github.com/huggingface/datasets/pull/660.patch" }
This adds [The OpenWebText Corpus](https://skylion007.github.io/OpenWebTextCorpus/), which is a clean and large text corpus for nlp pretraining. It is an open source effort to reproduce OpenAI’s WebText dataset used by GPT-2, and it is also needed to reproduce ELECTRA. It solves #132 . ### Besides dataset building script, I made some changes to the library. 1. Extract large amount of compressed files with multi processing I add a `num_proc` argument to `DownloadManager.extract` and pass this `num_proc` to `map_nested`. So I can decompress 20 thousands compressed files faster. `num_proc` I add is default to `None`, so it shouldn't break any other thing. 2. In `cached_path`, I change the order to deal with different kind of compressed files (zip, tar, gzip) Because there is no way to 100% detect a file is a zip file (see [this](https://stackoverflow.com/questions/18194688/how-can-i-determine-if-a-file-is-a-zip-file)), I found it wrongly detect `'./datasets/downloads/extracted/58764bd6898fa339b25d92e7fbbc3d8dbf64fb504edff1a30a1d7d99d1561027/openwebtext/urlsf_subset13-630_data.xz'` as a zip and try decompress it with zip, sure it will get error. So I made it detect wheter the file is tar or gzip first and detect zip in the last. 3. `MockDownloadManager.extract` Cuz I pass `num_proc` to `DownloadManager.extract`, I also have to make `MockDownloadManager.extract` to accept extra keywork arguments. So I make it `extract(path, *args, **kwargs)`, but just return the path as original implementation. **Note**: If there is better way for points mentioned above, thought I would like to help, unless we can solve point4 (make dataset building fast), I may not be able to afford rebuild the dataset again because of change of the dataset script (Building the dataset cost me 4 days). ### There is something I think we can improve 4. Long time to decompress compressed files Even I decompress those 20 thousands compressed files with 12 process on my 16 core 3.x Ghz server. It still took about 3 ~ 4days to complete dataset building. Most of time spent on decompress those files. ### Info about the source data The source data is an tar.xz file with following structure, files/directory beyond compressed file is what can we get after decompress it. ``` openwebtext.tar.xz |__ openwebtext |__subset000.xz | |__ ....txt | |__ ....txt | ... |__ subset001.xz | .... ``` And this the structure of dummy data, same as the original one. ``` dummy_data.zip |__ dummy_data |__ openwebtext |__fake_subset-1_data-dirxz # actually it is a directory | |__ ....txt | |__ ....txt |__ fake_subset-2_data-dirxz |__ ....txt |__ ....txt ```
https://api.github.com/repos/huggingface/datasets/issues/660/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/659
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/659/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/659/comments
https://api.github.com/repos/huggingface/datasets/issues/659/events
https://github.com/huggingface/datasets/pull/659
706,231,506
MDExOlB1bGxSZXF1ZXN0NDkwODE4NTY1
659
Keep new columns in transmit format
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,768,043,000
1,600,769,242,000
1,600,769,240,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/659", "html_url": "https://github.com/huggingface/datasets/pull/659", "diff_url": "https://github.com/huggingface/datasets/pull/659.diff", "patch_url": "https://github.com/huggingface/datasets/pull/659.patch" }
When a dataset is formatted with a list of columns that `__getitem__` should return, then calling `map` to add new columns doesn't add the new columns to this list. It caused `KeyError` issues in #620 I changed the logic to add those new columns to the list that `__getitem__` should return.
https://api.github.com/repos/huggingface/datasets/issues/659/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/658
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/658/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/658/comments
https://api.github.com/repos/huggingface/datasets/issues/658/events
https://github.com/huggingface/datasets/pull/658
706,206,247
MDExOlB1bGxSZXF1ZXN0NDkwNzk4MDc0
658
Fix squad metric's Features
{ "login": "tshrjn", "id": 8372098, "node_id": "MDQ6VXNlcjgzNzIwOTg=", "avatar_url": "https://avatars.githubusercontent.com/u/8372098?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tshrjn", "html_url": "https://github.com/tshrjn", "followers_url": "https://api.github.com/users/tshrjn/followers", "following_url": "https://api.github.com/users/tshrjn/following{/other_user}", "gists_url": "https://api.github.com/users/tshrjn/gists{/gist_id}", "starred_url": "https://api.github.com/users/tshrjn/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tshrjn/subscriptions", "organizations_url": "https://api.github.com/users/tshrjn/orgs", "repos_url": "https://api.github.com/users/tshrjn/repos", "events_url": "https://api.github.com/users/tshrjn/events{/privacy}", "received_events_url": "https://api.github.com/users/tshrjn/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Closing this one in favor of #670 \r\n\r\nThanks again for reporting the issue and proposing this fix !\r\nLet me know if you have other remarks" ]
1,600,765,792,000
1,601,395,110,000
1,601,395,110,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/658", "html_url": "https://github.com/huggingface/datasets/pull/658", "diff_url": "https://github.com/huggingface/datasets/pull/658.diff", "patch_url": "https://github.com/huggingface/datasets/pull/658.patch" }
Resolves issue [657](https://github.com/huggingface/datasets/issues/657).
https://api.github.com/repos/huggingface/datasets/issues/658/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/657
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/657/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/657/comments
https://api.github.com/repos/huggingface/datasets/issues/657/events
https://github.com/huggingface/datasets/issues/657
706,204,383
MDU6SXNzdWU3MDYyMDQzODM=
657
Squad Metric Description & Feature Mismatch
{ "login": "tshrjn", "id": 8372098, "node_id": "MDQ6VXNlcjgzNzIwOTg=", "avatar_url": "https://avatars.githubusercontent.com/u/8372098?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tshrjn", "html_url": "https://github.com/tshrjn", "followers_url": "https://api.github.com/users/tshrjn/followers", "following_url": "https://api.github.com/users/tshrjn/following{/other_user}", "gists_url": "https://api.github.com/users/tshrjn/gists{/gist_id}", "starred_url": "https://api.github.com/users/tshrjn/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tshrjn/subscriptions", "organizations_url": "https://api.github.com/users/tshrjn/orgs", "repos_url": "https://api.github.com/users/tshrjn/repos", "events_url": "https://api.github.com/users/tshrjn/events{/privacy}", "received_events_url": "https://api.github.com/users/tshrjn/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nThere indeed a mismatch between the features and the kwargs description\r\n\r\nI believe `answer_start` was added to match the squad dataset format for consistency, even though it is not used in the metric computation. I think I'd rather keep it this way, so that you can just give `references=squad[\"answers\"]` to `.compute()`.\r\nMaybe we can just fix the description then.", "But then providing the `answer_start` becomes mandatory since the format of the features is checked against the one provided in the squad [file](https://github.com/huggingface/datasets/pull/658/files)." ]
1,600,765,620,000
1,602,555,416,000
1,601,395,058,000
NONE
null
null
The [description](https://github.com/huggingface/datasets/blob/master/metrics/squad/squad.py#L39) doesn't mention `answer_start` in squad. However the `datasets.features` require [it](https://github.com/huggingface/datasets/blob/master/metrics/squad/squad.py#L68). It's also not used in the evaluation.
https://api.github.com/repos/huggingface/datasets/issues/657/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/656
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/656/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/656/comments
https://api.github.com/repos/huggingface/datasets/issues/656/events
https://github.com/huggingface/datasets/pull/656
705,736,319
MDExOlB1bGxSZXF1ZXN0NDkwNDEwODAz
656
Use multiprocess from pathos for multiprocessing
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "We can just install multiprocess actually, I'll change that", "Just an FYI: I remember that I wanted to try pathos a couple of years back and I ran into issues considering cross-platform; the code would just break on Windows. If I can verify this PR by running CPU tests on Windows, let me know!", "That's good to know thanks\r\nI guess we can just wait for #644 to be merged first. I'm working on fixing the tests for windows", "Looks like all the CI jobs on windows passed !\r\nI also tested locally on my windows and it works great :) \r\n\r\nI think this is ready to merge, let me know if you have any remarks @thomwolf @BramVanroy " ]
1,600,704,739,000
1,601,304,340,000
1,601,304,339,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/656", "html_url": "https://github.com/huggingface/datasets/pull/656", "diff_url": "https://github.com/huggingface/datasets/pull/656.diff", "patch_url": "https://github.com/huggingface/datasets/pull/656.patch" }
[Multiprocess](https://github.com/uqfoundation/multiprocess) (from the [pathos](https://github.com/uqfoundation/pathos) project) allows to use lambda functions in multiprocessed map. It was suggested to use it by @kandorm. We're already using dill which is its only dependency.
https://api.github.com/repos/huggingface/datasets/issues/656/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/655
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/655/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/655/comments
https://api.github.com/repos/huggingface/datasets/issues/655/events
https://github.com/huggingface/datasets/pull/655
705,672,208
MDExOlB1bGxSZXF1ZXN0NDkwMzU4OTQ3
655
added Winogrande debiased subset
{ "login": "TevenLeScao", "id": 26709476, "node_id": "MDQ6VXNlcjI2NzA5NDc2", "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TevenLeScao", "html_url": "https://github.com/TevenLeScao", "followers_url": "https://api.github.com/users/TevenLeScao/followers", "following_url": "https://api.github.com/users/TevenLeScao/following{/other_user}", "gists_url": "https://api.github.com/users/TevenLeScao/gists{/gist_id}", "starred_url": "https://api.github.com/users/TevenLeScao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TevenLeScao/subscriptions", "organizations_url": "https://api.github.com/users/TevenLeScao/orgs", "repos_url": "https://api.github.com/users/TevenLeScao/repos", "events_url": "https://api.github.com/users/TevenLeScao/events{/privacy}", "received_events_url": "https://api.github.com/users/TevenLeScao/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "To fix the CI you just have to copy the dummy data to the 1.1.0 folder, and maybe create the dummy ones for the `debiased` configuration", "Fixed! Thanks @lhoestq " ]
1,600,699,868,000
1,600,705,240,000
1,600,704,964,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/655", "html_url": "https://github.com/huggingface/datasets/pull/655", "diff_url": "https://github.com/huggingface/datasets/pull/655.diff", "patch_url": "https://github.com/huggingface/datasets/pull/655.patch" }
The [Winogrande](https://arxiv.org/abs/1907.10641) paper mentions a `debiased` subset that wasn't in the first release; this PR adds it.
https://api.github.com/repos/huggingface/datasets/issues/655/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/654
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/654/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/654/comments
https://api.github.com/repos/huggingface/datasets/issues/654/events
https://github.com/huggingface/datasets/pull/654
705,511,058
MDExOlB1bGxSZXF1ZXN0NDkwMjI1Nzk3
654
Allow empty inputs in metrics
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,687,596,000
1,601,956,308,000
1,600,704,818,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/654", "html_url": "https://github.com/huggingface/datasets/pull/654", "diff_url": "https://github.com/huggingface/datasets/pull/654.diff", "patch_url": "https://github.com/huggingface/datasets/pull/654.patch" }
There was an arrow error when trying to compute a metric with empty inputs. The error was occurring when reading the arrow file, before calling metric._compute.
https://api.github.com/repos/huggingface/datasets/issues/654/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/653
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/653/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/653/comments
https://api.github.com/repos/huggingface/datasets/issues/653/events
https://github.com/huggingface/datasets/pull/653
705,482,391
MDExOlB1bGxSZXF1ZXN0NDkwMjAxOTg4
653
handle data alteration when trying type
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,684,909,000
1,600,704,786,000
1,600,704,785,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/653", "html_url": "https://github.com/huggingface/datasets/pull/653", "diff_url": "https://github.com/huggingface/datasets/pull/653.diff", "patch_url": "https://github.com/huggingface/datasets/pull/653.patch" }
Fix #649 The bug came from the type inference that didn't handle a weird case in Pyarrow. Indeed this code runs without error but alters the data in arrow: ```python import pyarrow as pa type = pa.struct({"a": pa.struct({"b": pa.string()})}) array_with_altered_data = pa.array([{"a": {"b": "foo", "c": "bar"}}] * 10, type=type) print(array_with_altered_data[0].as_py()) # {'a': {'b': 'foo'}} -> the sub-field "c" is missing ``` (I don't know if this is intended in pyarrow tbh) We didn't take this case into account during type inference. By default it was keeping old features and maybe alter data. To fix that I added a line that checks that the first element of the array is not altered.
https://api.github.com/repos/huggingface/datasets/issues/653/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/652
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/652/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/652/comments
https://api.github.com/repos/huggingface/datasets/issues/652/events
https://github.com/huggingface/datasets/pull/652
705,390,850
MDExOlB1bGxSZXF1ZXN0NDkwMTI3MjIx
652
handle connection error in download_prepared_from_hf_gcs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,676,471,000
1,600,676,923,000
1,600,676,922,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/652", "html_url": "https://github.com/huggingface/datasets/pull/652", "diff_url": "https://github.com/huggingface/datasets/pull/652.diff", "patch_url": "https://github.com/huggingface/datasets/pull/652.patch" }
Fix #647
https://api.github.com/repos/huggingface/datasets/issues/652/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/651
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/651/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/651/comments
https://api.github.com/repos/huggingface/datasets/issues/651/events
https://github.com/huggingface/datasets/issues/651
705,212,034
MDU6SXNzdWU3MDUyMTIwMzQ=
651
Problem with JSON dataset format
{ "login": "vikigenius", "id": 12724810, "node_id": "MDQ6VXNlcjEyNzI0ODEw", "avatar_url": "https://avatars.githubusercontent.com/u/12724810?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vikigenius", "html_url": "https://github.com/vikigenius", "followers_url": "https://api.github.com/users/vikigenius/followers", "following_url": "https://api.github.com/users/vikigenius/following{/other_user}", "gists_url": "https://api.github.com/users/vikigenius/gists{/gist_id}", "starred_url": "https://api.github.com/users/vikigenius/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vikigenius/subscriptions", "organizations_url": "https://api.github.com/users/vikigenius/orgs", "repos_url": "https://api.github.com/users/vikigenius/repos", "events_url": "https://api.github.com/users/vikigenius/events{/privacy}", "received_events_url": "https://api.github.com/users/vikigenius/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Currently the `json` dataset doesn't support this format unfortunately.\r\nHowever you could load it with\r\n```python\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\n\r\ndf = pd.read_json(\"path_to_local.json\", orient=\"index\")\r\ndataset = Dataset.from_pandas(df)\r\n```", "or you can make a custom dataset script as explained in doc here: https://huggingface.co/docs/datasets/add_dataset.html" ]
1,600,646,234,000
1,600,690,464,000
null
NONE
null
null
I have a local json dataset with the following form. { 'id01234': {'key1': value1, 'key2': value2, 'key3': value3}, 'id01235': {'key1': value1, 'key2': value2, 'key3': value3}, . . . 'id09999': {'key1': value1, 'key2': value2, 'key3': value3} } Note that instead of a list of records it's basically a dictionary of key value pairs with the keys being the record_ids and the values being the corresponding record. Reading this with json: ``` data = datasets.load('json', data_files='path_to_local.json') ``` Throws an error and asks me to chose a field. What's the right way to handle this?
https://api.github.com/repos/huggingface/datasets/issues/651/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/650
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/650/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/650/comments
https://api.github.com/repos/huggingface/datasets/issues/650/events
https://github.com/huggingface/datasets/issues/650
704,861,844
MDU6SXNzdWU3MDQ4NjE4NDQ=
650
dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gists_url": "https://api.github.com/users/richarddwang/gists{/gist_id}", "starred_url": "https://api.github.com/users/richarddwang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/richarddwang/subscriptions", "organizations_url": "https://api.github.com/users/richarddwang/orgs", "repos_url": "https://api.github.com/users/richarddwang/repos", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "received_events_url": "https://api.github.com/users/richarddwang/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi :) \r\nIn your dummy data zip file you can just have `subset000.xz` as directories instead of compressed files.\r\nLet me know if it helps", "Thanks for your comment @lhoestq ,\r\nJust for confirmation, changing dummy data like this won't make dummy test test the functionality to extract `subsetxxx.xz` but actually kind of circumvent it. But since we will test the real data so it is ok ?", "Yes it's fine for now. We plan to add a job for slow tests.\r\nAnd at one point we'll also do another pass on the dummy data handling and consider extracting files.", "Thanks for the confirmation.\r\nAlso the suggestion works. Thank you." ]
1,600,513,623,000
1,600,775,650,000
1,600,775,649,000
CONTRIBUTOR
null
null
Hi, I recently want to add a dataset whose source data is like this ``` openwebtext.tar.xz |__ openwebtext |__subset000.xz | |__ ....txt | |__ ....txt | ... |__ subset001.xz | .... ``` So I wrote `openwebtext.py` like this ``` def _split_generators(self, dl_manager): dl_dir = dl_manager.download_and_extract(_URL) owt_dir = os.path.join(dl_dir, 'openwebtext') subset_xzs = [ os.path.join(owt_dir, file_name) for file_name in os.listdir(owt_dir) if file_name.endswith('xz') # filter out ...xz.lock ] ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count()*0.75)) nested_txt_files = [ [ os.path.join(ex_dir,txt_file_name) for txt_file_name in os.listdir(ex_dir) if txt_file_name.endswith('txt') ] for ex_dir in ex_dirs ] txt_files = chain(*nested_txt_files) return [ datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={"txt_files": txt_files} ), ] ``` All went good, I can load and use real openwebtext, except when I try to test with dummy data. The problem is `MockDownloadManager.extract` do nothing, so `ex_dirs = dl_manager.extract(subset_xzs)` won't decompress `subset_xxx.xz`s for me. How should I do ? Or you can modify `MockDownloadManager` to make it like a real `DownloadManager` ?
https://api.github.com/repos/huggingface/datasets/issues/650/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/649
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/649/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/649/comments
https://api.github.com/repos/huggingface/datasets/issues/649/events
https://github.com/huggingface/datasets/issues/649
704,838,415
MDU6SXNzdWU3MDQ4Mzg0MTU=
649
Inconsistent behavior in map
{ "login": "krandiash", "id": 10166085, "node_id": "MDQ6VXNlcjEwMTY2MDg1", "avatar_url": "https://avatars.githubusercontent.com/u/10166085?v=4", "gravatar_id": "", "url": "https://api.github.com/users/krandiash", "html_url": "https://github.com/krandiash", "followers_url": "https://api.github.com/users/krandiash/followers", "following_url": "https://api.github.com/users/krandiash/following{/other_user}", "gists_url": "https://api.github.com/users/krandiash/gists{/gist_id}", "starred_url": "https://api.github.com/users/krandiash/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/krandiash/subscriptions", "organizations_url": "https://api.github.com/users/krandiash/orgs", "repos_url": "https://api.github.com/users/krandiash/repos", "events_url": "https://api.github.com/users/krandiash/events{/privacy}", "received_events_url": "https://api.github.com/users/krandiash/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false } ]
null
[ "Thanks for reporting !\r\n\r\nThis issue must have appeared when we refactored type inference in `nlp`\r\nBy default the library tries to keep the same feature types when applying `map` but apparently it has troubles with nested structures. I'll try to fix that next week" ]
1,600,504,872,000
1,600,704,785,000
1,600,704,785,000
NONE
null
null
I'm observing inconsistent behavior when applying .map(). This happens specifically when I'm incrementally adding onto a feature that is a nested dictionary. Here's a simple example that reproduces the problem. ```python import datasets # Dataset with a single feature called 'field' consisting of two examples dataset = datasets.Dataset.from_dict({'field': ['a', 'b']}) print(dataset[0]) # outputs {'field': 'a'} # Map this dataset to create another feature called 'otherfield', which is a dictionary containing a key called 'capital' dataset = dataset.map(lambda example: {'otherfield': {'capital': example['field'].capitalize()}}) print(dataset[0]) # output is okay {'field': 'a', 'otherfield': {'capital': 'A'}} # Now I want to map again to modify 'otherfield', by adding another key called 'append_x' to the dictionary under 'otherfield' print(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x'}})[0]) # printing out the first example after applying the map shows that the new key 'append_x' doesn't get added # it also messes up the value stored at 'capital' {'field': 'a', 'otherfield': {'capital': None}} # Instead, I try to do the same thing by using a different mapped fn print(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['otherfield']['capital']}})[0]) # this preserves the value under capital, but still no 'append_x' {'field': 'a', 'otherfield': {'capital': 'A'}} # Instead, I try to pass 'otherfield' to remove_columns print(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['otherfield']['capital']}}, remove_columns=['otherfield'])[0]) # this still doesn't fix the problem {'field': 'a', 'otherfield': {'capital': 'A'}} # Alternately, here's what happens if I just directly map both 'capital' and 'append_x' on a fresh dataset. # Recreate the dataset dataset = datasets.Dataset.from_dict({'field': ['a', 'b']}) # Now map the entire 'otherfield' dict directly, instead of incrementally as before print(dataset.map(lambda example: {'otherfield': {'append_x': example['field'] + 'x', 'capital': example['field'].capitalize()}})[0]) # This looks good! {'field': 'a', 'otherfield': {'append_x': 'ax', 'capital': 'A'}} ``` This might be a new issue, because I didn't see this behavior in the `nlp` library. Any help is appreciated!
https://api.github.com/repos/huggingface/datasets/issues/649/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/648
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/648/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/648/comments
https://api.github.com/repos/huggingface/datasets/issues/648/events
https://github.com/huggingface/datasets/issues/648
704,753,123
MDU6SXNzdWU3MDQ3NTMxMjM=
648
offset overflow when multiprocessing batched map on large datasets.
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gists_url": "https://api.github.com/users/richarddwang/gists{/gist_id}", "starred_url": "https://api.github.com/users/richarddwang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/richarddwang/subscriptions", "organizations_url": "https://api.github.com/users/richarddwang/orgs", "repos_url": "https://api.github.com/users/richarddwang/repos", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "received_events_url": "https://api.github.com/users/richarddwang/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "This should be fixed with #645 ", "Feel free to re-open if it still occurs" ]
1,600,481,711,000
1,600,534,027,000
1,600,533,991,000
CONTRIBUTOR
null
null
It only happened when "multiprocessing" + "batched" + "large dataset" at the same time. ``` def bprocess(examples): examples['len'] = [] for text in examples['text']: examples['len'].append(len(text)) return examples wiki.map(brpocess, batched=True, num_proc=8) ``` ``` --------------------------------------------------------------------------- RemoteTraceback Traceback (most recent call last) RemoteTraceback: """ Traceback (most recent call last): File "/home/yisiang/miniconda3/envs/ml/lib/python3.7/multiprocessing/pool.py", line 121, in worker result = (True, func(*args, **kwds)) File "/home/yisiang/datasets/src/datasets/arrow_dataset.py", line 153, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/yisiang/datasets/src/datasets/fingerprint.py", line 163, in wrapper out = func(self, *args, **kwargs) File "/home/yisiang/datasets/src/datasets/arrow_dataset.py", line 1486, in _map_single batch = self[i : i + batch_size] File "/home/yisiang/datasets/src/datasets/arrow_dataset.py", line 1071, in __getitem__ format_kwargs=self._format_kwargs, File "/home/yisiang/datasets/src/datasets/arrow_dataset.py", line 972, in _getitem data_subset = self._data.take(indices_array) File "pyarrow/table.pxi", line 1145, in pyarrow.lib.Table.take File "/home/yisiang/miniconda3/envs/ml/lib/python3.7/site-packages/pyarrow/compute.py", line 268, in take return call_function('take', [data, indices], options) File "pyarrow/_compute.pyx", line 298, in pyarrow._compute.call_function File "pyarrow/_compute.pyx", line 192, in pyarrow._compute.Function.call File "pyarrow/error.pxi", line 122, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 84, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays """ The above exception was the direct cause of the following exception: ArrowInvalid Traceback (most recent call last) in 30 owt = datasets.load_dataset('/home/yisiang/datasets/datasets/openwebtext/openwebtext.py', cache_dir='./datasets')['train'] 31 print('load/create data from OpenWebText Corpus for ELECTRA') ---> 32 e_owt = ELECTRAProcessor(owt, apply_cleaning=False).map(cache_file_name=f"electra_owt_{c.max_length}.arrow") 33 dsets.append(e_owt) 34 ~/Reexamine_Attention/electra_pytorch/_utils/utils.py in map(self, **kwargs) 126 writer_batch_size=10**4, 127 num_proc=num_proc, --> 128 **kwargs 129 ) 130 ~/hugdatafast/hugdatafast/transform.py in my_map(self, *args, **kwargs) 21 if not cache_file_name.endswith('.arrow'): cache_file_name += '.arrow' 22 if '/' not in cache_file_name: cache_file_name = os.path.join(self.cache_directory(), cache_file_name) ---> 23 return self.map(*args, cache_file_name=cache_file_name, **kwargs) 24 25 @patch ~/datasets/src/datasets/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint) 1285 logger.info("Spawning {} processes".format(num_proc)) 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard] -> 1287 transformed_shards = [r.get() for r in results] 1288 logger.info("Concatenating {} shards from multiprocessing".format(num_proc)) 1289 result = concatenate_datasets(transformed_shards) ~/datasets/src/datasets/arrow_dataset.py in (.0) 1285 logger.info("Spawning {} processes".format(num_proc)) 1286 results = [pool.apply_async(self.__class__._map_single, kwds=kwds) for kwds in kwds_per_shard] -> 1287 transformed_shards = [r.get() for r in results] 1288 logger.info("Concatenating {} shards from multiprocessing".format(num_proc)) 1289 result = concatenate_datasets(transformed_shards) ~/miniconda3/envs/ml/lib/python3.7/multiprocessing/pool.py in get(self, timeout) 655 return self._value 656 else: --> 657 raise self._value 658 659 def _set(self, i, obj): ArrowInvalid: offset overflow while concatenating arrays ```
https://api.github.com/repos/huggingface/datasets/issues/648/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/647
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/647/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/647/comments
https://api.github.com/repos/huggingface/datasets/issues/647/events
https://github.com/huggingface/datasets/issues/647
704,734,764
MDU6SXNzdWU3MDQ3MzQ3NjQ=
647
Cannot download dataset_info.json
{ "login": "chiyuzhang94", "id": 33407613, "node_id": "MDQ6VXNlcjMzNDA3NjEz", "avatar_url": "https://avatars.githubusercontent.com/u/33407613?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chiyuzhang94", "html_url": "https://github.com/chiyuzhang94", "followers_url": "https://api.github.com/users/chiyuzhang94/followers", "following_url": "https://api.github.com/users/chiyuzhang94/following{/other_user}", "gists_url": "https://api.github.com/users/chiyuzhang94/gists{/gist_id}", "starred_url": "https://api.github.com/users/chiyuzhang94/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chiyuzhang94/subscriptions", "organizations_url": "https://api.github.com/users/chiyuzhang94/orgs", "repos_url": "https://api.github.com/users/chiyuzhang94/repos", "events_url": "https://api.github.com/users/chiyuzhang94/events{/privacy}", "received_events_url": "https://api.github.com/users/chiyuzhang94/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nWe should add support for servers without internet connection indeed\r\nI'll do that early next week", "Thanks, @lhoestq !\r\nPlease let me know when it is available. ", "Right now the recommended way is to create the dataset on a server with internet connection and then to save it and copy the serialized dataset to the server without internet connection.", "#652 should allow you to load text/json/csv/pandas datasets without an internet connection **IF** you've the dataset script locally.\r\n\r\nExample: \r\nIf you have `datasets/text/text.py` locally, then you can do `load_dataset(\"./datasets/text\", data_files=...)`" ]
1,600,479,315,000
1,600,676,922,000
1,600,676,922,000
NONE
null
null
I am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this: ``` ConnectionError: Couldn't reach https://storage.googleapis.com/huggingface-nlp/cache/datasets/text/default-53ee3045f07ba8ca/0.0.0/dataset_info.json ``` I tried to open this link manually, but I cannot access this file. How can I download this file and pass it through `dataset.load_dataset()` manually? Versions: Python version 3.7.3 PyTorch version 1.6.0 TensorFlow version 2.3.0 datasets version: 1.0.1
https://api.github.com/repos/huggingface/datasets/issues/647/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/646
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/646/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/646/comments
https://api.github.com/repos/huggingface/datasets/issues/646/events
https://github.com/huggingface/datasets/pull/646
704,607,371
MDExOlB1bGxSZXF1ZXN0NDg5NTAyMTM3
646
Fix docs typos
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,457,547,000
1,600,705,854,000
1,600,704,852,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/646", "html_url": "https://github.com/huggingface/datasets/pull/646", "diff_url": "https://github.com/huggingface/datasets/pull/646.diff", "patch_url": "https://github.com/huggingface/datasets/pull/646.patch" }
This PR fixes few typos in the docs and the error in the code snippet in the set_format section in docs/source/torch_tensorflow.rst. `torch.utils.data.Dataloader` expects padded batches so it throws an error due to not being able to stack the unpadded tensors. If we follow the Quick tour from the docs where they add the `truncation=True, padding='max_length'` arguments to the tokenizer before passing data to Dataloader, we can easily fix the issue.
https://api.github.com/repos/huggingface/datasets/issues/646/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/645
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/645/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/645/comments
https://api.github.com/repos/huggingface/datasets/issues/645/events
https://github.com/huggingface/datasets/pull/645
704,542,234
MDExOlB1bGxSZXF1ZXN0NDg5NDQ5MjAx
645
Don't use take on dataset table in pyarrow 1.0.x
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I tried lower batch sizes and it didn't accelerate filter (quite the opposite actually).\r\nThe slow-down also appears for pyarrow 0.17.1 for some reason, not sure it comes from these changes", "I just checked the benchmarks of other PRs and some of them had 300s (!!) for filter. This needs some investigation..", "Merging this one since it's not the cause of the the slow down", "@lhoestq What might be the reason for the slowdown? When I was training large batchsize, the slowdown was obvious, and my original assumption was to increase write_batch_size" ]
1,600,450,294,000
1,600,533,992,000
1,600,533,991,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/645", "html_url": "https://github.com/huggingface/datasets/pull/645", "diff_url": "https://github.com/huggingface/datasets/pull/645.diff", "patch_url": "https://github.com/huggingface/datasets/pull/645.patch" }
Fix #615
https://api.github.com/repos/huggingface/datasets/issues/645/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/644
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/644/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/644/comments
https://api.github.com/repos/huggingface/datasets/issues/644/events
https://github.com/huggingface/datasets/pull/644
704,534,501
MDExOlB1bGxSZXF1ZXN0NDg5NDQzMTk1
644
Better windows support
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "This PR is ready :)\r\nIt brings official support for windows.\r\n\r\nSome tests `AWSDatasetTest` are failing.\r\nThis is because I had to fix a few datasets that were not compatible with windows.\r\nThese test will pass once they got merged on master :)" ]
1,600,449,456,000
1,601,042,550,000
1,601,042,548,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/644", "html_url": "https://github.com/huggingface/datasets/pull/644", "diff_url": "https://github.com/huggingface/datasets/pull/644.diff", "patch_url": "https://github.com/huggingface/datasets/pull/644.patch" }
There are a few differences in the behavior of python and pyarrow on windows. For example there are restrictions when accessing/deleting files that are open Fix #590
https://api.github.com/repos/huggingface/datasets/issues/644/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/643
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/643/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/643/comments
https://api.github.com/repos/huggingface/datasets/issues/643/events
https://github.com/huggingface/datasets/issues/643
704,477,164
MDU6SXNzdWU3MDQ0NzcxNjQ=
643
Caching processed dataset at wrong folder
{ "login": "mrm8488", "id": 3653789, "node_id": "MDQ6VXNlcjM2NTM3ODk=", "avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mrm8488", "html_url": "https://github.com/mrm8488", "followers_url": "https://api.github.com/users/mrm8488/followers", "following_url": "https://api.github.com/users/mrm8488/following{/other_user}", "gists_url": "https://api.github.com/users/mrm8488/gists{/gist_id}", "starred_url": "https://api.github.com/users/mrm8488/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mrm8488/subscriptions", "organizations_url": "https://api.github.com/users/mrm8488/orgs", "repos_url": "https://api.github.com/users/mrm8488/repos", "events_url": "https://api.github.com/users/mrm8488/events{/privacy}", "received_events_url": "https://api.github.com/users/mrm8488/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[ "Thanks for reporting !\r\nIt uses a temporary file to write the data.\r\nHowever it looks like the temporary file is not placed in the right directory during the processing", "Well actually I just tested and the temporary file is placed in the same directory, so it should work as expected.\r\nWhich version of `datasets` are you using ?", "`datasets-1.0.1`\r\nHere you can reproduce it here:\r\nhttps://colab.research.google.com/drive/1O0KcepTFsmpkBbrbLLMq42iwTKmQh8d5?usp=sharing\r\n", "It looks like a pyarrow issue with google colab.\r\nFor some reason this code increases the disk usage of google colab while it actually writes into google drive:\r\n\r\n```python\r\nimport pyarrow as pa\r\n\r\nstream = pa.OSFile(\"/content/drive/My Drive/path/to/file.arrow\", \"wb\")\r\nwriter = pa.RecordBatchStreamWriter(stream, schema=pa.schema({\"text\": pa.string()}))\r\nwriter.write_table(pa.Table.from_pydict({\"text\": [\"a\"*511 + \"\\n\"] * ((1 << 30) // 512)})) # 1GiB\r\nwriter.close()\r\nstream.close()\r\n```\r\n\r\nMoreover if I `rm` the file on google drive, it frees disk space on google colab.", "It looks like replacing `pa.OSFile` by `open` fixes it, I'm going to open a PR", "Ok. Thank you so much!", "Actually I did more tests it doesn't >.<\r\nI'll let you know if I find a way to fix that", "Actually I also have the issue when writing a regular text file\r\n\r\n```python\r\nf = open(\"/content/drive/My Drive/path/to/file\", \"w\")\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) // 512)) # 1GiB\r\nf.close()\r\n```\r\n\r\nIs that supposed to happen ?", "The code you wrote should write a 1GB file in the Google Drive folder. Doesn't it? ", "Yes it does, but the disk usage of google colab also increases by 1GB", "I could check it and as you say as I write to te Drive disk the colab disk also increases...", "To reproduce it: \r\n```bash\r\n!df -h | grep sda1\r\n```\r\n```python\r\nf = open(\"/content/drive/My Drive/test_to_remove.txt\", \"w\")\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) // 512)) # 1GiB\r\nf.write((\"a\"*511 + \"\\n\") * ((1 << 30) // 512)) # 1GiB\r\nf.close()\r\n```\r\n```bash\r\n!ls -lh /content/drive/My\\ Drive/test_to_remove.txt\r\n\r\n!df -h | grep sda1\r\n\r\n!rm -rf /content/drive/My\\ Drive/test_to_remove.txt\r\n\r\n```\r\n[Colab](https://colab.research.google.com/drive/1D0UiweCYQwwWZ65EEhuqqbaDDbhJYXfm?usp=sharing)\r\n\r\n\r\n", "Apparently, Colab uses a local cache of the data files read/written from Google Drive. See:\r\n- https://github.com/googlecolab/colabtools/issues/2087#issuecomment-860818457\r\n- https://github.com/googlecolab/colabtools/issues/1915#issuecomment-804234540\r\n- https://github.com/googlecolab/colabtools/issues/2147#issuecomment-885052636" ]
1,600,443,686,000
1,601,309,680,000
null
NONE
null
null
Hi guys, I run this on my Colab (PRO): ```python from datasets import load_dataset dataset = load_dataset('text', data_files='/content/corpus.txt', cache_dir='/content/drive/My Drive', split='train') def encode(examples): return tokenizer(examples['text'], truncation=True, padding='max_length') dataset = dataset.map(encode, batched=True) ``` The file is about 4 GB, so I cannot process it on the Colab HD because there is no enough space. So I decided to mount my Google Drive fs and do it on it. The dataset is cached in the right place but by processing it (applying `encode` function) seems to use a different folder because Colab HD starts to grow and it crashes when it should be done in the Drive fs. What gets me crazy, it prints it is processing/encoding the dataset in the right folder: ``` Testing the mapped function outputs Testing finished, running the mapping function on the dataset Caching processed dataset at /content/drive/My Drive/text/default-ad3e69d6242ee916/0.0.0/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7/cache-b16341780a59747d.arrow ```
https://api.github.com/repos/huggingface/datasets/issues/643/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/642
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/642/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/642/comments
https://api.github.com/repos/huggingface/datasets/issues/642/events
https://github.com/huggingface/datasets/pull/642
704,397,499
MDExOlB1bGxSZXF1ZXN0NDg5MzMwMDAx
642
Rename wnut fields
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,437,091,000
1,600,449,511,000
1,600,449,510,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/642", "html_url": "https://github.com/huggingface/datasets/pull/642", "diff_url": "https://github.com/huggingface/datasets/pull/642.diff", "patch_url": "https://github.com/huggingface/datasets/pull/642.patch" }
As mentioned in #641 it would be cool to have it follow the naming of the other NER datasets
https://api.github.com/repos/huggingface/datasets/issues/642/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/641
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/641/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/641/comments
https://api.github.com/repos/huggingface/datasets/issues/641/events
https://github.com/huggingface/datasets/pull/641
704,373,940
MDExOlB1bGxSZXF1ZXN0NDg5MzExOTU3
641
Add Polyglot-NER Dataset
{ "login": "joeddav", "id": 9353833, "node_id": "MDQ6VXNlcjkzNTM4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joeddav", "html_url": "https://github.com/joeddav", "followers_url": "https://api.github.com/users/joeddav/followers", "following_url": "https://api.github.com/users/joeddav/following{/other_user}", "gists_url": "https://api.github.com/users/joeddav/gists{/gist_id}", "starred_url": "https://api.github.com/users/joeddav/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joeddav/subscriptions", "organizations_url": "https://api.github.com/users/joeddav/orgs", "repos_url": "https://api.github.com/users/joeddav/repos", "events_url": "https://api.github.com/users/joeddav/events{/privacy}", "received_events_url": "https://api.github.com/users/joeddav/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi @joeddav thanks for adding this! (I did a long webarchive.org session to actually find that dataset a while ago).\r\n\r\nOne question: should we manually correct the labeling scheme to (at least) IOB1?\r\n\r\nThat means \"LOC\" will be converted to \"I-LOC\". IOB1 is not explict. mentioned in the paper, but it is used in the documentation:\r\n\r\nhttps://polyglot.readthedocs.io/en/latest/NamedEntityRecognition.html", "@stefan-it I went back and forth on this. My biggest problem with it is that once you are in IOB, there is the expectation that the beginning of new entities are marked with a `B-` (at least in the case of two back-to-back entities):\r\n```\r\nToday O\r\nAlice I-PER\r\nBob B-PER\r\nand O\r\nI O \r\nate O\r\nlasagna O\r\n```\r\nIf we just prepend `I-` to everything, `Bob` would be incorrectly tagged `I-PER`, meaning `Bob Alice` is a single entity. The current format is bad but is at least clear that it does not contain that information.\r\n\r\nBut I could go either way if someone has a strong opinion.", "Indeed I'm not sure we can convert them to IOB because of this issue. I'm fine with keeping it like that", "I'll do a release later today, hopefully we can include this dataset in the release :)\r\n\r\nLet me know if you need help with the dummy data", "@lhoestq cool thanks, I think I've got it right now – just zipped them wrong. I'm running tests locally now and then will push.", "@lhoestq set to merge?", "@joeddav I'm fine with keeping the original labeling scheme :) " ]
1,600,435,304,000
1,600,571,083,000
1,600,571,083,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/641", "html_url": "https://github.com/huggingface/datasets/pull/641", "diff_url": "https://github.com/huggingface/datasets/pull/641.diff", "patch_url": "https://github.com/huggingface/datasets/pull/641.patch" }
Adds the [Polyglot-NER dataset](https://sites.google.com/site/rmyeid/projects/polylgot-ner) with named entity tags for 40 languages. I include separate configs for each language as well as a `combined` config which lumps them all together.
https://api.github.com/repos/huggingface/datasets/issues/641/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/640
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/640/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/640/comments
https://api.github.com/repos/huggingface/datasets/issues/640/events
https://github.com/huggingface/datasets/pull/640
704,311,758
MDExOlB1bGxSZXF1ZXN0NDg5MjYwNTc1
640
Make shuffle compatible with temp_seed
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,429,138,000
1,600,429,671,000
1,600,429,670,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/640", "html_url": "https://github.com/huggingface/datasets/pull/640", "diff_url": "https://github.com/huggingface/datasets/pull/640.diff", "patch_url": "https://github.com/huggingface/datasets/pull/640.patch" }
This code used to return different dataset at each run ```python import dataset as ds dataset = ... with ds.temp_seed(42): shuffled = dataset.shuffle() ``` Now it returns the same one since the seed is set
https://api.github.com/repos/huggingface/datasets/issues/640/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/639
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/639/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/639/comments
https://api.github.com/repos/huggingface/datasets/issues/639/events
https://github.com/huggingface/datasets/pull/639
704,217,963
MDExOlB1bGxSZXF1ZXN0NDg5MTgxOTY3
639
Update glue QQP checksum
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,420,095,000
1,600,429,028,000
1,600,429,027,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/639", "html_url": "https://github.com/huggingface/datasets/pull/639", "diff_url": "https://github.com/huggingface/datasets/pull/639.diff", "patch_url": "https://github.com/huggingface/datasets/pull/639.patch" }
Fix #638
https://api.github.com/repos/huggingface/datasets/issues/639/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/638
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/638/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/638/comments
https://api.github.com/repos/huggingface/datasets/issues/638/events
https://github.com/huggingface/datasets/issues/638
704,146,956
MDU6SXNzdWU3MDQxNDY5NTY=
638
GLUE/QQP dataset: NonMatchingChecksumError
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gists_url": "https://api.github.com/users/richarddwang/gists{/gist_id}", "starred_url": "https://api.github.com/users/richarddwang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/richarddwang/subscriptions", "organizations_url": "https://api.github.com/users/richarddwang/orgs", "repos_url": "https://api.github.com/users/richarddwang/repos", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "received_events_url": "https://api.github.com/users/richarddwang/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Hi ! Sure I'll take a look" ]
1,600,412,950,000
1,600,429,027,000
1,600,429,027,000
CONTRIBUTOR
null
null
Hi @lhoestq , I know you are busy and there are also other important issues. But if this is easy to be fixed, I am shamelessly wondering if you can give me some help , so I can evaluate my models and restart with my developing cycle asap. 😚 datasets version: editable install of master at 9/17 `datasets.load_dataset('glue','qqp', cache_dir='./datasets')` ``` Downloading and preparing dataset glue/qqp (download: 57.73 MiB, generated: 107.02 MiB, post-processed: Unknown size, total: 164.75 MiB) to ./datasets/glue/qqp/1.0.0/7c99657241149a24692c402a5c3f34d4c9f1df5ac2e4c3759fadea38f6cb29c4... --------------------------------------------------------------------------- NonMatchingChecksumError Traceback (most recent call last) in ----> 1 datasets.load_dataset('glue','qqp', cache_dir='./datasets') ~/datasets/src/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs) 609 download_config=download_config, 610 download_mode=download_mode, --> 611 ignore_verifications=ignore_verifications, 612 ) 613 ~/datasets/src/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs) 467 if not downloaded_from_gcs: 468 self._download_and_prepare( --> 469 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 470 ) 471 # Sync info ~/datasets/src/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 527 if verify_infos: 528 verify_checksums( --> 529 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), "dataset source files" 530 ) 531 ~/datasets/src/datasets/utils/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name) 37 if len(bad_urls) > 0: 38 error_msg = "Checksums didn't match" + for_verification_name + ":\n" ---> 39 raise NonMatchingChecksumError(error_msg + str(bad_urls)) 40 logger.info("All the checksums matched successfully" + for_verification_name) 41 NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://dl.fbaipublicfiles.com/glue/data/QQP-clean.zip'] ```
https://api.github.com/repos/huggingface/datasets/issues/638/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/637
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/637/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/637/comments
https://api.github.com/repos/huggingface/datasets/issues/637/events
https://github.com/huggingface/datasets/pull/637
703,539,909
MDExOlB1bGxSZXF1ZXN0NDg4NjMwNzk4
637
Add MATINF
{ "login": "JetRunner", "id": 22514219, "node_id": "MDQ6VXNlcjIyNTE0MjE5", "avatar_url": "https://avatars.githubusercontent.com/u/22514219?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JetRunner", "html_url": "https://github.com/JetRunner", "followers_url": "https://api.github.com/users/JetRunner/followers", "following_url": "https://api.github.com/users/JetRunner/following{/other_user}", "gists_url": "https://api.github.com/users/JetRunner/gists{/gist_id}", "starred_url": "https://api.github.com/users/JetRunner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JetRunner/subscriptions", "organizations_url": "https://api.github.com/users/JetRunner/orgs", "repos_url": "https://api.github.com/users/JetRunner/repos", "events_url": "https://api.github.com/users/JetRunner/events{/privacy}", "received_events_url": "https://api.github.com/users/JetRunner/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,345,493,000
1,600,348,998,000
1,600,348,997,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/637", "html_url": "https://github.com/huggingface/datasets/pull/637", "diff_url": "https://github.com/huggingface/datasets/pull/637.diff", "patch_url": "https://github.com/huggingface/datasets/pull/637.patch" }
https://api.github.com/repos/huggingface/datasets/issues/637/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/636
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/636/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/636/comments
https://api.github.com/repos/huggingface/datasets/issues/636/events
https://github.com/huggingface/datasets/pull/636
702,883,989
MDExOlB1bGxSZXF1ZXN0NDg4MDg3OTA5
636
Consistent ner features
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,271,785,000
1,600,336,379,000
1,600,336,378,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/636", "html_url": "https://github.com/huggingface/datasets/pull/636", "diff_url": "https://github.com/huggingface/datasets/pull/636.diff", "patch_url": "https://github.com/huggingface/datasets/pull/636.patch" }
As discussed in #613 , this PR aims at making NER feature names consistent across datasets. I changed the feature names of LinCE and XTREME/PAN-X
https://api.github.com/repos/huggingface/datasets/issues/636/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/635
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/635/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/635/comments
https://api.github.com/repos/huggingface/datasets/issues/635/events
https://github.com/huggingface/datasets/pull/635
702,822,439
MDExOlB1bGxSZXF1ZXN0NDg4MDM2OTE5
635
Loglevel
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "I think it's ready now @stas00, did you want to add something else ?\r\nThis PR includes your changes but with the level set to warning", "LGTM, thank you, @lhoestq " ]
1,600,267,073,000
1,600,336,339,000
1,600,336,338,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/635", "html_url": "https://github.com/huggingface/datasets/pull/635", "diff_url": "https://github.com/huggingface/datasets/pull/635.diff", "patch_url": "https://github.com/huggingface/datasets/pull/635.patch" }
Continuation of #618
https://api.github.com/repos/huggingface/datasets/issues/635/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/634
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/634/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/634/comments
https://api.github.com/repos/huggingface/datasets/issues/634/events
https://github.com/huggingface/datasets/pull/634
702,676,041
MDExOlB1bGxSZXF1ZXN0NDg3OTEzOTk4
634
Add ConLL-2000 dataset
{ "login": "vblagoje", "id": 458335, "node_id": "MDQ6VXNlcjQ1ODMzNQ==", "avatar_url": "https://avatars.githubusercontent.com/u/458335?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vblagoje", "html_url": "https://github.com/vblagoje", "followers_url": "https://api.github.com/users/vblagoje/followers", "following_url": "https://api.github.com/users/vblagoje/following{/other_user}", "gists_url": "https://api.github.com/users/vblagoje/gists{/gist_id}", "starred_url": "https://api.github.com/users/vblagoje/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vblagoje/subscriptions", "organizations_url": "https://api.github.com/users/vblagoje/orgs", "repos_url": "https://api.github.com/users/vblagoje/repos", "events_url": "https://api.github.com/users/vblagoje/events{/privacy}", "received_events_url": "https://api.github.com/users/vblagoje/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,254,851,000
1,600,339,090,000
1,600,339,090,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/634", "html_url": "https://github.com/huggingface/datasets/pull/634", "diff_url": "https://github.com/huggingface/datasets/pull/634.diff", "patch_url": "https://github.com/huggingface/datasets/pull/634.patch" }
Adds ConLL-2000 dataset used for text chunking. See https://www.clips.uantwerpen.be/conll2000/chunking/ for details and [motivation](https://github.com/huggingface/transformers/pull/7041#issuecomment-692710948) behind this PR
https://api.github.com/repos/huggingface/datasets/issues/634/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/633
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/633/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/633/comments
https://api.github.com/repos/huggingface/datasets/issues/633/events
https://github.com/huggingface/datasets/issues/633
702,440,484
MDU6SXNzdWU3MDI0NDA0ODQ=
633
Load large text file for LM pre-training resulting in OOM
{ "login": "leethu2012", "id": 29704017, "node_id": "MDQ6VXNlcjI5NzA0MDE3", "avatar_url": "https://avatars.githubusercontent.com/u/29704017?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leethu2012", "html_url": "https://github.com/leethu2012", "followers_url": "https://api.github.com/users/leethu2012/followers", "following_url": "https://api.github.com/users/leethu2012/following{/other_user}", "gists_url": "https://api.github.com/users/leethu2012/gists{/gist_id}", "starred_url": "https://api.github.com/users/leethu2012/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leethu2012/subscriptions", "organizations_url": "https://api.github.com/users/leethu2012/orgs", "repos_url": "https://api.github.com/users/leethu2012/repos", "events_url": "https://api.github.com/users/leethu2012/events{/privacy}", "received_events_url": "https://api.github.com/users/leethu2012/received_events", "type": "User", "site_admin": false }
[]
open
false
null
[]
null
[ "Not sure what could cause that on the `datasets` side. Could this be a `Trainer` issue ? cc @julien-c @sgugger ?", "There was a memory leak issue fixed recently in master. You should install from source and see if it fixes your problem.", "@lhoestq @sgugger Thanks for your comments. I have install from source code as you told, but the problem is still there.\r\nTo reproduce the issue, just replace [these lines](https://github.com/huggingface/transformers/blob/master/examples/language-modeling/run_language_modeling.py#L241-L258) with: \r\n(load_dataset and DataCollatorForDatasetsLanguageModeling as [above mentioned](https://github.com/huggingface/datasets/issues/633#issue-702440484))\r\n```python\r\n dataset = load_dataset(\"bookcorpus\")\r\n dataset = dataset.train_test_split(test_size=0.1)\r\n train_dataset = dataset['train']\r\n eval_dataset = dataset['test'] if training_args.do_eval else None\r\n\r\n data_collator = DataCollatorForDatasetsLanguageModeling(\r\n tokenizer=tokenizer,\r\n mlm=data_args.mlm,\r\n mlm_probability=data_args.mlm_probability,\r\n block_size=data_args.block_size\r\n )\r\n```\r\nand run by:\r\n```bash\r\npython run_language_modeling.py\r\n--output_dir=output \\\r\n--model_type=bert \\\r\n--model_name_or_path=bert-base-uncased \\\r\n--do_train \\\r\n--do_eval \\\r\n--mlm \r\n```", "Same here. Pre-training on wikitext-103 to do some test. At the end of the training it takes 32GB of RAM + ~30GB of SWAP. I installed dataset==1.1.0, not built from source. I will try uninstalling and building from source when it finish.", "This seems to be on the `transformers` library side.\r\n\r\nIf you have more informations (pip env) or even better, a colab reproducing the error we can investigate.", "It seems like it's solved with freshed versions of transformers. I have tried to replicate the error doing a fresh pip install transformers & datasets on colab and the error doesn't continue. On colab it keeps stable on 5GB! (Y)\r\n\r\nEdit: **Thanks for your great work**. Have a good day.", "@gaceladri witch version transformers and datasets are you using now? I want to try again. Thanks.", "transformers==3.3.1\r\ndatasets==1.1.0\r\ntokenizers==0.8.1rc2\r\n", "doing some modifications to mobilebert\r\nhttps://colab.research.google.com/drive/1ba09ZOpyHGAOQLcsxiQAHRXl10qnMU5o?usp=sharing ", "It does not happen to me anymore. Can we close? @leethu2012 ", "It's happening to me again. After 4 hours of pre-training, my ram memory gets full and the kernel dies. I am using the last transformers version as today. 4.4.0 and the last version of datasets 1.2.1, both installed from master. The memory consumption keeps increasing.", "It looks like it is something from pytorch/python itself :face_with_head_bandage: https://github.com/pytorch/pytorch/issues/13246 ", "Thanks for the investigation @gaceladri \r\n\r\nApparently this happens when `num_workers>0` and has to do with objects being copied-on-write.\r\nDid you try setting num_workers to 0 @gaceladri ?\r\nIf the issue doesn't happen with `num_workers=0` then this would confirm that it's indeed related to this python/pytorch issue.\r\n\r\nSince a `Dataset` object is a wrapper of a pyarrow Table, we should investigate if the data being copied comes from the Table itself or from metadata in the `Dataset` object. If it comes from the metadata in the `Dataset` object, we should be able to implement a workaround. But if it comes from the Table, we'll need to see with the pyarrow team what we can do... ", "@lhoestq I have tried and it keeps increasing also with `dataloader_num_workers=0`", "Hmmm so this might come from another issue...\r\nSince it doesn't seem to be related to multiprocessing it should be easier to investigate though.\r\nDo you have some ideas @gaceladri ?", "@lhoestq I looked quickly to a previously spoted bug in my env wandb /sdk/interface/interface.py, because sometimes when I load the dataset I got a multiprocessing error at line 510 in wandb...interface.py\r\n\r\nThis bug is reported here https://github.com/huggingface/datasets/issues/847\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nAssertionError Traceback (most recent call last)\r\n<timed eval> in <module>\r\n\r\n~/anaconda3/envs/tfm/lib/python3.6/site-packages/transformers/trainer.py in train(self, model_path, trial)\r\n 877 print(len(epoch_iterator))\r\n 878 \r\n--> 879 for step, inputs in enumerate(epoch_iterator):\r\n 880 \r\n 881 start_step = time.time()\r\n\r\n~/anaconda3/envs/tfm/lib/python3.6/site-packages/torch/utils/data/dataloader.py in __next__(self)\r\n 433 if self._sampler_iter is None:\r\n 434 self._reset()\r\n--> 435 data = self._next_data()\r\n 436 self._num_yielded += 1\r\n 437 if self._dataset_kind == _DatasetKind.Iterable and \\\r\n\r\n~/anaconda3/envs/tfm/lib/python3.6/site-packages/torch/utils/data/dataloader.py in _next_data(self)\r\n 1083 else:\r\n 1084 del self._task_info[idx]\r\n-> 1085 return self._process_data(data)\r\n 1086 \r\n 1087 def _try_put_index(self):\r\n\r\n~/anaconda3/envs/tfm/lib/python3.6/site-packages/torch/utils/data/dataloader.py in _process_data(self, data)\r\n 1109 self._try_put_index()\r\n 1110 if isinstance(data, ExceptionWrapper):\r\n-> 1111 data.reraise()\r\n 1112 return data\r\n 1113 \r\n\r\n~/anaconda3/envs/tfm/lib/python3.6/site-packages/torch/_utils.py in reraise(self)\r\n 426 # have message field\r\n 427 raise self.exc_type(message=msg)\r\n--> 428 raise self.exc_type(msg)\r\n 429 \r\n 430 \r\n\r\nAssertionError: Caught AssertionError in DataLoader worker process 0.\r\nOriginal Traceback (most recent call last):\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/torch/utils/data/_utils/worker.py\", line 198, in _worker_loop\r\n data = fetcher.fetch(index)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/torch/utils/data/_utils/fetch.py\", line 44, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/torch/utils/data/_utils/fetch.py\", line 44, in <listcomp>\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/datasets/arrow_dataset.py\", line 1083, in __getitem__\r\n format_kwargs=self._format_kwargs,\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/datasets/arrow_dataset.py\", line 1070, in _getitem\r\n format_kwargs=format_kwargs,\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/datasets/arrow_dataset.py\", line 886, in _convert_outputs\r\n v = map_nested(command, v, **map_nested_kwargs)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/datasets/utils/py_utils.py\", line 216, in map_nested\r\n return function(data_struct)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/datasets/arrow_dataset.py\", line 847, in command\r\n return torch.tensor(x, **format_kwargs)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/warnings.py\", line 101, in _showwarnmsg\r\n _showwarnmsg_impl(msg)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/warnings.py\", line 30, in _showwarnmsg_impl\r\n file.write(text)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/wandb/sdk/lib/redirect.py\", line 100, in new_write\r\n cb(name, data)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/wandb/sdk/wandb_run.py\", line 729, in _console_callback\r\n self._backend.interface.publish_output(name, data)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/wandb/sdk/interface/interface.py\", line 186, in publish_output\r\n self._publish_output(o)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/wandb/sdk/interface/interface.py\", line 191, in _publish_output\r\n self._publish(rec)\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/site-packages/wandb/sdk/interface/interface.py\", line 510, in _publish\r\n if self._process and not self._process.is_alive():\r\n File \"/home/ad/anaconda3/envs/tfm/lib/python3.6/multiprocessing/process.py\", line 134, in is_alive\r\n assert self._parent_pid == os.getpid(), 'can only test a child process'\r\nAssertionError: can only test a child process\r\n```\r\n\r\nMy workaround was to just comment those lines without looking to much into consecuences:\r\n\r\n```\r\ndef _publish(self, record: pb.Record, local: bool = None) -> None:\r\n #if self._process and not self._process.is_alive():\r\n # raise Exception(\"The wandb backend process has shutdown\")\r\n```\r\n\r\nIt worked so far... I need to try running without wandb and see if it could be causing something wrong with multiprocessing. I am going to try to launch the training setting wandb to false and I will let you know again.", "@lhoestq But despite this, I got lost into the [class Dataset()](https://huggingface.co/docs/datasets/_modules/datasets/arrow_dataset.html#Dataset) reading the pyarrow files.\r\n\r\nEdit: but you should be rigth, that it does not have to be related to multiprocessing since it keeps happening when `num_workers=0` ", "Or maybe wandb uses multiprocessing ? One process for wandb logging and one for actual training ? If this is the case then even setting `num_workers=0` would cause the process to be forked for wandb and therefore cause the memory issue.", "@lhoestq could be, but if we set wandb to false this should not happen. I am going to try.", "@lhoestq It keeps happening. I have uninstalled wandb from my env, setted `%env WANDB_DISABLED=true` on my notebook, and commented this func:\r\n\r\n```\r\ndef get_available_reporting_integrations():\r\n integrations = []\r\n if is_azureml_available():\r\n integrations.append(\"azure_ml\")\r\n if is_comet_available():\r\n integrations.append(\"comet_ml\")\r\n if is_mlflow_available():\r\n integrations.append(\"mlflow\")\r\n if is_tensorboard_available():\r\n integrations.append(\"tensorboard\")\r\n # if is_wandb_available():\r\n # integrations.append(\"wandb\")\r\n return integrations\r\n```\r\nAs a fast test and it keeps increasing the ram memory. Wandb could not be the blameworthy here.", "Thanks for checking @gaceladri . Let's investigate the single process setting then.\r\nIf you have some sort of colab notebook with a minimal code example that shows this behavior feel free to share it @gaceladri so that we can play around with it to find what causes this. Otherwise I'll probably try to reproduce on my side at one point", "@lhoestq sure. Here you have https://colab.research.google.com/drive/1ba09ZOpyHGAOQLcsxiQAHRXl10qnMU5o?usp=sharing let me know if the link works and it reproduces the issue. To me, it reproduces the issue, since if you start the training the ram memory keeps increasing.\r\n\r\nLet me know. Thanks!", "Could the bug be comming from tokenizers?\r\n\r\nI got this warning at the terminal from my jupyter notebook: \r\n```\r\nhuggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...\r\nTo disable this warning, you can either:\r\n\t- Avoid using `tokenizers` before the fork if possible\r\n\t- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)\r\n```", "I've never experienced memory issues with tokenizers so I don't know\r\nCc @n1t0 are you aware of any issue that would cause memory to keep increasing when the tokenizer is used in the Data Collator for language modeling ?", "@lhoestq Thanks for pointing to n1t0, just to clarify. That warning was doing fine-tuning, without collator:\r\n```\r\n\r\nfrom datasets import load_dataset, load_metric\r\nimport numpy as np\r\n\r\nGLUE_TASKS = [\r\n \"cola\",\r\n \"mnli\",\r\n \"mnli-mm\",\r\n \"mrpc\",\r\n \"qnli\",\r\n \"qqp\",\r\n \"rte\",\r\n \"sst2\",\r\n \"stsb\",\r\n \"wnli\",\r\n]\r\ntask = \"mnli\"\r\nactual_task = \"mnli\" if task == \"mnli-mm\" else task\r\ndataset = load_dataset(\"glue\", actual_task)\r\nmetric = load_metric(\"glue\", actual_task)\r\nbatch_size = 16\r\nattention_type = \"linear\"\r\n\r\nfrom transformers.models.mobilebert_mod import (\r\n MobileBertForSequenceClassification,\r\n MobileBertTokenizerFast,\r\n)\r\nfrom transformers.models.mobilebert_mod.configuration_mobilebert import (\r\n MobileBertConfigMod,\r\n)\r\nfrom transformers import TrainingArguments, Trainer\r\n\r\nnum_labels = 3 if task.startswith(\"mnli\") else 1 if task == \"stsb\" else 2\r\ntokenizer = MobileBertTokenizerFast.from_pretrained(\r\n \"/media/ad/00b5422b-9d54-4449-8b5d-08eab5cdac8c/training_trfm/big_linear_layerdrop_shared/checkpoint-23000/\",\r\n max_len=512,\r\n)\r\nmodel = MobileBertForSequenceClassification.from_pretrained(\r\n \"/media/ad/00b5422b-9d54-4449-8b5d-08eab5cdac8c/training_trfm/big_linear_layerdrop_shared/checkpoint-23000/\",\r\n num_labels=num_labels,\r\n)\r\nprint(model.num_parameters())\r\n\r\ntask_to_keys = {\r\n \"cola\": (\"sentence\", None),\r\n \"mnli\": (\"premise\", \"hypothesis\"),\r\n \"mnli-mm\": (\"premise\", \"hypothesis\"),\r\n \"mrpc\": (\"sentence1\", \"sentence2\"),\r\n \"qnli\": (\"question\", \"sentence\"),\r\n \"qqp\": (\"question1\", \"question2\"),\r\n \"rte\": (\"sentence1\", \"sentence2\"),\r\n \"sst2\": (\"sentence\", None),\r\n \"stsb\": (\"sentence1\", \"sentence2\"),\r\n \"wnli\": (\"sentence1\", \"sentence2\"),\r\n}\r\n\r\nsentence1_key, sentence2_key = task_to_keys[task]\r\nif sentence2_key is None:\r\n print(f\"Sentence: {dataset['train'][0][sentence1_key]}\")\r\nelse:\r\n print(f\"Sentence 1: {dataset['train'][0][sentence1_key]}\")\r\n print(f\"Sentence 2: {dataset['train'][0][sentence2_key]}\")\r\n\r\n\r\ndef preprocess_function(examples):\r\n if sentence2_key is None:\r\n return tokenizer(examples[sentence1_key], truncation=True)\r\n return tokenizer(examples[sentence1_key], examples[sentence2_key], truncation=True)\r\n\r\n\r\nencoded_dataset = dataset.map(preprocess_function, batched=True)\r\nmetric_name = (\r\n \"pearson\"\r\n if task == \"stsb\"\r\n else \"matthews_correlation\"\r\n if task == \"cola\"\r\n else \"accuracy\"\r\n)\r\n\r\nargs = TrainingArguments(\r\n f\"test-glue/{task}_{attention_type}\",\r\n evaluation_strategy=\"steps\",\r\n learning_rate=1e-5,\r\n per_device_train_batch_size=batch_size,\r\n per_device_eval_batch_size=batch_size,\r\n logging_steps=200,\r\n num_train_epochs=5,\r\n gradient_accumulation_steps=1,\r\n warmup_steps=10000,\r\n fp16=True,\r\n dataloader_num_workers=10,\r\n weight_decay=0.1,\r\n load_best_model_at_end=True,\r\n metric_for_best_model=metric_name,\r\n)\r\n\r\n\r\ndef compute_metrics(eval_pred):\r\n predictions, labels = eval_pred\r\n if task != \"stsb\":\r\n predictions = np.argmax(predictions, axis=1)\r\n else:\r\n predictions = predictions[:, 0]\r\n return metric.compute(predictions=predictions, references=labels)\r\n\r\n\r\nvalidation_key = (\r\n \"validation_mismatched\"\r\n if task == \"mnli-mm\"\r\n else \"validation_matched\"\r\n if task == \"mnli\"\r\n else \"validation\"\r\n)\r\n\r\ntrainer = Trainer(\r\n model,\r\n args,\r\n train_dataset=encoded_dataset[\"train\"],\r\n eval_dataset=encoded_dataset[validation_key],\r\n tokenizer=tokenizer,\r\n compute_metrics=compute_metrics,\r\n)\r\n\r\ntrainer.train()\r\n```\r\n\r\nNow, I have come back to pre-training. The changes that I think I have done are: not formatting the dataset to torch: ~~`big_dataset.set_format(type='torch', columns=[\"text\", \"input_ids\", \"attention_mask\", \"token_type_ids\"])`~~ so maybe some column is dropped and not freezed in memory and now I have not setted any validation dataset in the trainer. \r\n\r\nMy validation dataset before:\r\n```\r\nbook_corpus_eval = load_dataset(\r\n \"bookcorpus\",\r\n \"plain_text\",\r\n cache_dir=\"/home/ad/Desktop/bookcorpus\",\r\n split=\"train[98:99%]\",\r\n)\r\nbook_corpus_eval = book_corpus_eval.map(encode, batched=True)\r\nbook_corpus_eval.set_format(\r\n type=\"torch\", columns=[\"text\", \"input_ids\", \"attention_mask\", \"token_type_ids\"]\r\n)\r\n**book_corpus_eval = book_corpus_eval.select([i for i in range(1500)])**\r\n```\r\nMaybe _selecting_ or indexing the dataset before feeding it to the trainer, do something strange.\r\n\r\nMy trainer now:\r\n```\r\n\r\nbig_dataset = load_from_disk(\"/home/ad/Desktop/35percent_data.arrow/\")\r\n\r\nfrom transformers import DataCollatorForWholeWordMask\r\n\r\ndata_collator = DataCollatorForWholeWordMask(\r\n tokenizer=tokenizer, mlm=True, mlm_probability=0.15)\r\n\r\nfrom transformers import Trainer, TrainingArguments\r\n\r\ntraining_args = TrainingArguments(\r\n output_dir=\"./big_linear_layerdrop_shared_silu_secondtry\",\r\n overwrite_output_dir=True,\r\n per_device_train_batch_size=60,\r\n per_device_eval_batch_size=60,\r\n save_steps=500,\r\n save_total_limit=10,\r\n logging_first_step=True,\r\n logging_steps=100,\r\n# evaluation_strategy='steps',\r\n# eval_steps=250,\r\n gradient_accumulation_steps=8,\r\n fp16=True,\r\n dataloader_num_workers=10,\r\n warmup_steps=15000,\r\n learning_rate=6e-4,\r\n adam_epsilon=1e-6,\r\n adam_beta2=0.98,\r\n weight_decay=0.01,\r\n max_grad_norm=1.0,\r\n max_steps=500000, \r\n)\r\n\r\ntrainer = Trainer(\r\n model=model,\r\n args=training_args,\r\n data_collator=data_collator,\r\n train_dataset=big_dataset,\r\n# eval_dataset=book_corpus_eval,\r\n tokenizer=tokenizer)\r\n\r\nimport wandb\r\nwandb.login()\r\n\r\ntrainer.train()\r\n```\r\n\r\nAnd surprisingly, the ram now keeps going up and down. The training is up now for 12h without collapse the ram. I don't know what could cause the leakage. :mag: \r\n\r\nEdit: I didn't see the swap memory, that keeps increasing. So the problem persist. ", "Thanks for sharing your results.\r\nSo you still had the issue for fine-tuning ?\r\nAnd the issue still appears with a bare-bone dataset from an arrow file...", "Yes, on both cases. Fine-tuning a pre-trained model and pre-training from scratch with a local arrow file already pre-processed." ]
1,600,230,795,000
1,613,476,921,000
null
NONE
null
null
I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this: ```python from datasets import load_dataset @dataclass class DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling): """ Data collator used for language modeling based on DataCollatorForLazyLanguageModeling - collates batches of tensors, honoring their tokenizer's pad_token - preprocesses batches for masked language modeling """ block_size: int = 512 def __call__(self, examples: List[dict]) -> Dict[str, torch.Tensor]: examples = [example['text'] for example in examples] batch, attention_mask = self._tensorize_batch(examples) if self.mlm: inputs, labels = self.mask_tokens(batch) return {"input_ids": inputs, "labels": labels} else: labels = batch.clone().detach() if self.tokenizer.pad_token_id is not None: labels[labels == self.tokenizer.pad_token_id] = -100 return {"input_ids": batch, "labels": labels} def _tensorize_batch(self, examples: List[str]) -> Tuple[torch.Tensor, torch.Tensor]: if self.tokenizer._pad_token is None: raise ValueError( "You are attempting to pad samples but the tokenizer you are using" f" ({self.tokenizer.__class__.__name__}) does not have one." ) tensor_examples = self.tokenizer.batch_encode_plus( [ex for ex in examples if ex], max_length=self.block_size, return_tensors="pt", pad_to_max_length=True, return_attention_mask=True, truncation=True, ) input_ids, attention_mask = tensor_examples["input_ids"], tensor_examples["attention_mask"] return input_ids, attention_mask dataset = load_dataset('text', data_files='train.txt',cache_dir="./", , split='train') data_collator = DataCollatorForDatasetsLanguageModeling(tokenizer=tokenizer, mlm=True, mlm_probability=0.15, block_size=tokenizer.max_len) trainer = Trainer(model=model, args=args, data_collator=data_collator, train_dataset=train_dataset, prediction_loss_only=True, ) trainer.train(model_path=model_path) ``` This train.txt is about 1.1GB and has 90k lines where each line is a sequence of 4k words. During training, the memory usage increased fast as the following graph and resulted in OOM before the finish of training. ![image](https://user-images.githubusercontent.com/29704017/93292112-5576b280-f817-11ea-8da2-b2db9bf35665.png) Could you please give me any suggestions on why this happened and how to fix it? Thanks.
https://api.github.com/repos/huggingface/datasets/issues/633/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/632
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/632/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/632/comments
https://api.github.com/repos/huggingface/datasets/issues/632/events
https://github.com/huggingface/datasets/pull/632
702,358,124
MDExOlB1bGxSZXF1ZXN0NDg3NjQ5OTQ2
632
Fix typos in the loading datasets docs
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "repos_url": "https://api.github.com/users/mariosasko/repos", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "thanks!" ]
1,600,216,061,000
1,600,705,871,000
1,600,239,164,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/632", "html_url": "https://github.com/huggingface/datasets/pull/632", "diff_url": "https://github.com/huggingface/datasets/pull/632.diff", "patch_url": "https://github.com/huggingface/datasets/pull/632.patch" }
This PR fixes two typos in the loading datasets docs, one of them being a broken link to the `load_dataset` function.
https://api.github.com/repos/huggingface/datasets/issues/632/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/631
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/631/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/631/comments
https://api.github.com/repos/huggingface/datasets/issues/631/events
https://github.com/huggingface/datasets/pull/631
701,711,255
MDExOlB1bGxSZXF1ZXN0NDg3MTE3OTA0
631
Fix text delimiter
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "repos_url": "https://api.github.com/users/lhoestq/repos", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Which OS are you using ?@abhi1nandy2", "> Which OS are you using ?\r\n\r\nPRETTY_NAME=\"Debian GNU/Linux 9 (stretch)\"\r\nNAME=\"Debian GNU/Linux\"\r\nVERSION_ID=\"9\"\r\nVERSION=\"9 (stretch)\"\r\nVERSION_CODENAME=stretch\r\nID=debian\r\nHOME_URL=\"https://www.debian.org/\"\r\nSUPPORT_URL=\"https://www.debian.org/support\"\r\nBUG_REPORT_URL=\"https://bugs.debian.org/\"", "Do you mind sharing the data you used (or part of it), so I can try to reproduce ?\r\nOr at least some info about the text file you're using ? (size, n of lines, encoding)", "Lot of data, difficult to share. There are 46 shards, each having about 256000 lines. using `file` command gives this - `ASCII text, with very long lines`.", "Ok I see, no problem :) \r\nI'll see what I can do\r\n\r\nCould you just test with one single dummy text file with a few lines to see if you're having the issue ?\r\nAlso which version of `datasets` do you have ?" ]
1,600,157,322,000
1,600,786,986,000
1,600,158,385,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/631", "html_url": "https://github.com/huggingface/datasets/pull/631", "diff_url": "https://github.com/huggingface/datasets/pull/631.diff", "patch_url": "https://github.com/huggingface/datasets/pull/631.patch" }
I changed the delimiter in the `text` dataset script. It should fix the `pyarrow.lib.ArrowInvalid: CSV parse error` from #622 I changed the delimiter to an unused ascii character that is not present in text files : `\b`
https://api.github.com/repos/huggingface/datasets/issues/631/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/630
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/630/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/630/comments
https://api.github.com/repos/huggingface/datasets/issues/630/events
https://github.com/huggingface/datasets/issues/630
701,636,350
MDU6SXNzdWU3MDE2MzYzNTA=
630
Text dataset not working with large files
{ "login": "ksjae", "id": 17930170, "node_id": "MDQ6VXNlcjE3OTMwMTcw", "avatar_url": "https://avatars.githubusercontent.com/u/17930170?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ksjae", "html_url": "https://github.com/ksjae", "followers_url": "https://api.github.com/users/ksjae/followers", "following_url": "https://api.github.com/users/ksjae/following{/other_user}", "gists_url": "https://api.github.com/users/ksjae/gists{/gist_id}", "starred_url": "https://api.github.com/users/ksjae/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ksjae/subscriptions", "organizations_url": "https://api.github.com/users/ksjae/orgs", "repos_url": "https://api.github.com/users/ksjae/repos", "events_url": "https://api.github.com/users/ksjae/events{/privacy}", "received_events_url": "https://api.github.com/users/ksjae/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Seems like it works when setting ```block_size=2100000000``` or something arbitrarily large though.", "Can you give us some stats on the data files you use as inputs?", "Basically ~600MB txt files(UTF-8) * 59. \r\ncontents like ```안녕하세요, 이것은 예제로 한번 말해보는 텍스트입니다. 그냥 이렇다고요.<|endoftext|>\\n```\r\n\r\nAlso, it gets stuck for a loooong time at ```Testing the mapped function outputs```, for more than 12 hours(currently ongoing)", "It gets stuck while doing `.map()` ? Are you using multiprocessing ?\r\nIf you could provide a code snippet it could be very useful", "From transformers/examples/language-modeling/run-language-modeling.py :\r\n```\r\ndef get_dataset(\r\n args: DataTrainingArguments,\r\n tokenizer: PreTrainedTokenizer,\r\n evaluate: bool = False,\r\n cache_dir: Optional[str] = None,\r\n):\r\n file_path = args.eval_data_file if evaluate else args.train_data_file\r\n if True:\r\n dataset = load_dataset(\"text\", data_files=glob.glob(file_path), split='train', use_threads=True, \r\n ignore_verifications=True, save_infos=True, block_size=104857600)\r\n dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"], add_special_tokens=True,\r\n truncation=True, max_length=args.block_size), batched=True)\r\n dataset.set_format(type='torch', columns=['input_ids'])\r\n return dataset\r\n if args.line_by_line:\r\n return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size)\r\n else:\r\n return TextDataset(\r\n tokenizer=tokenizer,\r\n file_path=file_path,\r\n block_size=args.block_size,\r\n overwrite_cache=args.overwrite_cache,\r\n cache_dir=cache_dir,\r\n )\r\n```\r\n\r\nNo, I'm not using multiprocessing.", "I am not able to reproduce on my side :/\r\n\r\nCould you send the version of `datasets` and `pyarrow` you're using ?\r\nCould you try to update the lib and try again ?\r\nOr do you think you could try to reproduce it on google colab ?", "Huh, weird. It's fixed on my side too.\r\nBut now ```Caching processed dataset``` is taking forever - how can I disable it? Any flags?", "Right after `Caching processed dataset`, your function is applied to the dataset and there's a progress bar that shows how much time is left. How much time does it take for you ?\r\n\r\nAlso caching isn't supposed to slow down your processing. But if you still want to disable it you can do `.map(..., load_from_cache_file=False)`", "Ah, it’s much faster now(Takes around 15~20min). \r\nBTW, any way to set default tensor output as plain tensors with distributed training? The ragged tensors are incompatible with tpustrategy :(", "> Ah, it’s much faster now(Takes around 15~20min).\r\n\r\nGlad to see that it's faster now. What did you change exactly ?\r\n\r\n> BTW, any way to set default tensor output as plain tensors with distributed training? The ragged tensors are incompatible with tpustrategy :(\r\n\r\nOh I didn't know about that. Feel free to open an issue to mention that.\r\nI guess what you can do for now is set the dataset format to numpy instead of tensorflow, and use a wrapper of the dataset that converts the numpy arrays to tf tensors.\r\n\r\n", ">>> Glad to see that it's faster now. What did you change exactly ?\r\nI don't know, it just worked...? Sorry I couldn't be more helpful.\r\n\r\nSetting with numpy array is a great idea! Thanks." ]
1,600,149,756,000
1,601,072,503,000
1,601,072,503,000
NONE
null
null
``` Traceback (most recent call last): File "examples/language-modeling/run_language_modeling.py", line 333, in <module> main() File "examples/language-modeling/run_language_modeling.py", line 262, in main get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_train else None File "examples/language-modeling/run_language_modeling.py", line 144, in get_dataset dataset = load_dataset("text", data_files=file_path, split='train+test') File "/home/ksjae/.local/lib/python3.7/site-packages/datasets/load.py", line 611, in load_dataset ignore_verifications=ignore_verifications, File "/home/ksjae/.local/lib/python3.7/site-packages/datasets/builder.py", line 469, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "/home/ksjae/.local/lib/python3.7/site-packages/datasets/builder.py", line 546, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/ksjae/.local/lib/python3.7/site-packages/datasets/builder.py", line 888, in _prepare_split for key, table in utils.tqdm(generator, unit=" tables", leave=False, disable=not_verbose): File "/home/ksjae/.local/lib/python3.7/site-packages/tqdm/std.py", line 1129, in __iter__ for obj in iterable: File "/home/ksjae/.cache/huggingface/modules/datasets_modules/datasets/text/7e13bc0fa76783d4ef197f079dc8acfe54c3efda980f2c9adfab046ede2f0ff7/text.py", line 104, in _generate_tables convert_options=self.config.convert_options, File "pyarrow/_csv.pyx", line 714, in pyarrow._csv.read_csv File "pyarrow/error.pxi", line 122, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 84, in pyarrow.lib.check_status ``` **pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)** It gives the same message for both 200MB, 10GB .tx files but not for 700MB file. Can't upload due to size & copyright problem. sorry.
https://api.github.com/repos/huggingface/datasets/issues/630/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/629
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/629/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/629/comments
https://api.github.com/repos/huggingface/datasets/issues/629/events
https://github.com/huggingface/datasets/issues/629
701,517,550
MDU6SXNzdWU3MDE1MTc1NTA=
629
straddling object straddles two block boundaries
{ "login": "bharaniabhishek123", "id": 17970177, "node_id": "MDQ6VXNlcjE3OTcwMTc3", "avatar_url": "https://avatars.githubusercontent.com/u/17970177?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bharaniabhishek123", "html_url": "https://github.com/bharaniabhishek123", "followers_url": "https://api.github.com/users/bharaniabhishek123/followers", "following_url": "https://api.github.com/users/bharaniabhishek123/following{/other_user}", "gists_url": "https://api.github.com/users/bharaniabhishek123/gists{/gist_id}", "starred_url": "https://api.github.com/users/bharaniabhishek123/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bharaniabhishek123/subscriptions", "organizations_url": "https://api.github.com/users/bharaniabhishek123/orgs", "repos_url": "https://api.github.com/users/bharaniabhishek123/repos", "events_url": "https://api.github.com/users/bharaniabhishek123/events{/privacy}", "received_events_url": "https://api.github.com/users/bharaniabhishek123/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "sorry it's an apache arrow issue." ]
1,600,129,846,000
1,600,130,177,000
1,600,129,937,000
NONE
null
null
I am trying to read json data (it's an array with lots of dictionaries) and getting block boundaries issue as below : I tried calling read_json with readOptions but no luck . ``` table = json.read_json(fn) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "pyarrow/_json.pyx", line 246, in pyarrow._json.read_json File "pyarrow/error.pxi", line 122, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 84, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?) ```
https://api.github.com/repos/huggingface/datasets/issues/629/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/628
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/628/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/628/comments
https://api.github.com/repos/huggingface/datasets/issues/628/events
https://github.com/huggingface/datasets/pull/628
701,496,053
MDExOlB1bGxSZXF1ZXN0NDg2OTQyNzgx
628
Update docs links in the contribution guideline
{ "login": "M-Salti", "id": 9285264, "node_id": "MDQ6VXNlcjkyODUyNjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/9285264?v=4", "gravatar_id": "", "url": "https://api.github.com/users/M-Salti", "html_url": "https://github.com/M-Salti", "followers_url": "https://api.github.com/users/M-Salti/followers", "following_url": "https://api.github.com/users/M-Salti/following{/other_user}", "gists_url": "https://api.github.com/users/M-Salti/gists{/gist_id}", "starred_url": "https://api.github.com/users/M-Salti/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/M-Salti/subscriptions", "organizations_url": "https://api.github.com/users/M-Salti/orgs", "repos_url": "https://api.github.com/users/M-Salti/repos", "events_url": "https://api.github.com/users/M-Salti/events{/privacy}", "received_events_url": "https://api.github.com/users/M-Salti/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Thanks!" ]
1,600,126,039,000
1,604,351,003,000
1,600,150,775,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/628", "html_url": "https://github.com/huggingface/datasets/pull/628", "diff_url": "https://github.com/huggingface/datasets/pull/628.diff", "patch_url": "https://github.com/huggingface/datasets/pull/628.patch" }
Fixed the `add a dataset` and `share a dataset` links in the contribution guideline to refer to the new docs website.
https://api.github.com/repos/huggingface/datasets/issues/628/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/627
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/627/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/627/comments
https://api.github.com/repos/huggingface/datasets/issues/627/events
https://github.com/huggingface/datasets/pull/627
701,411,661
MDExOlB1bGxSZXF1ZXN0NDg2ODcxMTg2
627
fix (#619) MLQA features names
{ "login": "M-Salti", "id": 9285264, "node_id": "MDQ6VXNlcjkyODUyNjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/9285264?v=4", "gravatar_id": "", "url": "https://api.github.com/users/M-Salti", "html_url": "https://github.com/M-Salti", "followers_url": "https://api.github.com/users/M-Salti/followers", "following_url": "https://api.github.com/users/M-Salti/following{/other_user}", "gists_url": "https://api.github.com/users/M-Salti/gists{/gist_id}", "starred_url": "https://api.github.com/users/M-Salti/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/M-Salti/subscriptions", "organizations_url": "https://api.github.com/users/M-Salti/orgs", "repos_url": "https://api.github.com/users/M-Salti/repos", "events_url": "https://api.github.com/users/M-Salti/events{/privacy}", "received_events_url": "https://api.github.com/users/M-Salti/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,116,119,000
1,604,351,072,000
1,600,239,251,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/627", "html_url": "https://github.com/huggingface/datasets/pull/627", "diff_url": "https://github.com/huggingface/datasets/pull/627.diff", "patch_url": "https://github.com/huggingface/datasets/pull/627.patch" }
Fixed the features names as suggested in (#619) in the `_generate_examples` and `_info` methods in the MLQA loading script and also changed the names in the `dataset_infos.json` file.
https://api.github.com/repos/huggingface/datasets/issues/627/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/626
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/626/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/626/comments
https://api.github.com/repos/huggingface/datasets/issues/626/events
https://github.com/huggingface/datasets/pull/626
701,352,605
MDExOlB1bGxSZXF1ZXN0NDg2ODIzMTY1
626
Update GLUE URLs (now hosted on FB)
{ "login": "jeswan", "id": 57466294, "node_id": "MDQ6VXNlcjU3NDY2Mjk0", "avatar_url": "https://avatars.githubusercontent.com/u/57466294?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jeswan", "html_url": "https://github.com/jeswan", "followers_url": "https://api.github.com/users/jeswan/followers", "following_url": "https://api.github.com/users/jeswan/following{/other_user}", "gists_url": "https://api.github.com/users/jeswan/gists{/gist_id}", "starred_url": "https://api.github.com/users/jeswan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jeswan/subscriptions", "organizations_url": "https://api.github.com/users/jeswan/orgs", "repos_url": "https://api.github.com/users/jeswan/repos", "events_url": "https://api.github.com/users/jeswan/events{/privacy}", "received_events_url": "https://api.github.com/users/jeswan/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[]
1,600,110,339,000
1,600,239,198,000
1,600,239,198,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/626", "html_url": "https://github.com/huggingface/datasets/pull/626", "diff_url": "https://github.com/huggingface/datasets/pull/626.diff", "patch_url": "https://github.com/huggingface/datasets/pull/626.patch" }
NYU is switching dataset hosting from Google to FB. This PR closes https://github.com/huggingface/datasets/issues/608 and is necessary for https://github.com/jiant-dev/jiant/issues/161. This PR updates the data URLs based on changes made in https://github.com/nyu-mll/jiant/pull/1112. Note: rebased on huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/626/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/625
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/625/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/625/comments
https://api.github.com/repos/huggingface/datasets/issues/625/events
https://github.com/huggingface/datasets/issues/625
701,057,799
MDU6SXNzdWU3MDEwNTc3OTk=
625
dtype of tensors should be preserved
{ "login": "BramVanroy", "id": 2779410, "node_id": "MDQ6VXNlcjI3Nzk0MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BramVanroy", "html_url": "https://github.com/BramVanroy", "followers_url": "https://api.github.com/users/BramVanroy/followers", "following_url": "https://api.github.com/users/BramVanroy/following{/other_user}", "gists_url": "https://api.github.com/users/BramVanroy/gists{/gist_id}", "starred_url": "https://api.github.com/users/BramVanroy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BramVanroy/subscriptions", "organizations_url": "https://api.github.com/users/BramVanroy/orgs", "repos_url": "https://api.github.com/users/BramVanroy/repos", "events_url": "https://api.github.com/users/BramVanroy/events{/privacy}", "received_events_url": "https://api.github.com/users/BramVanroy/received_events", "type": "User", "site_admin": false }
[]
closed
false
null
[]
null
[ "Indeed we convert tensors to list to be able to write in arrow format. Because of this conversion we lose the dtype information. We should add the dtype detection when we do type inference. However it would require a bit of refactoring since currently the conversion happens before the type inference..\r\n\r\nAnd then for your information, when reading from arrow format we have to cast from arrow to numpy (which is fast since pyarrow has a numpy integration), and then to torch.\r\n\r\nHowever there's one thing that can help you: we make sure that the dtypes correspond to what is defined in `features`.\r\nTherefore what you can do is provide `features` in `.map(preprocess, feature=...)` to specify the output types.\r\n\r\nFor example in your case:\r\n```python\r\nfrom datasets import Features, Value, Sequence\r\n\r\nfeatures = Features({\r\n \"input_ids\": Sequence(Value(\"int32\")),\r\n \"sembedding\": Sequence(Value(\"float32\"))\r\n})\r\npreprocessed_dataset = dataset.map(preprocess, features=features)\r\n\r\npreprocessed_dataset.set_format(\"torch\", columns=[\"input_ids\", \"sembedding\"])\r\nprint(preprocessed_dataset[0][\"sembedding\"].dtype)\r\n# \"torch.float32\"\r\n```\r\n\r\nLet me know if it helps", "If the arrow format is basically lists, why is the intermediate step to numpy necessary? I am a bit confused about that part.\r\n\r\nThanks for your suggestion. as I have currently implemented this, I cast to torch.Tensor in my collate_fn to save disk space (so I do not have to save padded tensors to max_len but can pad up to max batch len in collate_fn) at the cost of a bit slower processing. So for me this is not relevant anymore, but I am sure it is for others!", "I'm glad you managed to figure something out :)\r\n\r\nCasting from arrow to numpy can be 100x faster than casting from arrow to list.\r\nThis is because arrow has an integration with numpy that allows it to instantiate numpy arrays with zero-copy from arrow.\r\nOn the other hand to create python lists it is slow since it has to recreate the list object by iterating through each element in python.", "Ah that is interesting. I have no direct experience with arrow so I didn't know. ", "I encountered a simliar issue: `datasets` converted my float numpy array to `torch.float64` tensors, while many pytorch operations require `torch.float32` inputs and it's very troublesome. \r\n\r\nI tried @lhoestq 's solution, but since it's mixed with the preprocess function, it's not very intuitive. \r\n\r\nI just want to share another possible simpler solution: directly cast the dtype of the processed dataset.\r\n\r\nNow I want to change the type of `labels` in `train_dataset` from float64 to float32, I can do this.\r\n\r\n```\r\nfrom datasets import Value, Sequence, Features\r\nfeats = train_dataset.features.copy()\r\nfeats['labels'].feature = Value(dtype='float32')\r\nfeats = Features(feats)\r\ntrain_dataset.cast_(feats)\r\n```\r\n", "Reopening since @bhavitvyamalik started looking into it !\r\n\r\nAlso I'm posting here a function that could be helpful to support preserving the dtype of tensors.\r\n\r\nIt's used to build a pyarrow array out of a numpy array and:\r\n- it doesn't convert the numpy array to a python list\r\n- it keeps the precision of the numpy array for the pyarrow array\r\n- it works with multidimensional arrays (while `pa.array` can only take a 1D array as input)\r\n- it builds the pyarrow ListArray from offsets created on-the-fly and values that come from the flattened numpy array\r\n\r\n```python\r\nfrom functools import reduce\r\nfrom operator import mul\r\n\r\nimport numpy as np\r\nimport pyarrow as pa\r\n\r\ndef pa_ndarray(a):\r\n \"\"\"Build a PyArrow ListArray from a multidimensional NumPy array\"\"\"\r\n values = pa.array(a.flatten()) \r\n for i in range(a.ndim - 1): \r\n n_offsets = reduce(mul, a.shape[:a.ndim - i - 1], 1) \r\n step_offsets = a.shape[a.ndim - i - 1] \r\n offsets = pa.array(np.arange(n_offsets + 1) * step_offsets, type=pa.int32()) \r\n values = pa.ListArray.from_arrays(offsets, values) \r\n return values \r\n\r\nnarr = np.arange(42).reshape(7, 2, 3).astype(np.uint8)\r\nparr = pa_ndarray(narr)\r\nassert isinstance(parr, pa.Array)\r\nassert parr.type == pa.list_(pa.list_(pa.uint8()))\r\nassert narr.tolist() == parr.to_pylist()\r\n```\r\n\r\nThe only costly operation is the offsets computations. Since it doesn't iterate on the numpy array values this function is pretty fast.", "@lhoestq Have you thought about this further?\r\n\r\nWe have a use case where we're attempting to load data containing numpy arrays using the `datasets` library.\r\n\r\nWhen using one of the \"standard\" methods (`[Value(...)]` or `Sequence()`) we see ~200 samples processed per second during the call to `_prepare_split`. This slowdown is caused by the vast number of calls to `encode_nested_example` (each sequence is converted to a list, and each element in the sequence...). \r\n\r\nUsing the `Feature` `ArrayND` improves this somewhat to ~500/s as it now uses numpy's `tolist()` rather than iterating over each value in the array and converting them individually.\r\n\r\nHowever, it's still pretty slow and in theory it should be possible to avoid the `numpy -> python -> arrow` dance altogether. To demonstrate this, if you keep the `Feature` set to an `ArrayND` but instead return a `pa_ndarray(...)` in `_generate_examples` it skips the conversion (`return obj, False`) and hits ~11_000/s. Two orders of magnitude speed up! The problem is this then fails later on when the `ArrowWriter` tries to write the examples to disk :-( \r\n\r\nIt would be nice to have first-class support for user-defined PyArrow objects. Is this a possibility? We have _large_ datasets where even an order of magnitude difference is important so settling on the middle ~500/s is less than ideal! \r\n\r\nIs there a workaround for this or another method that should be used instead that gets near-to or equal performance to returning PyArrow arrays?", "Note that manually generating the table using `pyarrow` achieves ~30_000/s", "Hi !\r\n\r\nIt would be awesome to achieve this speed for numpy arrays !\r\nFor now we have to use `encode_nested_example` to convert numpy arrays to python lists since pyarrow doesn't support multidimensional numpy arrays (only 1D).\r\n\r\nMaybe let's start a new PR from your PR @bhavitvyamalik (idk why we didn't answer your PR at that time, sorry about that).\r\nBasically the idea is to allow `TypedSequence` to support numpy arrays as you did, and remove the numpy->python casting in `_cast_to_python_objects`.\r\n\r\nThis is really important since we are starting to have a focus on other modalities than text as well (audio, images).\r\n\r\nThough until then @samgd, there is another feature that may interest you and that may give you the speed you want:\r\n\r\nIn a dataset script you can subclass either a GeneratorBasedBuilder (with the `_generate_examples ` method) or an ArrowBasedBuilder if you want. the ArrowBasedBuilder allows to yield arrow data by implementing the `_generate_tables` method (it's the same as `_generate_examples` except you must yield arrow tables). Since the data are already in arrow format, it doesn't call `encode_nested_example`. Let me know if that helps." ]
1,600,087,085,000
1,629,189,004,000
1,629,189,004,000
CONTRIBUTOR
null
null
After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https://discuss.pytorch.org/t/is-it-required-that-input-and-hidden-for-gru-have-the-same-dtype-float32/96221)). As a user I did not expect this bug. I have a `map` function that I call on the Dataset that looks like this: ```python def preprocess(sentences: List[str]): token_ids = [[vocab.to_index(t) for t in s.split()] for s in sentences] sembeddings = stransformer.encode(sentences) print(sembeddings.dtype) return {"input_ids": token_ids, "sembedding": sembeddings} ``` Given a list of `sentences` (`List[str]`), it converts those into token_ids on the one hand (list of lists of ints; `List[List[int]]`) and into sentence embeddings on the other (Tensor of dtype `torch.float32`). That means that I actually set the column "sembedding" to a tensor that I as a user expect to be a float32. It appears though that behind the scenes, this tensor is converted into a **list**. I did not find this documented anywhere but I might have missed it. From a user's perspective this is incredibly important though, because it means you cannot do any data_type or tensor casting yourself in a mapping function! Furthermore, this can lead to issues, as was my case. My model expected float32 precision, which I thought `sembedding` was because that is what `stransformer.encode` outputs. But behind the scenes this tensor is first cast to a list, and when we then set its format, as below, this column is cast not to float32 but to double precision float64. ```python dataset.set_format(type="torch", columns=["input_ids", "sembedding"]) ``` This happens because apparently there is an intermediate step of casting to a **numpy** array (?) **whose dtype creation/deduction is different from torch dtypes** (see the snippet below). As you can see, this means that the dtype is not preserved: if I got it right, the dataset goes from torch.float32 -> list -> float64 (numpy) -> torch.float64. ```python import torch import numpy as np l = [-0.03010837361216545, -0.035979013890028, -0.016949838027358055] torch_tensor = torch.tensor(l) np_array = np.array(l) np_to_torch = torch.from_numpy(np_array) print(torch_tensor.dtype) # torch.float32 print(np_array.dtype) # float64 print(np_to_torch.dtype) # torch.float64 ``` This might lead to unwanted behaviour. I understand that the whole library is probably built around casting from numpy to other frameworks, so this might be difficult to solve. Perhaps `set_format` should include a `dtypes` option where for each input column the user can specify the wanted precision. The alternative is that the user needs to cast manually after loading data from the dataset but that does not seem user-friendly, makes the dataset less portable, and might use more space in memory as well as on disk than is actually needed.
https://api.github.com/repos/huggingface/datasets/issues/625/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/624
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/624/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/624/comments
https://api.github.com/repos/huggingface/datasets/issues/624/events
https://github.com/huggingface/datasets/issues/624
700,541,628
MDU6SXNzdWU3MDA1NDE2Mjg=
624
Add learningq dataset
{ "login": "krrishdholakia", "id": 17561003, "node_id": "MDQ6VXNlcjE3NTYxMDAz", "avatar_url": "https://avatars.githubusercontent.com/u/17561003?v=4", "gravatar_id": "", "url": "https://api.github.com/users/krrishdholakia", "html_url": "https://github.com/krrishdholakia", "followers_url": "https://api.github.com/users/krrishdholakia/followers", "following_url": "https://api.github.com/users/krrishdholakia/following{/other_user}", "gists_url": "https://api.github.com/users/krrishdholakia/gists{/gist_id}", "starred_url": "https://api.github.com/users/krrishdholakia/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/krrishdholakia/subscriptions", "organizations_url": "https://api.github.com/users/krrishdholakia/orgs", "repos_url": "https://api.github.com/users/krrishdholakia/repos", "events_url": "https://api.github.com/users/krrishdholakia/events{/privacy}", "received_events_url": "https://api.github.com/users/krrishdholakia/received_events", "type": "User", "site_admin": false }
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[]
1,599,992,427,000
1,600,077,002,000
null
NONE
null
null
Hi, Thank you again for this amazing repo. Would it be possible for y'all to add the LearningQ dataset - https://github.com/AngusGLChen/LearningQ ?
https://api.github.com/repos/huggingface/datasets/issues/624/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/623
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/623/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/623/comments
https://api.github.com/repos/huggingface/datasets/issues/623/events
https://github.com/huggingface/datasets/issues/623
700,235,308
MDU6SXNzdWU3MDAyMzUzMDg=
623
Custom feature types in `load_dataset` from CSV
{ "login": "lvwerra", "id": 8264887, "node_id": "MDQ6VXNlcjgyNjQ4ODc=", "avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lvwerra", "html_url": "https://github.com/lvwerra", "followers_url": "https://api.github.com/users/lvwerra/followers", "following_url": "https://api.github.com/users/lvwerra/following{/other_user}", "gists_url": "https://api.github.com/users/lvwerra/gists{/gist_id}", "starred_url": "https://api.github.com/users/lvwerra/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lvwerra/subscriptions", "organizations_url": "https://api.github.com/users/lvwerra/orgs", "repos_url": "https://api.github.com/users/lvwerra/repos", "events_url": "https://api.github.com/users/lvwerra/events{/privacy}", "received_events_url": "https://api.github.com/users/lvwerra/received_events", "type": "User", "site_admin": false }
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
[ "Currently `csv` doesn't support the `features` attribute (unlike `json`).\r\nWhat you can do for now is cast the features using the in-place transform `cast_`\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'])\r\ndataset.cast_(emotion_features)\r\n```\r\n", "Thanks for the clarification!", "Hi @lhoestq we've tried out your suggestion but are now running into the following error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nValueError Traceback (most recent call last)\r\n<ipython-input-163-81ffd5ac18c9> in <module>\r\n----> 1 dataset.cast_(emotion_features)\r\n\r\n/usr/local/lib/python3.6/dist-packages/datasets/dataset_dict.py in cast_(self, features)\r\n 125 self._check_values_type()\r\n 126 for dataset in self.values():\r\n--> 127 dataset.cast_(features=features)\r\n 128 \r\n 129 def remove_columns_(self, column_names: Union[str, List[str]]):\r\n\r\n/usr/local/lib/python3.6/dist-packages/datasets/fingerprint.py in wrapper(*args, **kwargs)\r\n 161 # Call actual function\r\n 162 \r\n--> 163 out = func(self, *args, **kwargs)\r\n 164 \r\n 165 # Update fingerprint of in-place transforms + update in-place history of transforms\r\n\r\n/usr/local/lib/python3.6/dist-packages/datasets/arrow_dataset.py in cast_(self, features)\r\n 602 self._info.features = features\r\n 603 schema = pa.schema(features.type)\r\n--> 604 self._data = self._data.cast(schema)\r\n 605 \r\n 606 @fingerprint(inplace=True)\r\n\r\n/usr/local/lib/python3.6/dist-packages/pyarrow/table.pxi in pyarrow.lib.Table.cast()\r\n\r\nValueError: Target schema's field names are not matching the table's field names: ['text', 'label'], ['label', 'text']\r\n```\r\n\r\nLooking at the types in `emotion_features` we see that `label` and `text` appear to be swapped in the Arrow table:\r\n\r\n```\r\nemotion_features.type\r\nStructType(struct<label: int64, text: string>)\r\n```\r\n\r\nDid we define the `emotion_features` incorrectly? We just followed the instructions from the [docs](https://huggingface.co/docs/datasets/features.html?highlight=features#dataset-features), but perhaps we misunderstood something 😬 \r\n\r\n", "In general, I don't think there is any hard reason we don't allow to use `features` in the csv script, right @lhoestq?\r\n\r\nShould I add it?", "> In general, I don't think there is any hard reason we don't allow to use `features` in the csv script, right @lhoestq?\r\n> \r\n> Should I add it?\r\n\r\nSure let's add it. Setting the convert options should do the job\r\n\r\n> Hi @lhoestq we've tried out your suggestion but are now running into the following error:\r\n> \r\n> ```\r\n> ---------------------------------------------------------------------------\r\n> ValueError Traceback (most recent call last)\r\n> <ipython-input-163-81ffd5ac18c9> in <module>\r\n> ----> 1 dataset.cast_(emotion_features)\r\n>\r\n> /usr/local/lib/python3.6/dist-packages/pyarrow/table.pxi in pyarrow.lib.Table.cast()\r\n> \r\n> ValueError: Target schema's field names are not matching the table's field names: ['text', 'label'], ['label', 'text']\r\n> ```\r\n>\r\n> Did we define the `emotion_features` incorrectly? We just followed the instructions from the [docs](https://huggingface.co/docs/datasets/features.html?highlight=features#dataset-features), but perhaps we misunderstood something 😬\r\n\r\nThanks for reporting, that's a bug :) I'm fixing it right now", "PR is open for the `ValueError: Target schema's field names are not matching the table's field names` error.\r\n\r\nI'm adding the features parameter to csv", "Thanks a lot for the PR and quick fix @lhoestq!" ]
1,599,916,894,000
1,601,495,503,000
1,601,455,194,000
CONTRIBUTOR
null
null
I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. I am working with the local files from the emotion dataset. To get the data you can use the following code: ```Python from pathlib import Path import wget EMOTION_PATH = Path("./data/emotion") DOWNLOAD_URLS = [ "https://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt?dl=1", "https://www.dropbox.com/s/2mzialpsgf9k5l3/val.txt?dl=1", "https://www.dropbox.com/s/ikkqxfdbdec3fuj/test.txt?dl=1", ] if not Path.is_dir(EMOTION_PATH): Path.mkdir(EMOTION_PATH) for url in DOWNLOAD_URLS: wget.download(url, str(EMOTION_PATH)) ``` The first five lines of the train set are: ``` i didnt feel humiliated;sadness i can go from feeling so hopeless to so damned hopeful just from being around someone who cares and is awake;sadness im grabbing a minute to post i feel greedy wrong;anger i am ever feeling nostalgic about the fireplace i will know that it is still on the property;love i am feeling grouchy;anger ``` Here the code to reproduce the issue: ```Python from datasets import Features, Value, ClassLabel, load_dataset class_names = ["sadness", "joy", "love", "anger", "fear", "surprise"] emotion_features = Features({'text': Value('string'), 'label': ClassLabel(names=class_names)}) file_dict = {'train': EMOTION_PATH/'train.txt'} dataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label'], features=emotion_features) ``` **Observed behaviour:** ```Python dataset['train'].features ``` ```Python {'text': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)} ``` **Expected behaviour:** ```Python dataset['train'].features ``` ```Python {'text': Value(dtype='string', id=None), 'label': ClassLabel(num_classes=6, names=['sadness', 'joy', 'love', 'anger', 'fear', 'surprise'], names_file=None, id=None)} ``` **Things I've tried:** - deleting the cache - trying other types such as `int64` Am I missing anything? Thanks for any pointer in the right direction.
https://api.github.com/repos/huggingface/datasets/issues/623/timeline
null
false