id
int64
1.14B
2.23B
labels_url
stringlengths
75
75
body
stringlengths
2
33.9k
updated_at
stringlengths
20
20
number
int64
3.76k
6.79k
milestone
dict
repository_url
stringclasses
1 value
draft
bool
2 classes
labels
listlengths
0
4
created_at
stringlengths
20
20
comments_url
stringlengths
70
70
assignee
dict
timeline_url
stringlengths
70
70
title
stringlengths
1
290
events_url
stringlengths
68
68
active_lock_reason
null
user
dict
assignees
listlengths
0
3
performed_via_github_app
null
state_reason
stringclasses
3 values
author_association
stringclasses
3 values
closed_at
stringlengths
20
20
pull_request
dict
node_id
stringlengths
18
19
comments
sequencelengths
0
30
reactions
dict
state
stringclasses
2 values
locked
bool
1 class
url
stringlengths
61
61
html_url
stringlengths
49
51
is_pull_request
bool
2 classes
1,224,244,460
https://api.github.com/repos/huggingface/datasets/issues/4270/labels{/name}
CI in PR: - #4259 was green, but after merging it to master, a code quality error appeared.
2022-05-06T08:38:06Z
4,270
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-05-03T15:21:34Z
https://api.github.com/repos/huggingface/datasets/issues/4270/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4270/timeline
Fix style in openbookqa dataset
https://api.github.com/repos/huggingface/datasets/issues/4270/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-05-03T16:20:52Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4270.diff", "html_url": "https://github.com/huggingface/datasets/pull/4270", "merged_at": "2022-05-03T16:20:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/4270.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4270" }
PR_kwDODunzps43PC5V
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4270/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4270
https://github.com/huggingface/datasets/pull/4270
true
1,223,865,145
https://api.github.com/repos/huggingface/datasets/issues/4269/labels{/name}
Update metadata of big_patent dataset with: - license - point of contact
2022-05-06T08:38:09Z
4,269
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-05-03T09:24:07Z
https://api.github.com/repos/huggingface/datasets/issues/4269/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4269/timeline
Add license and point of contact to big_patent dataset
https://api.github.com/repos/huggingface/datasets/issues/4269/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-05-03T11:16:19Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4269.diff", "html_url": "https://github.com/huggingface/datasets/pull/4269", "merged_at": "2022-05-03T11:16:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/4269.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4269" }
PR_kwDODunzps43Nzwh
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4269/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4269
https://github.com/huggingface/datasets/pull/4269
true
1,223,331,964
https://api.github.com/repos/huggingface/datasets/issues/4268/labels{/name}
## Describe the bug Error generated when attempting to download dataset ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("bigscience-catalogue-lm-data/lm_en_wiktionary_filtered") ``` ## Expected results A clear and concise description of the expected results. ## Actual results ``` ExpectedMoreDownloadedFiles Traceback (most recent call last) [<ipython-input-62-4ac5cf959477>](https://localhost:8080/#) in <module>() 1 from datasets import load_dataset 2 ----> 3 dataset = load_dataset("bigscience-catalogue-lm-data/lm_en_wiktionary_filtered") 3 frames [/usr/local/lib/python3.7/dist-packages/datasets/utils/info_utils.py](https://localhost:8080/#) in verify_checksums(expected_checksums, recorded_checksums, verification_name) 31 return 32 if len(set(expected_checksums) - set(recorded_checksums)) > 0: ---> 33 raise ExpectedMoreDownloadedFiles(str(set(expected_checksums) - set(recorded_checksums))) 34 if len(set(recorded_checksums) - set(expected_checksums)) > 0: 35 raise UnexpectedDownloadedFile(str(set(recorded_checksums) - set(expected_checksums))) ExpectedMoreDownloadedFiles: {'/home/leandro/catalogue_data/datasets/lm_en_wiktionary_filtered/data/file-01.jsonl.gz', '/home/leandro/catalogue_data/datasets/lm_en_wiktionary_filtered/data/file-01.jsonl.gz.lock'} ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.18.3 - Platform: Linux-5.4.188+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - PyArrow version: 6.0.1
2022-05-06T15:53:30Z
4,268
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
2022-05-02T20:34:25Z
https://api.github.com/repos/huggingface/datasets/issues/4268/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4268/timeline
error downloading bigscience-catalogue-lm-data/lm_en_wiktionary_filtered
https://api.github.com/repos/huggingface/datasets/issues/4268/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/102043285?v=4", "events_url": "https://api.github.com/users/i-am-neo/events{/privacy}", "followers_url": "https://api.github.com/users/i-am-neo/followers", "following_url": "https://api.github.com/users/i-am-neo/following{/other_user}", "gists_url": "https://api.github.com/users/i-am-neo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/i-am-neo", "id": 102043285, "login": "i-am-neo", "node_id": "U_kgDOBhUOlQ", "organizations_url": "https://api.github.com/users/i-am-neo/orgs", "received_events_url": "https://api.github.com/users/i-am-neo/received_events", "repos_url": "https://api.github.com/users/i-am-neo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/i-am-neo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/i-am-neo/subscriptions", "type": "User", "url": "https://api.github.com/users/i-am-neo" }
[]
null
completed
NONE
2022-05-03T11:23:48Z
null
I_kwDODunzps5I6pB8
[ "It would help a lot to be able to preview the dataset - I'd like to see if the pronunciations are in the dataset, eg. for [\"word\"](https://en.wiktionary.org/wiki/word),\r\n\r\nPronunciation\r\n([Received Pronunciation](https://en.wikipedia.org/wiki/Received_Pronunciation)) [IPA](https://en.wiktionary.org/wiki/Wiktionary:International_Phonetic_Alphabet)([key](https://en.wiktionary.org/wiki/Appendix:English_pronunciation)): /wɜːd/\r\n([General American](https://en.wikipedia.org/wiki/General_American)) [enPR](https://en.wiktionary.org/wiki/Appendix:English_pronunciation): wûrd, [IPA](https://en.wiktionary.org/wiki/Wiktionary:International_Phonetic_Alphabet)([key](https://en.wiktionary.org/wiki/Appendix:English_pronunciation)): /wɝd/", "Hi @i-am-neo, thanks for reporting.\r\n\r\nNormally this dataset should be private and not accessible for public use. @cakiki, @lvwerra, any reason why is it public? I see many other Wikimedia datasets are also public.\r\n\r\nAlso note that last commit \"Add metadata\" (https://huggingface.co/datasets/bigscience-catalogue-lm-data/lm_en_wiktionary_filtered/commit/dc2f458dab50e00f35c94efb3cd4009996858609) introduced buggy data files (`data/file-01.jsonl.gz.lock`, `data/file-01.jsonl.gz.lock.lock`). The same bug appears in other datasets as well.\r\n\r\n@i-am-neo, please note that in the near future we are planning to make public all datasets used for the BigScience project (at least all of them whose license allows to do that). Once public, they will be accessible for all the NLP community.", "Ah this must be a bug introduced at creation time since the repos were created programmatically; I'll go ahead and make them private; sorry about that!", "All datasets are private now. \r\n\r\nRe:that bug I think we're currently avoiding it by avoiding verifications. (i.e. `ignore_verifications=True`)", "Thanks a lot, @cakiki.\r\n\r\n@i-am-neo, I'm closing this issue for now because the dataset is not publicly available yet. Just stay tuned, as we will soon release all the BigScience open-license datasets. ", "Thanks for letting me know, @albertvillanova @cakiki.\r\nAny chance of having a subset alpha version in the meantime? \r\nI only need two dicts out of wiktionary: 1) phoneme(as key): word, and 2) word(as key): its phonemes.\r\n\r\nWould like to use it for a mini-poc [Robust ASR](https://github.com/huggingface/transformers/issues/13162#issuecomment-1096881290) decoding, cc @patrickvonplaten. \r\n\r\n(Patrick, possible to email you so as not to litter github with comments? I have some observations after experiments training hubert on some YT AMI-like data (11.44% wer). Also wonder if a robust ASR is on your/HG's roadmap). Thanks!", "Hey @i-am-neo,\r\n\r\nCool to hear that you're working on Robust ASR! Feel free to drop me a mail :-)", "@i-am-neo This particular subset of the dataset was taken from the [CirrusSearch dumps](https://dumps.wikimedia.org/other/cirrussearch/current/)\r\nYou're specifically after the [enwiktionary-20220425-cirrussearch-content.json.gz](https://dumps.wikimedia.org/other/cirrussearch/current/enwiktionary-20220425-cirrussearch-content.json.gz) file", "thanks @cakiki ! <del>I could access the gz file yesterday (but neglected to tuck it away somewhere safe), and today the link is throwing a 404. Can you help? </del> Never mind, got it!", "thanks @patrickvonplaten. will do - getting my observations together." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4268/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4268
https://github.com/huggingface/datasets/issues/4268
false
1,223,214,275
https://api.github.com/repos/huggingface/datasets/issues/4267/labels{/name}
Replace data URL with one in the same repository.
2022-05-06T08:38:13Z
4,267
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-05-02T18:38:08Z
https://api.github.com/repos/huggingface/datasets/issues/4267/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4267/timeline
Replace data URL in SAMSum dataset within the same repository
https://api.github.com/repos/huggingface/datasets/issues/4267/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-05-02T19:03:49Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4267.diff", "html_url": "https://github.com/huggingface/datasets/pull/4267", "merged_at": "2022-05-02T19:03:49Z", "patch_url": "https://github.com/huggingface/datasets/pull/4267.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4267" }
PR_kwDODunzps43LzOR
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4267/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4267
https://github.com/huggingface/datasets/pull/4267
true
1,223,116,436
https://api.github.com/repos/huggingface/datasets/issues/4266/labels{/name}
Adds the HF Speech Bench to Librispeech Dataset Card in place of the Papers With Code Leaderboard. Should improve usage and visibility of this leaderboard! Wondering whether this can also be done for [Common Voice 7](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0) and [8](https://huggingface.co/datasets/mozilla-foundation/common_voice_8_0) through someone with permissions? cc @patrickvonplaten: more leaderboard promotion!
2022-05-05T08:47:20Z
4,266
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-05-02T16:59:31Z
https://api.github.com/repos/huggingface/datasets/issues/4266/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4266/timeline
Add HF Speech Bench to Librispeech Dataset Card
https://api.github.com/repos/huggingface/datasets/issues/4266/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4", "events_url": "https://api.github.com/users/sanchit-gandhi/events{/privacy}", "followers_url": "https://api.github.com/users/sanchit-gandhi/followers", "following_url": "https://api.github.com/users/sanchit-gandhi/following{/other_user}", "gists_url": "https://api.github.com/users/sanchit-gandhi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sanchit-gandhi", "id": 93869735, "login": "sanchit-gandhi", "node_id": "U_kgDOBZhWpw", "organizations_url": "https://api.github.com/users/sanchit-gandhi/orgs", "received_events_url": "https://api.github.com/users/sanchit-gandhi/received_events", "repos_url": "https://api.github.com/users/sanchit-gandhi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sanchit-gandhi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sanchit-gandhi/subscriptions", "type": "User", "url": "https://api.github.com/users/sanchit-gandhi" }
[]
null
null
CONTRIBUTOR
2022-05-05T08:40:09Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4266.diff", "html_url": "https://github.com/huggingface/datasets/pull/4266", "merged_at": "2022-05-05T08:40:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/4266.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4266" }
PR_kwDODunzps43LeXK
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4266/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4266
https://github.com/huggingface/datasets/pull/4266
true
1,222,723,083
https://api.github.com/repos/huggingface/datasets/issues/4263/labels{/name}
On the Hugging Face Hub, users refer to imagenet2012 (from #4178 ) as imagenet-1k in their model tags. To correctly link models to imagenet, we should rename this dataset `imagenet-1k`. Later we can add `imagenet-21k` as a new dataset if we want. Once this one is merged we can delete the `imagenet2012` dataset repository on the Hub. EDIT: to complete the rationale on why we should name it `imagenet-1k`: If users specifically added the tag `imagenet-1k` , then it could be for two reasons (not sure which one is predominant), either they - wanted to make it explicit that it’s not 21k -> the distinction is important for the community - or they have been following this convention from other models -> the convention implicitly exists already
2022-05-02T17:50:46Z
4,263
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-05-02T10:26:21Z
https://api.github.com/repos/huggingface/datasets/issues/4263/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4263/timeline
Rename imagenet2012 -> imagenet-1k
https://api.github.com/repos/huggingface/datasets/issues/4263/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-05-02T16:32:57Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4263.diff", "html_url": "https://github.com/huggingface/datasets/pull/4263", "merged_at": "2022-05-02T16:32:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/4263.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4263" }
PR_kwDODunzps43KLnD
[ "> Later we can add imagenet-21k as a new dataset if we want.\r\n\r\nisn't it what models refer to as `imagenet` already?", "> isn't it what models refer to as imagenet already?\r\n\r\nI wasn't sure, but it looks like it indeed. Therefore having a dataset `imagenet` for ImageNet 21k makes sense actually.\r\n\r\nEDIT: actually not all `imagenet` tag refer to ImageNet 21k - we will need to correct some of them", "_The documentation is not available anymore as the PR was closed or merged._", "should we remove the repo mirror on the hub side or will you do it?" ]
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 4, "url": "https://api.github.com/repos/huggingface/datasets/issues/4263/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4263
https://github.com/huggingface/datasets/pull/4263
true
1,222,130,749
https://api.github.com/repos/huggingface/datasets/issues/4262/labels{/name}
The dataset card for the rotten tomatoes / MR movie review dataset had some missing YAML tags. Hopefully, this also improves the visibility of this dataset now that paperswithcode and huggingface link to eachother.
2022-05-03T14:27:33Z
4,262
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-05-01T11:59:08Z
https://api.github.com/repos/huggingface/datasets/issues/4262/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4262/timeline
Add YAML tags to Dataset Card rotten tomatoes
https://api.github.com/repos/huggingface/datasets/issues/4262/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/10004251?v=4", "events_url": "https://api.github.com/users/mo6zes/events{/privacy}", "followers_url": "https://api.github.com/users/mo6zes/followers", "following_url": "https://api.github.com/users/mo6zes/following{/other_user}", "gists_url": "https://api.github.com/users/mo6zes/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mo6zes", "id": 10004251, "login": "mo6zes", "node_id": "MDQ6VXNlcjEwMDA0MjUx", "organizations_url": "https://api.github.com/users/mo6zes/orgs", "received_events_url": "https://api.github.com/users/mo6zes/received_events", "repos_url": "https://api.github.com/users/mo6zes/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mo6zes/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mo6zes/subscriptions", "type": "User", "url": "https://api.github.com/users/mo6zes" }
[]
null
null
CONTRIBUTOR
2022-05-03T14:20:35Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4262.diff", "html_url": "https://github.com/huggingface/datasets/pull/4262", "merged_at": "2022-05-03T14:20:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/4262.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4262" }
PR_kwDODunzps43IOye
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4262/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4262
https://github.com/huggingface/datasets/pull/4262
true
1,221,883,779
https://api.github.com/repos/huggingface/datasets/issues/4261/labels{/name}
## Describe the bug Some samples (argument-conclusion pairs) in the *training* split of the `webis/conclugen` dataset are present in both the *validation* and *test* splits, creating data leakage and distorting model results. Furthermore, all splits contain duplicate samples. ## Steps to reproduce the bug ```python from datasets import load_dataset training = load_dataset("webis/conclugen", "base", split="train") validation = load_dataset("webis/conclugen", "base", split="validation") testing = load_dataset("webis/conclugen", "base", split="test") # collect which sample id's are present in the training split ids_validation = list() ids_testing = list() for train_sample in training: train_argument = train_sample["argument"] train_conclusion = train_sample["conclusion"] train_id = train_sample["id"] # test if current sample is in validation split if train_argument in validation["argument"]: for validation_sample in validation: validation_argument = validation_sample["argument"] validation_conclusion = validation_sample["conclusion"] validation_id = validation_sample["id"] if train_argument == validation_argument and train_conclusion == validation_conclusion: ids_validation.append(validation_id) # test if current sample is in test split if train_argument in testing["argument"]: for testing_sample in testing: testing_argument = testing_sample["argument"] testing_conclusion = testing_sample["conclusion"] testing_id = testing_sample["id"] if train_argument == testing_argument and train_conclusion == testing_conclusion: ids_testing.append(testing_id) ``` ## Expected results Length of both lists `ids_validation` and `ids_testing` should be zero. ## Actual results Length of `ids_validation` = `2556` Length of `ids_testing` = `287` Furthermore, there seems to be duplicate samples in (at least) the *training* split, since: `print(len(set(ids_validation)))` = `950` `print(len(set(ids_testing)))` = `101` All in all, around 7% of the samples of each the *validation* and *test* split seems to be present in the *training* split. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.18.4 - Platform: macOS-12.3.1-arm64-arm-64bit - Python version: 3.9.10 - PyArrow version: 7.0.0
2022-05-03T06:04:26Z
4,261
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
2022-04-30T17:43:37Z
https://api.github.com/repos/huggingface/datasets/issues/4261/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4261/timeline
data leakage in `webis/conclugen` dataset
https://api.github.com/repos/huggingface/datasets/issues/4261/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/54585776?v=4", "events_url": "https://api.github.com/users/xflashxx/events{/privacy}", "followers_url": "https://api.github.com/users/xflashxx/followers", "following_url": "https://api.github.com/users/xflashxx/following{/other_user}", "gists_url": "https://api.github.com/users/xflashxx/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/xflashxx", "id": 54585776, "login": "xflashxx", "node_id": "MDQ6VXNlcjU0NTg1Nzc2", "organizations_url": "https://api.github.com/users/xflashxx/orgs", "received_events_url": "https://api.github.com/users/xflashxx/received_events", "repos_url": "https://api.github.com/users/xflashxx/repos", "site_admin": false, "starred_url": "https://api.github.com/users/xflashxx/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xflashxx/subscriptions", "type": "User", "url": "https://api.github.com/users/xflashxx" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
completed
NONE
2022-05-03T06:04:26Z
null
I_kwDODunzps5I1HeD
[ "Hi @xflashxx, thanks for reporting.\r\n\r\nPlease note that this dataset was generated and shared by Webis Group: https://huggingface.co/webis\r\n\r\nWe are contacting the dataset owners to inform them about the issue you found. We'll keep you updated of their reply.", "i'd suggest just pinging the authors here in the issue if possible?", "Thanks for reporting this @xflashxx. I'll have a look and get back to you on this.", "Hi @xflashxx and @albertvillanova,\r\n\r\nI have updated the files with de-duplicated splits. Apparently the debate portals from which part of the examples were sourced had unique timestamps for some examples (up to 6%; updated counts in the README) without any actual content updated that lead to \"new\" items. The length of `ids_validation` and `ids_testing` is zero.\r\n\r\nRegarding impact on scores:\r\n1. We employed automatic evaluation (on a separate set of 1000 examples) only to justify the exclusion of the smaller models for manual evaluation (due to budget constraints). I am confident the ranking still stands (unsurprisingly, the bigger models doing better than those trained on the smaller splits). We also highlight this in the paper. \r\n\r\n2. The examples used for manual evaluation have no overlap with any splits (also because they do not have any ground truth as we applied the trained models on an unlabeled sample to test its practical usage). I've added these two files to the dataset repository.\r\n\r\nHope this helps!", "Thanks @shahbazsyed for your fast fix.\r\n\r\nAs a side note:\r\n- Your email appearing as Point of Contact in the dataset README has a typo: @uni.leipzig.de instead of @uni-leipzig.de\r\n- Your commits on the Hub are not linked to your profile on the Hub: this is because we use the email address to make this link; the email address used in your commit author and the email address set on your Hub account settings." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4261/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4261
https://github.com/huggingface/datasets/issues/4261
false
1,221,830,292
https://api.github.com/repos/huggingface/datasets/issues/4260/labels{/name}
Add the MR (Movie Review) dataset. The original dataset contains sentences from Rotten Tomatoes labeled as either "positive" or "negative". Homepage: [https://www.cs.cornell.edu/people/pabo/movie-review-data/](https://www.cs.cornell.edu/people/pabo/movie-review-data/) paperswithcode: [https://paperswithcode.com/dataset/mr](https://paperswithcode.com/dataset/mr) - [ ] I was not able to generate dummy data, the original dataset files have ".pos" and ".neg" as file extensions so the auto-generator does not work. Is it fine like this or should dummy data be added?
2022-04-30T14:16:25Z
4,260
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-30T13:19:33Z
https://api.github.com/repos/huggingface/datasets/issues/4260/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4260/timeline
Add mr_polarity movie review sentiment classification
https://api.github.com/repos/huggingface/datasets/issues/4260/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/10004251?v=4", "events_url": "https://api.github.com/users/mo6zes/events{/privacy}", "followers_url": "https://api.github.com/users/mo6zes/followers", "following_url": "https://api.github.com/users/mo6zes/following{/other_user}", "gists_url": "https://api.github.com/users/mo6zes/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mo6zes", "id": 10004251, "login": "mo6zes", "node_id": "MDQ6VXNlcjEwMDA0MjUx", "organizations_url": "https://api.github.com/users/mo6zes/orgs", "received_events_url": "https://api.github.com/users/mo6zes/received_events", "repos_url": "https://api.github.com/users/mo6zes/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mo6zes/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mo6zes/subscriptions", "type": "User", "url": "https://api.github.com/users/mo6zes" }
[]
null
null
CONTRIBUTOR
2022-04-30T14:16:25Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4260.diff", "html_url": "https://github.com/huggingface/datasets/pull/4260", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4260.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4260" }
PR_kwDODunzps43HSfs
[ "whoops just found https://huggingface.co/datasets/rotten_tomatoes" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4260/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4260
https://github.com/huggingface/datasets/pull/4260
true
1,221,768,025
https://api.github.com/repos/huggingface/datasets/issues/4259/labels{/name}
This PR fixes the Bug in the openbookqa dataset as mentioned in this issue #3550. Fix #3550. cc. @lhoestq @mariosasko
2022-05-04T06:31:31Z
4,259
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-30T07:41:39Z
https://api.github.com/repos/huggingface/datasets/issues/4259/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4259/timeline
Fix bug in choices labels in openbookqa dataset
https://api.github.com/repos/huggingface/datasets/issues/4259/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6687858?v=4", "events_url": "https://api.github.com/users/manandey/events{/privacy}", "followers_url": "https://api.github.com/users/manandey/followers", "following_url": "https://api.github.com/users/manandey/following{/other_user}", "gists_url": "https://api.github.com/users/manandey/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/manandey", "id": 6687858, "login": "manandey", "node_id": "MDQ6VXNlcjY2ODc4NTg=", "organizations_url": "https://api.github.com/users/manandey/orgs", "received_events_url": "https://api.github.com/users/manandey/received_events", "repos_url": "https://api.github.com/users/manandey/repos", "site_admin": false, "starred_url": "https://api.github.com/users/manandey/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/manandey/subscriptions", "type": "User", "url": "https://api.github.com/users/manandey" }
[]
null
null
CONTRIBUTOR
2022-05-03T15:14:21Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4259.diff", "html_url": "https://github.com/huggingface/datasets/pull/4259", "merged_at": "2022-05-03T15:14:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/4259.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4259" }
PR_kwDODunzps43HHGc
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4259/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4259
https://github.com/huggingface/datasets/pull/4259
true
1,221,637,727
https://api.github.com/repos/huggingface/datasets/issues/4258/labels{/name}
This pr fixes a couple bugs: 1) the perplexity was calculated with a 0 in the attention mask for the start token, which was causing high perplexity scores that were not correct 2) the documentation was not updated
2022-05-02T16:33:20Z
4,258
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-29T22:42:44Z
https://api.github.com/repos/huggingface/datasets/issues/4258/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4258/timeline
Fix/start token mask issue and update documentation
https://api.github.com/repos/huggingface/datasets/issues/4258/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/20826878?v=4", "events_url": "https://api.github.com/users/TristanThrush/events{/privacy}", "followers_url": "https://api.github.com/users/TristanThrush/followers", "following_url": "https://api.github.com/users/TristanThrush/following{/other_user}", "gists_url": "https://api.github.com/users/TristanThrush/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/TristanThrush", "id": 20826878, "login": "TristanThrush", "node_id": "MDQ6VXNlcjIwODI2ODc4", "organizations_url": "https://api.github.com/users/TristanThrush/orgs", "received_events_url": "https://api.github.com/users/TristanThrush/received_events", "repos_url": "https://api.github.com/users/TristanThrush/repos", "site_admin": false, "starred_url": "https://api.github.com/users/TristanThrush/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TristanThrush/subscriptions", "type": "User", "url": "https://api.github.com/users/TristanThrush" }
[]
null
null
CONTRIBUTOR
2022-05-02T16:26:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4258.diff", "html_url": "https://github.com/huggingface/datasets/pull/4258", "merged_at": "2022-05-02T16:26:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/4258.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4258" }
PR_kwDODunzps43Gstg
[ "_The documentation is not available anymore as the PR was closed or merged._", "> Good catch ! Thanks :)\r\n> \r\n> Next time can you describe your fix in the Pull Request description please ?\r\n\r\nThanks. Also whoops, sorry about not being very descriptive. I updated the pull request description, and will keep this in mind for future PRs." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4258/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4258
https://github.com/huggingface/datasets/pull/4258
true
1,221,393,137
https://api.github.com/repos/huggingface/datasets/issues/4257/labels{/name}
proposing a metric card to better explain how Mahalanobis distance works (last one for now :sweat_smile:
2022-05-02T14:50:18Z
4,257
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-29T18:37:27Z
https://api.github.com/repos/huggingface/datasets/issues/4257/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4257/timeline
Create metric card for Mahalanobis Distance
https://api.github.com/repos/huggingface/datasets/issues/4257/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
null
null
NONE
2022-05-02T14:43:24Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4257.diff", "html_url": "https://github.com/huggingface/datasets/pull/4257", "merged_at": "2022-05-02T14:43:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/4257.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4257" }
PR_kwDODunzps43GATC
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4257/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4257
https://github.com/huggingface/datasets/pull/4257
true
1,221,379,625
https://api.github.com/repos/huggingface/datasets/issues/4256/labels{/name}
Proposing a metric card for Mean Squared Error
2022-05-02T14:55:42Z
4,256
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-29T18:21:22Z
https://api.github.com/repos/huggingface/datasets/issues/4256/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4256/timeline
Create metric card for MSE
https://api.github.com/repos/huggingface/datasets/issues/4256/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
null
null
NONE
2022-05-02T14:48:47Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4256.diff", "html_url": "https://github.com/huggingface/datasets/pull/4256", "merged_at": "2022-05-02T14:48:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/4256.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4256" }
PR_kwDODunzps43F9Zw
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4256/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4256
https://github.com/huggingface/datasets/pull/4256
true
1,221,142,899
https://api.github.com/repos/huggingface/datasets/issues/4255/labels{/name}
I hosted the data files in https://huggingface.co/datasets/pubmed_qa. This is allowed because the data is under the MIT license. cc @stas00
2022-04-29T16:24:55Z
4,255
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-29T15:55:46Z
https://api.github.com/repos/huggingface/datasets/issues/4255/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4255/timeline
No google drive URL for pubmed_qa
https://api.github.com/repos/huggingface/datasets/issues/4255/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-04-29T16:18:56Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4255.diff", "html_url": "https://github.com/huggingface/datasets/pull/4255", "merged_at": "2022-04-29T16:18:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/4255.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4255" }
PR_kwDODunzps43FHgR
[ "_The documentation is not available anymore as the PR was closed or merged._", "CI is failing because some sections are missing in the dataset card, but this is unrelated to this PR - Merging !" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4255/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4255
https://github.com/huggingface/datasets/pull/4255
true
1,220,204,395
https://api.github.com/repos/huggingface/datasets/issues/4254/labels{/name}
This PR replaces data URL in SAMSum dataset: - original host (arxiv.org) does not allow HTTP Range requests - we have hosted the data on the Hub (license: CC BY-NC-ND 4.0) Moreover, it implements support for streaming. Fix #4146. Related to: #4236. CC: @severo
2022-05-06T08:38:16Z
4,254
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-29T08:21:43Z
https://api.github.com/repos/huggingface/datasets/issues/4254/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4254/timeline
Replace data URL in SAMSum dataset and support streaming
https://api.github.com/repos/huggingface/datasets/issues/4254/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-04-29T16:26:09Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4254.diff", "html_url": "https://github.com/huggingface/datasets/pull/4254", "merged_at": "2022-04-29T16:26:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/4254.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4254" }
PR_kwDODunzps43Bwnj
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4254/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4254
https://github.com/huggingface/datasets/pull/4254
true
1,219,286,408
https://api.github.com/repos/huggingface/datasets/issues/4253/labels{/name}
Proposing a metric card for mIoU :rocket: sorry for spamming you with review requests, @albertvillanova ! :hugs:
2022-04-29T17:44:47Z
4,253
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-28T20:58:27Z
https://api.github.com/repos/huggingface/datasets/issues/4253/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4253/timeline
Create metric cards for mean IOU
https://api.github.com/repos/huggingface/datasets/issues/4253/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
null
null
NONE
2022-04-29T17:38:06Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4253.diff", "html_url": "https://github.com/huggingface/datasets/pull/4253", "merged_at": "2022-04-29T17:38:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/4253.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4253" }
PR_kwDODunzps42-c8Q
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4253/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4253
https://github.com/huggingface/datasets/pull/4253
true
1,219,151,100
https://api.github.com/repos/huggingface/datasets/issues/4252/labels{/name}
Initial proposal for MAE metric card
2022-04-29T16:59:11Z
4,252
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-28T19:04:33Z
https://api.github.com/repos/huggingface/datasets/issues/4252/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4252/timeline
Creating metric card for MAE
https://api.github.com/repos/huggingface/datasets/issues/4252/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
null
null
NONE
2022-04-29T16:52:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4252.diff", "html_url": "https://github.com/huggingface/datasets/pull/4252", "merged_at": "2022-04-29T16:52:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/4252.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4252" }
PR_kwDODunzps429--I
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4252/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4252
https://github.com/huggingface/datasets/pull/4252
true
1,219,116,354
https://api.github.com/repos/huggingface/datasets/issues/4251/labels{/name}
Proposing a metric card for the XTREME-S dataset :hugs:
2022-04-29T16:46:11Z
4,251
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-28T18:32:19Z
https://api.github.com/repos/huggingface/datasets/issues/4251/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4251/timeline
Metric card for the XTREME-S dataset
https://api.github.com/repos/huggingface/datasets/issues/4251/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
null
null
NONE
2022-04-29T16:38:46Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4251.diff", "html_url": "https://github.com/huggingface/datasets/pull/4251", "merged_at": "2022-04-29T16:38:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/4251.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4251" }
PR_kwDODunzps4293dB
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4251/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4251
https://github.com/huggingface/datasets/pull/4251
true
1,219,093,830
https://api.github.com/repos/huggingface/datasets/issues/4250/labels{/name}
Fixes #4152 This PR updates the PyArrow version to 6 in setup.py, CI job files .circleci/config.yaml and .github/workflows/benchmarks.yaml files. This will fix ArrayND error which exists in pyarrow 5.
2022-05-04T09:36:52Z
4,250
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-28T18:10:50Z
https://api.github.com/repos/huggingface/datasets/issues/4250/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4250/timeline
Bump PyArrow Version to 6
https://api.github.com/repos/huggingface/datasets/issues/4250/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/17746528?v=4", "events_url": "https://api.github.com/users/dnaveenr/events{/privacy}", "followers_url": "https://api.github.com/users/dnaveenr/followers", "following_url": "https://api.github.com/users/dnaveenr/following{/other_user}", "gists_url": "https://api.github.com/users/dnaveenr/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dnaveenr", "id": 17746528, "login": "dnaveenr", "node_id": "MDQ6VXNlcjE3NzQ2NTI4", "organizations_url": "https://api.github.com/users/dnaveenr/orgs", "received_events_url": "https://api.github.com/users/dnaveenr/received_events", "repos_url": "https://api.github.com/users/dnaveenr/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dnaveenr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dnaveenr/subscriptions", "type": "User", "url": "https://api.github.com/users/dnaveenr" }
[]
null
null
CONTRIBUTOR
2022-05-04T09:29:46Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4250.diff", "html_url": "https://github.com/huggingface/datasets/pull/4250", "merged_at": "2022-05-04T09:29:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/4250.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4250" }
PR_kwDODunzps429yjN
[ "_The documentation is not available anymore as the PR was closed or merged._", "Updated meta.yaml as well. Thanks.", "I'm OK with bumping PyArrow to version 6 to match the version in Colab, but maybe a better solution would be to stop using extension types in our codebase to avoid similar issues.", "> but maybe a better solution would be to stop using extension types in our codebase to avoid similar issues.\r\n\r\nI agree, not much attention has been payed to extension arrays in the latest developments of Arrow anyway.\r\n\r\nLet's not use them more that what we do right now, and try to remove them at one point" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4250/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4250
https://github.com/huggingface/datasets/pull/4250
true
1,218,524,424
https://api.github.com/repos/huggingface/datasets/issues/4249/labels{/name}
Support streaming XGLUE dataset. Fix #4247. CC: @severo
2022-05-06T08:38:21Z
4,249
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-28T10:27:23Z
https://api.github.com/repos/huggingface/datasets/issues/4249/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4249/timeline
Support streaming XGLUE dataset
https://api.github.com/repos/huggingface/datasets/issues/4249/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-04-28T16:08:03Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4249.diff", "html_url": "https://github.com/huggingface/datasets/pull/4249", "merged_at": "2022-04-28T16:08:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/4249.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4249" }
PR_kwDODunzps42742y
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4249/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4249
https://github.com/huggingface/datasets/pull/4249
true
1,218,460,444
https://api.github.com/repos/huggingface/datasets/issues/4248/labels{/name}
## Describe the bug I load `conll2003` dataset to use refined data like [this](https://huggingface.co/datasets/conll2003/viewer/conll2003/train) preview, but it is original data that contains `'-DOCSTART- -X- -X- O'` text. Is this a bug or should I use another dataset_name like `lhoestq/conll2003` ? ## Steps to reproduce the bug ```python import datasets from datasets import load_dataset dataset = load_dataset("conll2003") ``` ## Expected results { "chunk_tags": [11, 12, 12, 21, 13, 11, 11, 21, 13, 11, 12, 13, 11, 21, 22, 11, 12, 17, 11, 21, 17, 11, 12, 12, 21, 22, 22, 13, 11, 0], "id": "0", "ner_tags": [0, 3, 4, 0, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], "pos_tags": [12, 22, 22, 38, 15, 22, 28, 38, 15, 16, 21, 35, 24, 35, 37, 16, 21, 15, 24, 41, 15, 16, 21, 21, 20, 37, 40, 35, 21, 7], "tokens": ["The", "European", "Commission", "said", "on", "Thursday", "it", "disagreed", "with", "German", "advice", "to", "consumers", "to", "shun", "British", "lamb", "until", "scientists", "determine", "whether", "mad", "cow", "disease", "can", "be", "transmitted", "to", "sheep", "."] } ## Actual results ```python print(dataset) DatasetDict({ train: Dataset({ features: ['text'], num_rows: 219554 }) test: Dataset({ features: ['text'], num_rows: 50350 }) validation: Dataset({ features: ['text'], num_rows: 55044 }) }) ``` ```python for i in range(20): print(dataset['train'][i]) {'text': '-DOCSTART- -X- -X- O'} {'text': ''} {'text': 'EU NNP B-NP B-ORG'} {'text': 'rejects VBZ B-VP O'} {'text': 'German JJ B-NP B-MISC'} {'text': 'call NN I-NP O'} {'text': 'to TO B-VP O'} {'text': 'boycott VB I-VP O'} {'text': 'British JJ B-NP B-MISC'} {'text': 'lamb NN I-NP O'} {'text': '. . O O'} {'text': ''} {'text': 'Peter NNP B-NP B-PER'} {'text': 'Blackburn NNP I-NP I-PER'} {'text': ''} {'text': 'BRUSSELS NNP B-NP B-LOC'} {'text': '1996-08-22 CD I-NP O'} {'text': ''} {'text': 'The DT B-NP O'} {'text': 'European NNP I-NP B-ORG'} ```
2022-07-18T07:15:48Z
4,248
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-28T09:33:31Z
https://api.github.com/repos/huggingface/datasets/issues/4248/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4248/timeline
conll2003 dataset loads original data.
https://api.github.com/repos/huggingface/datasets/issues/4248/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/26458611?v=4", "events_url": "https://api.github.com/users/sue991/events{/privacy}", "followers_url": "https://api.github.com/users/sue991/followers", "following_url": "https://api.github.com/users/sue991/following{/other_user}", "gists_url": "https://api.github.com/users/sue991/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sue991", "id": 26458611, "login": "sue991", "node_id": "MDQ6VXNlcjI2NDU4NjEx", "organizations_url": "https://api.github.com/users/sue991/orgs", "received_events_url": "https://api.github.com/users/sue991/received_events", "repos_url": "https://api.github.com/users/sue991/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sue991/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sue991/subscriptions", "type": "User", "url": "https://api.github.com/users/sue991" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
completed
NONE
2022-07-18T07:15:48Z
null
I_kwDODunzps5IoDsc
[ "Thanks for reporting @sue99.\r\n\r\nUnfortunately. I'm not able to reproduce your problem:\r\n```python\r\nIn [1]: import datasets\r\n ...: from datasets import load_dataset\r\n ...: dataset = load_dataset(\"conll2003\")\r\n\r\nIn [2]: dataset\r\nOut[2]: \r\nDatasetDict({\r\n train: Dataset({\r\n features: ['id', 'tokens', 'pos_tags', 'chunk_tags', 'ner_tags'],\r\n num_rows: 14042\r\n })\r\n validation: Dataset({\r\n features: ['id', 'tokens', 'pos_tags', 'chunk_tags', 'ner_tags'],\r\n num_rows: 3251\r\n })\r\n test: Dataset({\r\n features: ['id', 'tokens', 'pos_tags', 'chunk_tags', 'ner_tags'],\r\n num_rows: 3454\r\n })\r\n})\r\n\r\nIn [3]: dataset[\"train\"][0]\r\nOut[3]: \r\n{'id': '0',\r\n 'tokens': ['EU',\r\n 'rejects',\r\n 'German',\r\n 'call',\r\n 'to',\r\n 'boycott',\r\n 'British',\r\n 'lamb',\r\n '.'],\r\n 'pos_tags': [22, 42, 16, 21, 35, 37, 16, 21, 7],\r\n 'chunk_tags': [11, 21, 11, 12, 21, 22, 11, 12, 0],\r\n 'ner_tags': [3, 0, 7, 0, 0, 0, 7, 0, 0]}\r\n```\r\n\r\nJust guessing: might be the case that you are calling `load_dataset` from a working directory that contains a local folder named `conll2003` (containing the raw data files)? If that is the case, `datasets` library gives precedence to the local folder over the dataset on the Hub. " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4248/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4248
https://github.com/huggingface/datasets/issues/4248
false
1,218,320,882
https://api.github.com/repos/huggingface/datasets/issues/4247/labels{/name}
It seems that something wrong with the data previvew of XGLUE
2022-04-29T08:23:28Z
4,247
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-04-28T07:30:50Z
https://api.github.com/repos/huggingface/datasets/issues/4247/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4247/timeline
The data preview of XGLUE
https://api.github.com/repos/huggingface/datasets/issues/4247/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/49108847?v=4", "events_url": "https://api.github.com/users/czq1999/events{/privacy}", "followers_url": "https://api.github.com/users/czq1999/followers", "following_url": "https://api.github.com/users/czq1999/following{/other_user}", "gists_url": "https://api.github.com/users/czq1999/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/czq1999", "id": 49108847, "login": "czq1999", "node_id": "MDQ6VXNlcjQ5MTA4ODQ3", "organizations_url": "https://api.github.com/users/czq1999/orgs", "received_events_url": "https://api.github.com/users/czq1999/received_events", "repos_url": "https://api.github.com/users/czq1999/repos", "site_admin": false, "starred_url": "https://api.github.com/users/czq1999/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/czq1999/subscriptions", "type": "User", "url": "https://api.github.com/users/czq1999" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
completed
NONE
2022-04-28T16:08:03Z
null
I_kwDODunzps5Inhny
[ "![image](https://user-images.githubusercontent.com/49108847/165700611-915b4343-766f-4b81-bdaa-b31950250f06.png)\r\n", "Thanks for reporting @czq1999.\r\n\r\nNote that the dataset viewer uses the dataset in streaming mode and that not all datasets support streaming yet.\r\n\r\nThat is the case for XGLUE dataset (as the error message points out): this must be refactored to support streaming. ", "Fixed, thanks @albertvillanova !\r\n\r\nhttps://huggingface.co/datasets/xglue\r\n\r\n<img width=\"824\" alt=\"Capture d’écran 2022-04-29 à 10 23 14\" src=\"https://user-images.githubusercontent.com/1676121/165909391-9f98d98a-665a-4e57-822d-8baa2dc9b7c9.png\">\r\n" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4247/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4247
https://github.com/huggingface/datasets/issues/4247
false
1,218,320,293
https://api.github.com/repos/huggingface/datasets/issues/4246/labels{/name}
This PR implements support to load a dataset (w/o script) containing TSV files by passing only the dataset name (no need to pass `sep='\t'`): ```python ds = load_dataset("dataset/name") ``` The refactoring allows for future builder kwargs customizations based on file extension. Related to #4238.
2022-05-06T08:38:28Z
4,246
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-28T07:30:15Z
https://api.github.com/repos/huggingface/datasets/issues/4246/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4246/timeline
Support to load dataset with TSV files by passing only dataset name
https://api.github.com/repos/huggingface/datasets/issues/4246/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-05-06T08:14:07Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4246.diff", "html_url": "https://github.com/huggingface/datasets/pull/4246", "merged_at": "2022-05-06T08:14:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/4246.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4246" }
PR_kwDODunzps427NiD
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4246/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4246
https://github.com/huggingface/datasets/pull/4246
true
1,217,959,400
https://api.github.com/repos/huggingface/datasets/issues/4245/labels{/name}
This PR adds code examples for `DatasetDict` in the API reference :)
2022-04-29T18:19:34Z
4,245
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
2022-04-27T22:52:22Z
https://api.github.com/repos/huggingface/datasets/issues/4245/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4245/timeline
Add code examples for DatasetDict
https://api.github.com/repos/huggingface/datasets/issues/4245/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2022-04-29T18:13:03Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4245.diff", "html_url": "https://github.com/huggingface/datasets/pull/4245", "merged_at": "2022-04-29T18:13:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/4245.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4245" }
PR_kwDODunzps426AUR
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4245/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4245
https://github.com/huggingface/datasets/pull/4245
true
1,217,732,221
https://api.github.com/repos/huggingface/datasets/issues/4244/labels{/name}
changed multi input text classification as task id instead of category
2022-05-04T10:43:53Z
4,244
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T18:28:14Z
https://api.github.com/repos/huggingface/datasets/issues/4244/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4244/timeline
task id update
https://api.github.com/repos/huggingface/datasets/issues/4244/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3278583?v=4", "events_url": "https://api.github.com/users/nazneenrajani/events{/privacy}", "followers_url": "https://api.github.com/users/nazneenrajani/followers", "following_url": "https://api.github.com/users/nazneenrajani/following{/other_user}", "gists_url": "https://api.github.com/users/nazneenrajani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nazneenrajani", "id": 3278583, "login": "nazneenrajani", "node_id": "MDQ6VXNlcjMyNzg1ODM=", "organizations_url": "https://api.github.com/users/nazneenrajani/orgs", "received_events_url": "https://api.github.com/users/nazneenrajani/received_events", "repos_url": "https://api.github.com/users/nazneenrajani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nazneenrajani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nazneenrajani/subscriptions", "type": "User", "url": "https://api.github.com/users/nazneenrajani" }
[]
null
null
NONE
2022-05-04T10:36:37Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4244.diff", "html_url": "https://github.com/huggingface/datasets/pull/4244", "merged_at": "2022-05-04T10:36:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/4244.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4244" }
PR_kwDODunzps425Po6
[ "Reverted the multi-input-text-classification tag from task_categories and added it as task_ids @lhoestq ", "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4244/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4244
https://github.com/huggingface/datasets/pull/4244
true
1,217,689,909
https://api.github.com/repos/huggingface/datasets/issues/4243/labels{/name}
null
2022-10-03T09:36:35Z
4,243
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-04-27T17:45:43Z
https://api.github.com/repos/huggingface/datasets/issues/4243/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4243/timeline
WIP: Initial shades loading script and readme
https://api.github.com/repos/huggingface/datasets/issues/4243/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/69018523?v=4", "events_url": "https://api.github.com/users/shayne-longpre/events{/privacy}", "followers_url": "https://api.github.com/users/shayne-longpre/followers", "following_url": "https://api.github.com/users/shayne-longpre/following{/other_user}", "gists_url": "https://api.github.com/users/shayne-longpre/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shayne-longpre", "id": 69018523, "login": "shayne-longpre", "node_id": "MDQ6VXNlcjY5MDE4NTIz", "organizations_url": "https://api.github.com/users/shayne-longpre/orgs", "received_events_url": "https://api.github.com/users/shayne-longpre/received_events", "repos_url": "https://api.github.com/users/shayne-longpre/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shayne-longpre/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shayne-longpre/subscriptions", "type": "User", "url": "https://api.github.com/users/shayne-longpre" }
[]
null
null
NONE
2022-10-03T09:36:35Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4243.diff", "html_url": "https://github.com/huggingface/datasets/pull/4243", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4243.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4243" }
PR_kwDODunzps425Gkn
[ "Thanks for your contribution, @shayne-longpre.\r\n\r\nAre you still interested in adding this dataset? As we are transferring the dataset scripts from this GitHub repo, we would recommend you to add this to the Hugging Face Hub: https://huggingface.co/datasets" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4243/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4243
https://github.com/huggingface/datasets/pull/4243
true
1,217,665,960
https://api.github.com/repos/huggingface/datasets/issues/4242/labels{/name}
We don't need to use extraHeaders anymore for rate limits anymore. Anyway extraHeaders was not working with git LFS because it was passing the wrong auth to S3.
2022-04-27T17:37:04Z
4,242
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T17:22:31Z
https://api.github.com/repos/huggingface/datasets/issues/4242/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4242/timeline
Update auth when mirroring datasets on the hub
https://api.github.com/repos/huggingface/datasets/issues/4242/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-04-27T17:30:42Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4242.diff", "html_url": "https://github.com/huggingface/datasets/pull/4242", "merged_at": "2022-04-27T17:30:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/4242.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4242" }
PR_kwDODunzps425BYf
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4242/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4242
https://github.com/huggingface/datasets/pull/4242
true
1,217,423,686
https://api.github.com/repos/huggingface/datasets/issues/4241/labels{/name}
## Describe the bug I am trying to download the GLUE dataset from the NLP module but get an error (see below). ## Steps to reproduce the bug ```python import nlp nlp.__version__ # '0.2.0' nlp.load_dataset('glue', name="rte", download_mode="force_redownload") ``` ## Expected results I expect the dataset to download without an error. ## Actual results ``` INFO:nlp.load:Checking /home/richier/.cache/huggingface/datasets/5fe6ab0df8a32a3371b2e6a969d31d855a19563724fb0d0f163748c270c0ac60.2ea96febf19981fae5f13f0a43d4e2aa58bc619bc23acf06de66675f425a5538.py for additional imports. INFO:nlp.load:Found main folder for dataset https://s3.amazonaws.com/datasets.huggingface.co/nlp/datasets/glue/glue.py at /home/richier/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/datasets/glue INFO:nlp.load:Found specific version folder for dataset https://s3.amazonaws.com/datasets.huggingface.co/nlp/datasets/glue/glue.py at /home/richier/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/datasets/glue/637080968c182118f006d3ea39dd9937940e81cfffc8d79836eaae8bba307fc4 INFO:nlp.load:Found script file from https://s3.amazonaws.com/datasets.huggingface.co/nlp/datasets/glue/glue.py to /home/richier/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/datasets/glue/637080968c182118f006d3ea39dd9937940e81cfffc8d79836eaae8bba307fc4/glue.py INFO:nlp.load:Found dataset infos file from https://s3.amazonaws.com/datasets.huggingface.co/nlp/datasets/glue/dataset_infos.json to /home/richier/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/datasets/glue/637080968c182118f006d3ea39dd9937940e81cfffc8d79836eaae8bba307fc4/dataset_infos.json INFO:nlp.load:Found metadata file for dataset https://s3.amazonaws.com/datasets.huggingface.co/nlp/datasets/glue/glue.py at /home/richier/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/datasets/glue/637080968c182118f006d3ea39dd9937940e81cfffc8d79836eaae8bba307fc4/glue.json INFO:nlp.info:Loading Dataset Infos from /home/richier/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/datasets/glue/637080968c182118f006d3ea39dd9937940e81cfffc8d79836eaae8bba307fc4 INFO:nlp.builder:Generating dataset glue (/home/richier/.cache/huggingface/datasets/glue/rte/1.0.0) INFO:nlp.builder:Dataset not on Hf google storage. Downloading and preparing it from source INFO:nlp.utils.file_utils:Couldn't get ETag version for url https://firebasestorage.googleapis.com/v0/b/mtl-sentence-representations.appspot.com/o/data%2FRTE.zip?alt=media&token=5efa7e85-a0bb-4f19-8ea2-9e1840f077fb INFO:nlp.utils.file_utils:https://firebasestorage.googleapis.com/v0/b/mtl-sentence-representations.appspot.com/o/data%2FRTE.zip?alt=media&token=5efa7e85-a0bb-4f19-8ea2-9e1840f077fb not found in cache or force_download set to True, downloading to /home/richier/.cache/huggingface/datasets/downloads/tmpldt3n805 Downloading and preparing dataset glue/rte (download: 680.81 KiB, generated: 1.83 MiB, total: 2.49 MiB) to /home/richier/.cache/huggingface/datasets/glue/rte/1.0.0... Downloading: 100%|██████████| 73.0/73.0 [00:00<00:00, 73.9kB/s] INFO:nlp.utils.file_utils:storing https://firebasestorage.googleapis.com/v0/b/mtl-sentence-representations.appspot.com/o/data%2FRTE.zip?alt=media&token=5efa7e85-a0bb-4f19-8ea2-9e1840f077fb in cache at /home/richier/.cache/huggingface/datasets/downloads/e8b62ee44e6f8b6aea761935928579ffe1aa55d161808c482e0725abbdcf9c64 INFO:nlp.utils.file_utils:creating metadata file for /home/richier/.cache/huggingface/datasets/downloads/e8b62ee44e6f8b6aea761935928579ffe1aa55d161808c482e0725abbdcf9c64 --------------------------------------------------------------------------- NonMatchingChecksumError Traceback (most recent call last) <ipython-input-7-669a8343dcc1> in <module> ----> 1 nlp.load_dataset('glue', name="rte", download_mode="force_redownload") ~/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs) 518 download_mode=download_mode, 519 ignore_verifications=ignore_verifications, --> 520 save_infos=save_infos, 521 ) 522 ~/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, save_infos, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs) 418 verify_infos = not save_infos and not ignore_verifications 419 self._download_and_prepare( --> 420 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 421 ) 422 # Sync info ~/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 458 # Checksums verification 459 if verify_infos: --> 460 verify_checksums(self.info.download_checksums, dl_manager.get_recorded_sizes_checksums()) 461 for split_generator in split_generators: 462 if str(split_generator.split_info.name).lower() == "all": ~/anaconda3/envs/py36_bert_ee_torch1_11/lib/python3.6/site-packages/nlp/utils/info_utils.py in verify_checksums(expected_checksums, recorded_checksums) 34 bad_urls = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] 35 if len(bad_urls) > 0: ---> 36 raise NonMatchingChecksumError(str(bad_urls)) 37 logger.info("All the checksums matched successfully.") 38 NonMatchingChecksumError: ['https://firebasestorage.googleapis.com/v0/b/mtl-sentence-representations.appspot.com/o/data%2FRTE.zip?alt=media&token=5efa7e85-a0bb-4f19-8ea2-9e1840f077fb'] ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.0.0 - Platform: Linux-4.18.0-348.20.1.el8_5.x86_64-x86_64-with-redhat-8.5-Ootpa - Python version: 3.6.13 - PyArrow version: 6.0.1 - Pandas version: 1.1.5
2022-04-28T07:45:27Z
4,241
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-27T14:14:21Z
https://api.github.com/repos/huggingface/datasets/issues/4241/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4241/timeline
NonMatchingChecksumError when attempting to download GLUE
https://api.github.com/repos/huggingface/datasets/issues/4241/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/9650729?v=4", "events_url": "https://api.github.com/users/drussellmrichie/events{/privacy}", "followers_url": "https://api.github.com/users/drussellmrichie/followers", "following_url": "https://api.github.com/users/drussellmrichie/following{/other_user}", "gists_url": "https://api.github.com/users/drussellmrichie/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/drussellmrichie", "id": 9650729, "login": "drussellmrichie", "node_id": "MDQ6VXNlcjk2NTA3Mjk=", "organizations_url": "https://api.github.com/users/drussellmrichie/orgs", "received_events_url": "https://api.github.com/users/drussellmrichie/received_events", "repos_url": "https://api.github.com/users/drussellmrichie/repos", "site_admin": false, "starred_url": "https://api.github.com/users/drussellmrichie/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/drussellmrichie/subscriptions", "type": "User", "url": "https://api.github.com/users/drussellmrichie" }
[]
null
completed
NONE
2022-04-28T07:45:27Z
null
I_kwDODunzps5IkGlG
[ "Hi :)\r\n\r\nI think your issue may be related to the older `nlp` library. I was able to download `glue` with the latest version of `datasets`. Can you try updating with:\r\n\r\n```py\r\npip install -U datasets\r\n```\r\n\r\nThen you can download:\r\n\r\n```py\r\nfrom datasets import load_dataset\r\nds = load_dataset(\"glue\", \"rte\")\r\n```", "This appears to work. Thank you!\n\nOn Wed, Apr 27, 2022, 1:18 PM Steven Liu ***@***.***> wrote:\n\n> Hi :)\n>\n> I think your issue may be related to the older nlp library. I was able to\n> download glue with the latest version of datasets. Can you try updating\n> with:\n>\n> pip install -U datasets\n>\n> Then you can download:\n>\n> from datasets import load_datasetds = load_dataset(\"glue\", \"rte\")\n>\n> —\n> Reply to this email directly, view it on GitHub\n> <https://github.com/huggingface/datasets/issues/4241#issuecomment-1111267650>,\n> or unsubscribe\n> <https://github.com/notifications/unsubscribe-auth/ACJUEKLUP2EL7ES3RRWJRPTVHFZHBANCNFSM5UPJBYXA>\n> .\n> You are receiving this because you authored the thread.Message ID:\n> ***@***.***>\n>\n" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4241/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4241
https://github.com/huggingface/datasets/issues/4241
false
1,217,287,594
https://api.github.com/repos/huggingface/datasets/issues/4240/labels{/name}
Modified the `_generate_examples` function to consider all the turns for a chunk id as a single example Modified the features accordingly ``` "turns": [ { "names": datasets.features.Sequence(datasets.Value("string")), "utterances": datasets.features.Sequence(datasets.Value("string")), "number": datasets.Value("int32"), } ], } ``` I wasn't able to run `datasets-cli dummy_data datasets` command. Is there a workaround for this?
2022-04-29T12:41:41Z
4,240
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T12:31:36Z
https://api.github.com/repos/huggingface/datasets/issues/4240/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4240/timeline
Fix yield for crd3
https://api.github.com/repos/huggingface/datasets/issues/4240/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/21066979?v=4", "events_url": "https://api.github.com/users/shanyas10/events{/privacy}", "followers_url": "https://api.github.com/users/shanyas10/followers", "following_url": "https://api.github.com/users/shanyas10/following{/other_user}", "gists_url": "https://api.github.com/users/shanyas10/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shanyas10", "id": 21066979, "login": "shanyas10", "node_id": "MDQ6VXNlcjIxMDY2OTc5", "organizations_url": "https://api.github.com/users/shanyas10/orgs", "received_events_url": "https://api.github.com/users/shanyas10/received_events", "repos_url": "https://api.github.com/users/shanyas10/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shanyas10/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shanyas10/subscriptions", "type": "User", "url": "https://api.github.com/users/shanyas10" }
[]
null
null
CONTRIBUTOR
2022-04-29T12:41:41Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4240.diff", "html_url": "https://github.com/huggingface/datasets/pull/4240", "merged_at": "2022-04-29T12:41:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/4240.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4240" }
PR_kwDODunzps423xRl
[ "I don't think you need to generate new dummy data, since they're in the same format as the original data.\r\n\r\nThe CI is failing because of this error:\r\n```python\r\n> turn[\"names\"] = turn[\"NAMES\"]\r\nE TypeError: tuple indices must be integers or slices, not str\r\n```\r\n\r\nDo you know what could cause this ? If I understand correctly, `turn` is supposed to be a list of dictionaries right ?", "> ``` \r\n> \r\n> Do you know what could cause this ? If I understand correctly, turn is supposed to be a list of dictionaries right ?\r\n> ```\r\n\r\nThis is strange. Let me look into this. As per https://github.com/RevanthRameshkumar/CRD3/blob/master/data/aligned%20data/c%3D2/C1E001_2_0.json TURNS is a list of dictionaries. So when we iterate over `row[\"TURNS]\"` each `turn` is essentially a dictionary. Not sure why it's being considered a tuple here." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4240/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4240
https://github.com/huggingface/datasets/pull/4240
true
1,217,269,689
https://api.github.com/repos/huggingface/datasets/issues/4239/labels{/name}
The list of use cases did not render on GitHub with the prepended spacing. Additionally, some typo's we're fixed.
2022-05-02T13:28:57Z
4,239
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T12:15:50Z
https://api.github.com/repos/huggingface/datasets/issues/4239/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4239/timeline
Small fixes in ROC AUC docs
https://api.github.com/repos/huggingface/datasets/issues/4239/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/9478856?v=4", "events_url": "https://api.github.com/users/wschella/events{/privacy}", "followers_url": "https://api.github.com/users/wschella/followers", "following_url": "https://api.github.com/users/wschella/following{/other_user}", "gists_url": "https://api.github.com/users/wschella/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wschella", "id": 9478856, "login": "wschella", "node_id": "MDQ6VXNlcjk0Nzg4NTY=", "organizations_url": "https://api.github.com/users/wschella/orgs", "received_events_url": "https://api.github.com/users/wschella/received_events", "repos_url": "https://api.github.com/users/wschella/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wschella/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wschella/subscriptions", "type": "User", "url": "https://api.github.com/users/wschella" }
[]
null
null
CONTRIBUTOR
2022-05-02T13:22:03Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4239.diff", "html_url": "https://github.com/huggingface/datasets/pull/4239", "merged_at": "2022-05-02T13:22:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/4239.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4239" }
PR_kwDODunzps423tZr
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4239/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4239
https://github.com/huggingface/datasets/pull/4239
true
1,217,168,123
https://api.github.com/repos/huggingface/datasets/issues/4238/labels{/name}
## Describe the bug I cannot clean cache of my datasets files, despite I have updated the `csv` files on the repository [here](https://huggingface.co/datasets/loretoparisi/tatoeba-sentences). The original file had a line with bad characters, causing the following error ``` [/usr/local/lib/python3.7/dist-packages/datasets/features/features.py](https://localhost:8080/#) in str2int(self, values) 852 if value not in self._str2int: 853 value = str(value).strip() --> 854 output.append(self._str2int[str(value)]) 855 else: 856 # No names provided, try to integerize KeyError: '\\N' ``` The file now is cleanup up, but I still get the error. This happens even if I inspect the local cached contents, and cleanup the files locally: ```python from datasets import load_dataset_builder dataset_builder = load_dataset_builder("loretoparisi/tatoeba-sentences") print(dataset_builder.cache_dir) print(dataset_builder.info.features) print(dataset_builder.info.splits) ``` ``` Using custom data configuration loretoparisi--tatoeba-sentences-e59b8ad92f1bb8dd /root/.cache/huggingface/datasets/csv/loretoparisi--tatoeba-sentences-e59b8ad92f1bb8dd/0.0.0/433e0ccc46f9880962cc2b12065189766fbb2bee57a221866138fb9203c83519 None None ``` and removing files located at `/root/.cache/huggingface/datasets/csv/loretoparisi--tatoeba-sentences-*`. Is there any remote file caching policy in place? If so, is it possibile to programmatically disable it? Currently it seems that the file `test.csv` on the repo [here](https://huggingface.co/datasets/loretoparisi/tatoeba-sentences/blob/main/test.csv) is cached remotely. In fact I download locally the file from raw link, the file is up-to-date; but If I use it within `datasets` as shown above, it gives to me always the first revision of the file, not the last. Thank you. ## Steps to reproduce the bug ```python from datasets import load_dataset,Features,Value,ClassLabel class_names = ["cmn","deu","rus","fra","eng","jpn","spa","ita","kor","vie","nld","epo","por","tur","heb","hun","ell","ind","ara","arz","fin","bul","yue","swe","ukr","bel","que","ces","swh","nno","wuu","nob","zsm","est","kat","pol","lat","urd","sqi","isl","fry","afr","ron","fao","san","bre","tat","yid","uig","uzb","srp","qya","dan","pes","slk","eus","cycl","acm","tgl","lvs","kaz","hye","hin","lit","ben","cat","bos","hrv","tha","orv","cha","mon","lzh","scn","gle","mkd","slv","frm","glg","vol","ain","jbo","tok","ina","nds","mal","tlh","roh","ltz","oss","ido","gla","mlt","sco","ast","jav","oci","ile","ota","xal","tel","sjn","nov","khm","tpi","ang","aze","tgk","tuk","chv","hsb","dsb","bod","sme","cym","mri","ksh","kmr","ewe","kab","ber","tpw","udm","lld","pms","lad","grn","mlg","xho","pnb","grc","hat","lao","npi","cor","nah","avk","mar","guj","pan","kir","myv","prg","sux","crs","ckt","bak","zlm","hil","cbk","chr","nav","lkt","enm","arq","lin","abk","pcd","rom","gsw","tam","zul","awa","wln","amh","bar","hbo","mhr","bho","mrj","ckb","osx","pfl","mgm","sna","mah","hau","kan","nog","sin","glv","dng","kal","liv","vro","apc","jdt","fur","che","haw","yor","crh","pdc","ppl","kin","shs","mnw","tet","sah","kum","ngt","nya","pus","hif","mya","moh","wol","tir","ton","lzz","oar","lug","brx","non","mww","hak","nlv","ngu","bua","aym","vec","ibo","tkl","bam","kha","ceb","lou","fuc","smo","gag","lfn","arg","umb","tyv","kjh","oji","cyo","urh","kzj","pam","srd","lmo","swg","mdf","gil","snd","tso","sot","zza","tsn","pau","som","egl","ady","asm","ori","dtp","cho","max","kam","niu","sag","ilo","kaa","fuv","nch","hoc","iba","gbm","sun","war","mvv","pap","ary","kxi","csb","pag","cos","rif","kek","krc","aii","ban","ssw","tvl","mfe","tah","bvy","bcl","hnj","nau","nst","afb","quc","min","tmw","mad","bjn","mai","cjy","got","hsn","gan","tzl","dws","ldn","afh","sgs","krl","vep","rue","tly","mic","ext","izh","sma","jam","cmo","mwl","kpv","koi","bis","ike","run","evn","ryu","mnc","aoz","otk","kas","aln","akl","yua","shy","fkv","gos","fij","thv","zgh","gcf","cay","xmf","tig","div","lij","rap","hrx","cpi","tts","gaa","tmr","iii","ltg","bzt","syc","emx","gom","chg","osp","stq","frr","fro","nys","toi","new","phn","jpa","rel","drt","chn","pli","laa","bal","hdn","hax","mik","ajp","xqa","pal","crk","mni","lut","ayl","ood","sdh","ofs","nus","kiu","diq","qxq","alt","bfz","klj","mus","srn","guc","lim","zea","shi","mnr","bom","sat","szl"] features = Features({ 'label': ClassLabel(names=class_names), 'text': Value('string')}) num_labels = features['label'].num_classes data_files = { "train": "train.csv", "test": "test.csv" } sentences = load_dataset( "loretoparisi/tatoeba-sentences", data_files=data_files, delimiter='\t', column_names=['label', 'text'], ) # You can make this part faster with num_proc=<some int> sentences = sentences.map(lambda ex: {"label" : features["label"].str2int(ex["label"]) if ex["label"] is not None else None}, features=features) sentences = sentences.shuffle() ``` ## Expected results Properly tokenize dataset file `test.csv` without issues. ## Actual results Specify the actual results or traceback. ``` Downloading data files: 100% 2/2 [00:16<00:00, 7.34s/it] Downloading data: 100% 391M/391M [00:12<00:00, 36.6MB/s] Downloading data: 100% 92.4M/92.4M [00:02<00:00, 40.0MB/s] Extracting data files: 100% 2/2 [00:00<00:00, 47.66it/s] Dataset csv downloaded and prepared to /root/.cache/huggingface/datasets/csv/loretoparisi--tatoeba-sentences-efeff8965c730a2c/0.0.0/433e0ccc46f9880962cc2b12065189766fbb2bee57a221866138fb9203c83519. Subsequent calls will reuse this data. 100% 2/2 [00:00<00:00, 25.94it/s] 11% 942339/8256449 [01:55<13:11, 9245.85ex/s] --------------------------------------------------------------------------- KeyError Traceback (most recent call last) [<ipython-input-3-6a9867fad8d6>](https://localhost:8080/#) in <module>() 12 ) 13 # You can make this part faster with num_proc=<some int> ---> 14 sentences = sentences.map(lambda ex: {"label" : features["label"].str2int(ex["label"]) if ex["label"] is not None else None}, features=features) 15 sentences = sentences.shuffle() 10 frames [/usr/local/lib/python3.7/dist-packages/datasets/features/features.py](https://localhost:8080/#) in str2int(self, values) 852 if value not in self._str2int: 853 value = str(value).strip() --> 854 output.append(self._str2int[str(value)]) 855 else: 856 # No names provided, try to integerize KeyError: '\\N' ``` ## Environment info ``` - `datasets` version: 2.1.0 - Platform: Linux-5.4.144+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - PyArrow version: 6.0.1 - Pandas version: 1.3.5 - ``` ``` - `transformers` version: 4.18.0 - Platform: Linux-5.4.144+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - Huggingface_hub version: 0.5.1 - PyTorch version (GPU?): 1.11.0+cu113 (True) - Tensorflow version (GPU?): 2.8.0 (True) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using GPU in script?: <fill in> - Using distributed or parallel set-up in script?: <fill in> - ```
2022-04-27T16:29:25Z
4,238
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-27T10:42:11Z
https://api.github.com/repos/huggingface/datasets/issues/4238/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4238/timeline
Dataset caching policy
https://api.github.com/repos/huggingface/datasets/issues/4238/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/163333?v=4", "events_url": "https://api.github.com/users/loretoparisi/events{/privacy}", "followers_url": "https://api.github.com/users/loretoparisi/followers", "following_url": "https://api.github.com/users/loretoparisi/following{/other_user}", "gists_url": "https://api.github.com/users/loretoparisi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/loretoparisi", "id": 163333, "login": "loretoparisi", "node_id": "MDQ6VXNlcjE2MzMzMw==", "organizations_url": "https://api.github.com/users/loretoparisi/orgs", "received_events_url": "https://api.github.com/users/loretoparisi/received_events", "repos_url": "https://api.github.com/users/loretoparisi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/loretoparisi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/loretoparisi/subscriptions", "type": "User", "url": "https://api.github.com/users/loretoparisi" }
[]
null
completed
NONE
2022-04-27T16:28:50Z
null
I_kwDODunzps5IjIL7
[ "Hi @loretoparisi, thanks for reporting.\r\n\r\nThere is an option to force the redownload of the data files (and thus not using previously download and cached data files): `load_dataset(..., download_mode=\"force_redownload\")`.\r\n\r\nPlease, let me know if this fixes your problem.\r\n\r\nI can confirm you that your dataset loads without any problem for me:\r\n```python\r\nIn [2]: ds = load_dataset(\"loretoparisi/tatoeba-sentences\", data_files={\"train\": \"train.csv\", \"test\": \"test.csv\"}, delimiter=\"\\t\", column_names=['label', 'text'])\r\n\r\nIn [3]: ds\r\nOut[3]: \r\nDatasetDict({\r\n train: Dataset({\r\n features: ['label', 'text'],\r\n num_rows: 8256449\r\n })\r\n test: Dataset({\r\n features: ['label', 'text'],\r\n num_rows: 2061204\r\n })\r\n})\r\n``` ", "@albertvillanova thank you, it seems it still does not work using:\r\n\r\n```python\r\nsentences = load_dataset(\r\n \"loretoparisi/tatoeba-sentences\",\r\n data_files=data_files,\r\n delimiter='\\t', \r\n column_names=['label', 'text'],\r\n download_mode=\"force_redownload\"\r\n)\r\n```\r\n[This](https://colab.research.google.com/drive/1EA6FWo5pHxU8rPHHRn24NlHqRPiOlPTr?usp=sharing) is my notebook!\r\n\r\nThe problem is that the download file's revision for `test.csv` is not correctly parsed\r\n\r\n![Schermata 2022-04-27 alle 18 09 41](https://user-images.githubusercontent.com/163333/165563507-0be53eb6-8f61-49b0-b959-306e59281de3.png)\r\n\r\nIf you download that file `test.csv` from the repo, the line `\\\\N` is not there anymore (it was there at the first file upload).\r\n\r\nMy impression is that the Apache Arrow file is still cached - so server side, despite of enabling a forced download. For what I can see I get those two arrow files, but I cannot grep the bad line (`\\\\N`) since are binary files:\r\n\r\n```\r\n!ls -l /root/.cache/huggingface/datasets/csv/loretoparisi--tatoeba-sentences-efeff8965c730a2c/0.0.0/433e0ccc46f9880962cc2b12065189766fbb2bee57a221866138fb9203c83519\r\n!ls -l /root/.cache/huggingface/datasets/csv/loretoparisi--tatoeba-sentences-efeff8965c730a2c/0.0.0/433e0ccc46f9880962cc2b12065189766fbb2bee57a221866138fb9203c83519/csv-test.arrow\r\n!head /root/.cache/huggingface/datasets/csv/loretoparisi--tatoeba-sentences-efeff8965c730a2c/0.0.0/433e0ccc46f9880962cc2b12065189766fbb2bee57a221866138fb9203c83519/dataset_info.json\r\n```\r\n", "SOLVED! The problem was the with the file itself, using caching parameter helped indeed.\r\nThanks for helping!" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4238/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4238
https://github.com/huggingface/datasets/issues/4238
false
1,217,121,044
https://api.github.com/repos/huggingface/datasets/issues/4237/labels{/name}
https://huggingface.co/datasets/mozilla-foundation/common_voice_8_0
2022-05-10T12:17:05Z
4,237
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
2022-04-27T10:05:20Z
https://api.github.com/repos/huggingface/datasets/issues/4237/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4237/timeline
Common Voice 8 doesn't show datasets viewer
https://api.github.com/repos/huggingface/datasets/issues/4237/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[]
null
completed
CONTRIBUTOR
2022-05-10T12:17:04Z
null
I_kwDODunzps5Ii8sU
[ "Thanks for reporting. I understand it's an error in the dataset script. To reproduce:\r\n\r\n```python\r\n>>> import datasets as ds\r\n>>> split_names = ds.get_dataset_split_names(\"mozilla-foundation/common_voice_8_0\", use_auth_token=\"**********\")\r\nDownloading builder script: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 10.9k/10.9k [00:00<00:00, 10.9MB/s]\r\nDownloading extra modules: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2.98k/2.98k [00:00<00:00, 3.36MB/s]\r\nDownloading extra modules: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 53.1k/53.1k [00:00<00:00, 650kB/s]\r\nNo config specified, defaulting to: common_voice/en\r\nTraceback (most recent call last):\r\n File \"/home/slesage/hf/datasets-preview-backend/libs/libmodels/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 280, in get_dataset_config_info\r\n for split_generator in builder._split_generators(\r\n File \"/home/slesage/.cache/huggingface/modules/datasets_modules/datasets/mozilla-foundation--common_voice_8_0/720589e6e5ad674019008b719053303a71716db1b27e63c9846df02fdf93f2f3/common_voice_8_0.py\", line 153, in _split_generators\r\n self._log_download(self.config.name, bundle_version, hf_auth_token)\r\n File \"/home/slesage/.cache/huggingface/modules/datasets_modules/datasets/mozilla-foundation--common_voice_8_0/720589e6e5ad674019008b719053303a71716db1b27e63c9846df02fdf93f2f3/common_voice_8_0.py\", line 139, in _log_download\r\n email = HfApi().whoami(auth_token)[\"email\"]\r\nKeyError: 'email'\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/slesage/hf/datasets-preview-backend/libs/libmodels/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 323, in get_dataset_split_names\r\n info = get_dataset_config_info(\r\n File \"/home/slesage/hf/datasets-preview-backend/libs/libmodels/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 285, in get_dataset_config_info\r\n raise SplitsNotFoundError(\"The split names could not be parsed from the dataset config.\") from err\r\ndatasets.inspect.SplitsNotFoundError: The split names could not be parsed from the dataset config.\r\n```", "Thanks for reporting @patrickvonplaten and thanks for the investigation @severo.\r\n\r\nUnfortunately I'm not able to reproduce the error.\r\n\r\nI think the error has to do with authentication with `huggingface_hub`, because the exception is thrown from these code lines: https://huggingface.co/datasets/mozilla-foundation/common_voice_8_0/blob/main/common_voice_8_0.py#L137-L139\r\n```python\r\nfrom huggingface_hub import HfApi, HfFolder\r\n\r\nif isinstance(auth_token, bool):\r\n email = HfApi().whoami(auth_token)\r\nemail = HfApi().whoami(auth_token)[\"email\"]\r\n```\r\n\r\nCould you please verify the previous code with the `auth_token` you pass to `load_dataset(..., use_auth_token=auth_token,...`?", "OK, thanks for digging a bit into it. Indeed, the error occurs with the dataset-viewer, but not with a normal user token, because we use an app token, and it does not have a related email!\r\n\r\n```python\r\n>>> from huggingface_hub import HfApi, HfFolder\r\n>>> auth_token = \"hf_app_******\"\r\n>>> t = HfApi().whoami(auth_token)\r\n>>> t\r\n{'type': 'app', 'name': 'dataset-preview-backend'}\r\n>>> t[\"email\"]\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\nKeyError: 'email'\r\n```\r\n\r\nNote also that the doc (https://huggingface.co/docs/huggingface_hub/package_reference/hf_api#huggingface_hub.HfApi.whoami) does not state that `whoami` should return an `email` key.\r\n\r\n@SBrandeis @julien-c: do you think the app token should have an email associated, like the users?", "We can workaround this with\r\n```python\r\nemail = HfApi().whoami(auth_token).get(\"email\", \"[email protected]\")\r\n```\r\nin the common voice scripts", "Hmmm, does this mean that any person who downloads the common voice dataset will be logged as \"[email protected]\"? If so, it would defeat the purpose of sending the user's email to the commonvoice API, right?", "I agree with @severo: we cannot set our system email as default, allowing anybody not authenticated to by-pass the Common Voice usage policy.\r\n\r\nAdditionally, looking at the code, I think we should implement a more robust way to send user email to Common Voice: currently anybody can tweak the script and send somebody else email instead.\r\n\r\nCC: @patrickvonplaten @lhoestq @SBrandeis @julien-c ", "Hmm I don't agree here. \r\n\r\nAnybody can always just bypass the system by setting whatever email. As soon as someone has access to the downloading script it's trivial to tweak the code to not send the \"correct\" email but to just whatever and it would work.\r\n\r\nNote that someone only has visibility on the code after having \"signed\" the access-mechanism so I think we can expect the users to have agreed to not do anything malicious. \r\n\r\nI'm fine with both @lhoestq's solution or we find a way that forces the user to be logged in + being able to load the data for the datasets viewer. Wdyt @lhoestq @severo @albertvillanova ?", "> Additionally, looking at the code, I think we should implement a more robust way to send user email to Common Voice: currently anybody can tweak the script and send somebody else email instead.\r\n\r\nYes, I agree we can forget about this @patrickvonplaten. After having had a look at Common Voice website, I've seen they only require sending an email (no auth is inplace on their side, contrary to what I had previously thought). Therefore, currently we impose stronger requirements than them: we require the user having logged in and accepted the access mechanism.\r\n\r\nCurrently the script as it is already requires the user being logged in:\r\n```python\r\nHfApi().whoami(auth_token)\r\n```\r\nthrows an exception if None/invalid auth_token is passed.\r\n\r\nOn the other hand, we should agree on the way to allow the viewer to stream the data.", "The preview is back now, thanks !" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4237/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4237
https://github.com/huggingface/datasets/issues/4237
false
1,217,115,691
https://api.github.com/repos/huggingface/datasets/issues/4236/labels{/name}
This PR replaces the Google Drive URL with our Hub one, once the data owners have approved to host their data on the Hub. Moreover, this PR makes the dataset streamable. Fix #4217.
2022-06-10T08:10:55Z
4,236
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T10:01:13Z
https://api.github.com/repos/huggingface/datasets/issues/4236/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4236/timeline
Replace data URL in big_patent dataset and support streaming
https://api.github.com/repos/huggingface/datasets/issues/4236/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-05-02T18:21:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4236.diff", "html_url": "https://github.com/huggingface/datasets/pull/4236", "merged_at": "2022-05-02T18:21:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/4236.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4236" }
PR_kwDODunzps423MOc
[ "_The documentation is not available anymore as the PR was closed or merged._", "I first uploaded the data files to the Hub: I think it is a good option because we have git lfs to track versions and changes. Moreover people will be able to make PRs to propose updates on the data files.\r\n- I would have preferred to upload it it to the \"data\" org namespace, but it is already taken (although not used): might be possible to take it?\r\n\r\nAs an alternative (and to be consistent with previous datasets), I also uploaded the data files to our AWS bucket.\r\n\r\nWe should decide which to use (now and for future datasets) and set it here before merging. We should remove the data files for the non-chosen option.\r\n\r\nCC: @lhoestq @mariosasko @polinaeterna ", "Would it make sense to make the dataset a community one (so, create an organization for it) and store the script and the data in a single repository? Just as it is for most of the datasets. That way we can also access the data using a relative path inside the repo (that's not the point though). The point is that to me it seems a bit more straightforward to store everything in one place. \r\n\r\nI guess the strong argument against this logic is that in this case the canonical version won't work... But maybe there is some redirecting mechanism I don't know about? :)\r\n\r\nAnyway, I'm in favor of hosting data on the Hub instead of AWS :) ", "I also think storing everything in one place/single repository is the best option.\r\n\r\n@polinaeterna Canonical datasets also support data files (see the [`red_caps` repo](https://huggingface.co/datasets/red_caps/tree/main) for instance) ", "Thanks @polinaeterna and @mariosasko for your comments.\r\n\r\nYes, definitely it is much better to have everything in the same repo. \r\n\r\nI'm transferring their data files to the Hub under \"big_patent\" and deleting them from the other repo and AWS." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4236/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4236
https://github.com/huggingface/datasets/pull/4236
true
1,216,952,640
https://api.github.com/repos/huggingface/datasets/issues/4235/labels{/name}
### System Info ```shell I am using transformer trainer while meeting the issue. The trainer requests torch.utils.data.Dataset as input, which loads the whole dataset into the memory at once. Therefore, when the dataset is too large to load, there's nothing I can do except using IterDataset, which loads samples of data seperately, and results in low efficiency. I wonder if there are any tricks like Sharding in huggingface trainer. Looking forward to your reply. ``` ### Who can help? Trainer: @sgugger ### Information - [ ] The official example scripts - [ ] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [ ] My own task or dataset (give details below) ### Reproduction None ### Expected behavior ```shell I wonder if there are any tricks like fairseq Sharding very large datasets https://fairseq.readthedocs.io/en/latest/getting_started.html. Thanks a lot! ```
2023-07-25T15:07:57Z
4,235
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-27T07:50:13Z
https://api.github.com/repos/huggingface/datasets/issues/4235/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4235/timeline
How to load VERY LARGE dataset?
https://api.github.com/repos/huggingface/datasets/issues/4235/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/45160643?v=4", "events_url": "https://api.github.com/users/CaoYiqingT/events{/privacy}", "followers_url": "https://api.github.com/users/CaoYiqingT/followers", "following_url": "https://api.github.com/users/CaoYiqingT/following{/other_user}", "gists_url": "https://api.github.com/users/CaoYiqingT/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CaoYiqingT", "id": 45160643, "login": "CaoYiqingT", "node_id": "MDQ6VXNlcjQ1MTYwNjQz", "organizations_url": "https://api.github.com/users/CaoYiqingT/orgs", "received_events_url": "https://api.github.com/users/CaoYiqingT/received_events", "repos_url": "https://api.github.com/users/CaoYiqingT/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CaoYiqingT/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CaoYiqingT/subscriptions", "type": "User", "url": "https://api.github.com/users/CaoYiqingT" }
[]
null
completed
NONE
2023-07-25T15:07:57Z
null
I_kwDODunzps5IiTlA
[ "The `Trainer` support `IterableDataset`, not just datasets." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4235/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4235
https://github.com/huggingface/datasets/issues/4235
false
1,216,818,846
https://api.github.com/repos/huggingface/datasets/issues/4234/labels{/name}
Added autoeval config to imdb as pilot
2022-05-06T13:20:31Z
4,234
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T05:32:10Z
https://api.github.com/repos/huggingface/datasets/issues/4234/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4234/timeline
Autoeval config
https://api.github.com/repos/huggingface/datasets/issues/4234/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3278583?v=4", "events_url": "https://api.github.com/users/nazneenrajani/events{/privacy}", "followers_url": "https://api.github.com/users/nazneenrajani/followers", "following_url": "https://api.github.com/users/nazneenrajani/following{/other_user}", "gists_url": "https://api.github.com/users/nazneenrajani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nazneenrajani", "id": 3278583, "login": "nazneenrajani", "node_id": "MDQ6VXNlcjMyNzg1ODM=", "organizations_url": "https://api.github.com/users/nazneenrajani/orgs", "received_events_url": "https://api.github.com/users/nazneenrajani/received_events", "repos_url": "https://api.github.com/users/nazneenrajani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nazneenrajani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nazneenrajani/subscriptions", "type": "User", "url": "https://api.github.com/users/nazneenrajani" }
[]
null
null
NONE
2022-05-05T18:20:58Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4234.diff", "html_url": "https://github.com/huggingface/datasets/pull/4234", "merged_at": "2022-05-05T18:20:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/4234.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4234" }
PR_kwDODunzps422Mwn
[ "_The documentation is not available anymore as the PR was closed or merged._", "Related to: https://github.com/huggingface/autonlp-backend/issues/414 and https://github.com/huggingface/autonlp-backend/issues/424", "The tests are failing due to the changed metadata:\r\n\r\n```\r\ngot an unexpected keyword argument 'train-eval-index'\r\n```\r\n\r\nI think you can fix this by updating the `DatasetMetadata` class and implementing an appropriate `validate_train_eval_index()` function\r\n\r\n@lhoestq we are working with an arbitrary set of tags for `autoeval config`. See https://github.com/huggingface/autonlp-backend/issues/414\r\nI need to add a validator function though for the tests to pass. Our set is not well-defined as in the rest https://github.com/huggingface/datasets/tree/master/src/datasets/utils/resources. What's a workaround for this?", "On the question of validating the `train-eval-index` metadata, I think the simplest approach would be to validate that the required fields exist and not worry about their values (which are open-ended).\r\n\r\nFor me, the required fields include:\r\n\r\n* `config`\r\n* `task`\r\n* `task_id`\r\n* `splits` (train / validation / eval)\r\n* `col_mapping`\r\n* `metrics` (checking that each one has `type`, `name`) \r\n\r\nHere I'm using the spec defined in https://github.com/huggingface/autonlp-backend/issues/414 as a guide.\r\n\r\nWDYT @lhoestq ?", "Makes sense ! Currently the metadata type validator doesn't support subfields - let me open a PR to add it", "I ended up improving the metadata validation in this PR x)\r\n\r\nIn particular:\r\n- I added support YAML keys with dashes instead of underscores for `train-eval-index`\r\n- I added `train-eval-index` validation with `validate_train_eval_index`. It does nothing fancy, it just checks that it is a list if it exists in the YAML, but feel free to improve it if you want\r\n\r\nLet me know if it sounds good to you ! I think we can improve `validate_train_eval_index` in another PR", "Come on windows... I didn't do anything advanced...\r\n\r\nAnyway, will try to fix this when I get back home x)", "> Come on windows... I didn't do anything advanced...\r\n> \r\n> Anyway, will try to fix this when I get back home x)\r\n\r\nHehe, thanks!", "Thanks, @lhoestq this is great! ", "Did I just fix it for windows and now it fails on linux ? xD", "> Did I just fix it for windows and now it fails on linux ? xD\r\n\r\nLooks like the Heisenberg uncertainty principle is at play here - you cannot simultaneously have unit tests passing in both Linux and Windows 😅 ", "The worst is that the tests pass locally both on my windows and my linux x)", "Ok fixed it, the issue came from python 3.6 that doesn't return the right `__origin__` for Dict and List types", "> Alright thanks for adding the first Autoeval config ! :D\r\n\r\nWoohoo! Thank you so much 🤗 ", "This is cool!" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4234/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4234
https://github.com/huggingface/datasets/pull/4234
true
1,216,665,044
https://api.github.com/repos/huggingface/datasets/issues/4233/labels{/name}
null
2022-04-27T05:29:30Z
4,233
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T01:32:09Z
https://api.github.com/repos/huggingface/datasets/issues/4233/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4233/timeline
Autoeval
https://api.github.com/repos/huggingface/datasets/issues/4233/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3278583?v=4", "events_url": "https://api.github.com/users/nazneenrajani/events{/privacy}", "followers_url": "https://api.github.com/users/nazneenrajani/followers", "following_url": "https://api.github.com/users/nazneenrajani/following{/other_user}", "gists_url": "https://api.github.com/users/nazneenrajani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nazneenrajani", "id": 3278583, "login": "nazneenrajani", "node_id": "MDQ6VXNlcjMyNzg1ODM=", "organizations_url": "https://api.github.com/users/nazneenrajani/orgs", "received_events_url": "https://api.github.com/users/nazneenrajani/received_events", "repos_url": "https://api.github.com/users/nazneenrajani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nazneenrajani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nazneenrajani/subscriptions", "type": "User", "url": "https://api.github.com/users/nazneenrajani" }
[]
null
null
NONE
2022-04-27T01:32:23Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4233.diff", "html_url": "https://github.com/huggingface/datasets/pull/4233", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4233.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4233" }
PR_kwDODunzps421r-6
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4233). All of your documentation changes will be reflected on that endpoint." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4233/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4233
https://github.com/huggingface/datasets/pull/4233
true
1,216,659,444
https://api.github.com/repos/huggingface/datasets/issues/4232/labels{/name}
null
2022-05-03T14:23:56Z
4,232
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T01:21:09Z
https://api.github.com/repos/huggingface/datasets/issues/4232/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4232/timeline
adding new tag to tasks.json and modified for existing datasets
https://api.github.com/repos/huggingface/datasets/issues/4232/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3278583?v=4", "events_url": "https://api.github.com/users/nazneenrajani/events{/privacy}", "followers_url": "https://api.github.com/users/nazneenrajani/followers", "following_url": "https://api.github.com/users/nazneenrajani/following{/other_user}", "gists_url": "https://api.github.com/users/nazneenrajani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nazneenrajani", "id": 3278583, "login": "nazneenrajani", "node_id": "MDQ6VXNlcjMyNzg1ODM=", "organizations_url": "https://api.github.com/users/nazneenrajani/orgs", "received_events_url": "https://api.github.com/users/nazneenrajani/received_events", "repos_url": "https://api.github.com/users/nazneenrajani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nazneenrajani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nazneenrajani/subscriptions", "type": "User", "url": "https://api.github.com/users/nazneenrajani" }
[]
null
null
NONE
2022-05-03T14:16:39Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4232.diff", "html_url": "https://github.com/huggingface/datasets/pull/4232", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4232.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4232" }
PR_kwDODunzps421qz4
[ "_The documentation is not available anymore as the PR was closed or merged._", "closing in favor of https://github.com/huggingface/datasets/pull/4244" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4232/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4232
https://github.com/huggingface/datasets/pull/4232
true
1,216,651,960
https://api.github.com/repos/huggingface/datasets/issues/4231/labels{/name}
The CC-Aligned dataset url has changed to https://data.statmt.org/cc-aligned/, the old address http://www.statmt.org/cc-aligned/ is no longer valid
2022-05-16T17:01:13Z
4,231
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T01:07:01Z
https://api.github.com/repos/huggingface/datasets/issues/4231/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4231/timeline
Fix invalid url to CC-Aligned dataset
https://api.github.com/repos/huggingface/datasets/issues/4231/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/44451229?v=4", "events_url": "https://api.github.com/users/juntang-zhuang/events{/privacy}", "followers_url": "https://api.github.com/users/juntang-zhuang/followers", "following_url": "https://api.github.com/users/juntang-zhuang/following{/other_user}", "gists_url": "https://api.github.com/users/juntang-zhuang/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/juntang-zhuang", "id": 44451229, "login": "juntang-zhuang", "node_id": "MDQ6VXNlcjQ0NDUxMjI5", "organizations_url": "https://api.github.com/users/juntang-zhuang/orgs", "received_events_url": "https://api.github.com/users/juntang-zhuang/received_events", "repos_url": "https://api.github.com/users/juntang-zhuang/repos", "site_admin": false, "starred_url": "https://api.github.com/users/juntang-zhuang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/juntang-zhuang/subscriptions", "type": "User", "url": "https://api.github.com/users/juntang-zhuang" }
[]
null
null
CONTRIBUTOR
2022-05-16T16:53:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4231.diff", "html_url": "https://github.com/huggingface/datasets/pull/4231", "merged_at": "2022-05-16T16:53:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/4231.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4231" }
PR_kwDODunzps421pUX
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4231/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4231
https://github.com/huggingface/datasets/pull/4231
true
1,216,643,661
https://api.github.com/repos/huggingface/datasets/issues/4230/labels{/name}
![image](https://user-images.githubusercontent.com/37113676/165416606-96b5db18-b16c-4b6b-928c-de8620fd943e.png) But on huggingface datasets: ![image](https://user-images.githubusercontent.com/37113676/165416649-8fd77980-ca0d-43f0-935e-f398ba8323a4.png) Where is the German data?
2023-07-25T15:10:15Z
4,230
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-04-27T00:53:52Z
https://api.github.com/repos/huggingface/datasets/issues/4230/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4230/timeline
Why the `conll2003` dataset on huggingface only contains the `en` subset? Where is the German data?
https://api.github.com/repos/huggingface/datasets/issues/4230/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/37113676?v=4", "events_url": "https://api.github.com/users/beyondguo/events{/privacy}", "followers_url": "https://api.github.com/users/beyondguo/followers", "following_url": "https://api.github.com/users/beyondguo/following{/other_user}", "gists_url": "https://api.github.com/users/beyondguo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/beyondguo", "id": 37113676, "login": "beyondguo", "node_id": "MDQ6VXNlcjM3MTEzNjc2", "organizations_url": "https://api.github.com/users/beyondguo/orgs", "received_events_url": "https://api.github.com/users/beyondguo/received_events", "repos_url": "https://api.github.com/users/beyondguo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/beyondguo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/beyondguo/subscriptions", "type": "User", "url": "https://api.github.com/users/beyondguo" }
[]
null
completed
NONE
2023-07-25T15:10:15Z
null
I_kwDODunzps5IhIJN
[ "Thanks for reporting @beyondguo.\r\n\r\nIndeed, we generate this dataset from this raw data file URL: https://data.deepai.org/conll2003.zip\r\nAnd that URL only contains the English version.", "The German data requires payment\r\n\r\nThe [original task page](https://www.clips.uantwerpen.be/conll2003/ner/) states \"The German data is a collection of articles from the Frankfurter Rundschau. The named entities have been annotated by people of the University of Antwerp. Only the annotations are available here. In order to build these data sets you need access to the ECI Multilingual Text Corpus. It can be ordered from the Linguistic Data Consortium (2003 non-member price: US$ 35.00).\"\r\n\r\nInflation since 2003 has also affected LDC's prices, and today the dataset [LDC94T5](https://catalog.ldc.upenn.edu/LDC94T5) is available under license for $75 a copy. The [license](https://catalog.ldc.upenn.edu/license/eci-slash-mci-user-agreement.pdf) includes a non-distribution condition, which is probably why the data has not turned up openly.\r\n\r\nThe ACL hold copyright of this data; I'll mail them and anyone I can find at ECI to see if they'll open this up now. After all, it worked with Microsoft 3DMM, why not here too, after 28 years? :)\r\n", "Closing this issue as we are not allowed to share publicly the German subset." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4230/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4230
https://github.com/huggingface/datasets/issues/4230
false
1,216,638,968
https://api.github.com/repos/huggingface/datasets/issues/4229/labels{/name}
multi-input-text-classification tag for classification datasets that take more than one input
2022-04-27T00:48:28Z
4,229
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-27T00:47:08Z
https://api.github.com/repos/huggingface/datasets/issues/4229/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4229/timeline
new task tag
https://api.github.com/repos/huggingface/datasets/issues/4229/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3278583?v=4", "events_url": "https://api.github.com/users/nazneenrajani/events{/privacy}", "followers_url": "https://api.github.com/users/nazneenrajani/followers", "following_url": "https://api.github.com/users/nazneenrajani/following{/other_user}", "gists_url": "https://api.github.com/users/nazneenrajani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nazneenrajani", "id": 3278583, "login": "nazneenrajani", "node_id": "MDQ6VXNlcjMyNzg1ODM=", "organizations_url": "https://api.github.com/users/nazneenrajani/orgs", "received_events_url": "https://api.github.com/users/nazneenrajani/received_events", "repos_url": "https://api.github.com/users/nazneenrajani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nazneenrajani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nazneenrajani/subscriptions", "type": "User", "url": "https://api.github.com/users/nazneenrajani" }
[]
null
null
NONE
2022-04-27T00:48:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4229.diff", "html_url": "https://github.com/huggingface/datasets/pull/4229", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4229.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4229" }
PR_kwDODunzps421mjM
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4229/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4229
https://github.com/huggingface/datasets/pull/4229
true
1,216,523,043
https://api.github.com/repos/huggingface/datasets/issues/4228/labels{/name}
multi-input-text-classification tag for classification datasets that take more than one input
2022-04-27T00:48:31Z
4,228
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-26T22:00:33Z
https://api.github.com/repos/huggingface/datasets/issues/4228/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4228/timeline
new task tag
https://api.github.com/repos/huggingface/datasets/issues/4228/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3278583?v=4", "events_url": "https://api.github.com/users/nazneenrajani/events{/privacy}", "followers_url": "https://api.github.com/users/nazneenrajani/followers", "following_url": "https://api.github.com/users/nazneenrajani/following{/other_user}", "gists_url": "https://api.github.com/users/nazneenrajani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nazneenrajani", "id": 3278583, "login": "nazneenrajani", "node_id": "MDQ6VXNlcjMyNzg1ODM=", "organizations_url": "https://api.github.com/users/nazneenrajani/orgs", "received_events_url": "https://api.github.com/users/nazneenrajani/received_events", "repos_url": "https://api.github.com/users/nazneenrajani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nazneenrajani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nazneenrajani/subscriptions", "type": "User", "url": "https://api.github.com/users/nazneenrajani" }
[]
null
null
NONE
2022-04-27T00:46:31Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4228.diff", "html_url": "https://github.com/huggingface/datasets/pull/4228", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4228.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4228" }
PR_kwDODunzps421NKL
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4228/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4228
https://github.com/huggingface/datasets/pull/4228
true
1,216,455,316
https://api.github.com/repos/huggingface/datasets/issues/4227/labels{/name}
null
2022-05-03T12:50:23Z
4,227
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-26T20:41:03Z
https://api.github.com/repos/huggingface/datasets/issues/4227/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4227/timeline
Add f1 metric card, update docstring in py file
https://api.github.com/repos/huggingface/datasets/issues/4227/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
null
null
CONTRIBUTOR
2022-05-03T12:43:33Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4227.diff", "html_url": "https://github.com/huggingface/datasets/pull/4227", "merged_at": "2022-05-03T12:43:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/4227.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4227" }
PR_kwDODunzps420-mc
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4227/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4227
https://github.com/huggingface/datasets/pull/4227
true
1,216,331,073
https://api.github.com/repos/huggingface/datasets/issues/4226/labels{/name}
- adds pearsonr metric card - adds ability to return p-value - p-value was mentioned in the original docs as a return value, but there was no option to return it. I updated the _compute function slightly to have an option to return the p-value.
2022-05-03T17:09:24Z
4,226
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-26T18:30:46Z
https://api.github.com/repos/huggingface/datasets/issues/4226/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4226/timeline
Add pearsonr mc, update functionality to match the original docs
https://api.github.com/repos/huggingface/datasets/issues/4226/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
null
null
CONTRIBUTOR
2022-05-03T17:02:28Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4226.diff", "html_url": "https://github.com/huggingface/datasets/pull/4226", "merged_at": "2022-05-03T17:02:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/4226.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4226" }
PR_kwDODunzps420kAv
[ "_The documentation is not available anymore as the PR was closed or merged._", "thank you @lhoestq!! :hugs: " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4226/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4226
https://github.com/huggingface/datasets/pull/4226
true
1,216,213,464
https://api.github.com/repos/huggingface/datasets/issues/4225/labels{/name}
add train eval index for autoeval
2022-04-27T00:48:31Z
4,225
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-26T16:38:34Z
https://api.github.com/repos/huggingface/datasets/issues/4225/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4225/timeline
autoeval config
https://api.github.com/repos/huggingface/datasets/issues/4225/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3278583?v=4", "events_url": "https://api.github.com/users/nazneenrajani/events{/privacy}", "followers_url": "https://api.github.com/users/nazneenrajani/followers", "following_url": "https://api.github.com/users/nazneenrajani/following{/other_user}", "gists_url": "https://api.github.com/users/nazneenrajani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nazneenrajani", "id": 3278583, "login": "nazneenrajani", "node_id": "MDQ6VXNlcjMyNzg1ODM=", "organizations_url": "https://api.github.com/users/nazneenrajani/orgs", "received_events_url": "https://api.github.com/users/nazneenrajani/received_events", "repos_url": "https://api.github.com/users/nazneenrajani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nazneenrajani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nazneenrajani/subscriptions", "type": "User", "url": "https://api.github.com/users/nazneenrajani" }
[]
null
null
NONE
2022-04-26T22:00:26Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4225.diff", "html_url": "https://github.com/huggingface/datasets/pull/4225", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4225.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4225" }
PR_kwDODunzps420LNM
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4225/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4225
https://github.com/huggingface/datasets/pull/4225
true
1,216,209,667
https://api.github.com/repos/huggingface/datasets/issues/4224/labels{/name}
add train eval index for autoeval
2022-04-26T16:36:45Z
4,224
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-26T16:35:19Z
https://api.github.com/repos/huggingface/datasets/issues/4224/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4224/timeline
autoeval config
https://api.github.com/repos/huggingface/datasets/issues/4224/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3278583?v=4", "events_url": "https://api.github.com/users/nazneenrajani/events{/privacy}", "followers_url": "https://api.github.com/users/nazneenrajani/followers", "following_url": "https://api.github.com/users/nazneenrajani/following{/other_user}", "gists_url": "https://api.github.com/users/nazneenrajani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nazneenrajani", "id": 3278583, "login": "nazneenrajani", "node_id": "MDQ6VXNlcjMyNzg1ODM=", "organizations_url": "https://api.github.com/users/nazneenrajani/orgs", "received_events_url": "https://api.github.com/users/nazneenrajani/received_events", "repos_url": "https://api.github.com/users/nazneenrajani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nazneenrajani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nazneenrajani/subscriptions", "type": "User", "url": "https://api.github.com/users/nazneenrajani" }
[]
null
null
NONE
2022-04-26T16:36:45Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4224.diff", "html_url": "https://github.com/huggingface/datasets/pull/4224", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4224.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4224" }
PR_kwDODunzps420KX2
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4224/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4224
https://github.com/huggingface/datasets/pull/4224
true
1,216,107,082
https://api.github.com/repos/huggingface/datasets/issues/4223/labels{/name}
- adds accuracy metric card - updates docstring in accuracy.py - adds .json file with metric card and docstring information
2022-05-03T14:27:45Z
4,223
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-26T15:10:46Z
https://api.github.com/repos/huggingface/datasets/issues/4223/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4223/timeline
Add Accuracy Metric Card
https://api.github.com/repos/huggingface/datasets/issues/4223/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
null
null
CONTRIBUTOR
2022-05-03T14:20:47Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4223.diff", "html_url": "https://github.com/huggingface/datasets/pull/4223", "merged_at": "2022-05-03T14:20:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/4223.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4223" }
PR_kwDODunzps42z0YV
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4223/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4223
https://github.com/huggingface/datasets/pull/4223
true
1,216,056,439
https://api.github.com/repos/huggingface/datasets/issues/4222/labels{/name}
I noticed many links were not properly displayed (only text, no link) on the Hub because of wrong syntax, e.g.: https://huggingface.co/datasets/big_patent This PR fixes all description links in dataset cards.
2022-05-06T08:38:38Z
4,222
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-26T14:36:25Z
https://api.github.com/repos/huggingface/datasets/issues/4222/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4222/timeline
Fix description links in dataset cards
https://api.github.com/repos/huggingface/datasets/issues/4222/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-04-26T16:52:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4222.diff", "html_url": "https://github.com/huggingface/datasets/pull/4222", "merged_at": "2022-04-26T16:52:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/4222.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4222" }
PR_kwDODunzps42zpcd
[ "_The documentation is not available anymore as the PR was closed or merged._", "Non passing tests are due to other pre-existing errors in dataset cards: not related to this PR." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4222/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4222
https://github.com/huggingface/datasets/pull/4222
true
1,215,911,182
https://api.github.com/repos/huggingface/datasets/issues/4221/labels{/name}
Hi, I'm trying to create the loading script for a dataset in which one feature is a list of dictionaries, which afaik doesn't fit very well the values and structures supported by Value and Sequence. Is there any suggested workaround, am I missing something? Thank you in advance.
2022-04-29T14:52:19Z
4,221
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d876e3", "default": true, "description": "Further information is requested", "id": 1935892912, "name": "question", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question" } ]
2022-04-26T12:50:18Z
https://api.github.com/repos/huggingface/datasets/issues/4221/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4221/timeline
Dictionary Feature
https://api.github.com/repos/huggingface/datasets/issues/4221/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/2944532?v=4", "events_url": "https://api.github.com/users/jordiae/events{/privacy}", "followers_url": "https://api.github.com/users/jordiae/followers", "following_url": "https://api.github.com/users/jordiae/following{/other_user}", "gists_url": "https://api.github.com/users/jordiae/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jordiae", "id": 2944532, "login": "jordiae", "node_id": "MDQ6VXNlcjI5NDQ1MzI=", "organizations_url": "https://api.github.com/users/jordiae/orgs", "received_events_url": "https://api.github.com/users/jordiae/received_events", "repos_url": "https://api.github.com/users/jordiae/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jordiae/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jordiae/subscriptions", "type": "User", "url": "https://api.github.com/users/jordiae" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
completed
NONE
2022-04-28T17:04:58Z
null
I_kwDODunzps5IeVUO
[ "Hi @jordiae,\r\n\r\nInstead of the `Sequence` feature, you can use just a regular list: put the dict between `[` and `]`:\r\n```python\r\n\"list_of_dict_feature\": [\r\n {\r\n \"key1_in_dict\": datasets.Value(\"string\"),\r\n \"key2_in_dict\": datasets.Value(\"int32\"),\r\n ...\r\n }\r\n],\r\n```\r\n\r\nFeel free to re-open this issue if that does not work for your use case.", "> Hi @jordiae,\r\n> \r\n> Instead of the `Sequence` feature, you can use just a regular list: put the dict between `[` and `]`:\r\n> \r\n> ```python\r\n> \"list_of_dict_feature\": [\r\n> {\r\n> \"key1_in_dict\": datasets.Value(\"string\"),\r\n> \"key2_in_dict\": datasets.Value(\"int32\"),\r\n> ...\r\n> }\r\n> ],\r\n> ```\r\n> \r\n> Feel free to re-open this issue if that does not work for your use case.\r\n\r\nThank you" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4221/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4221
https://github.com/huggingface/datasets/issues/4221
false
1,215,225,802
https://api.github.com/repos/huggingface/datasets/issues/4220/labels{/name}
null
2022-05-09T17:29:34Z
4,220
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-26T01:20:43Z
https://api.github.com/repos/huggingface/datasets/issues/4220/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4220/timeline
Altered faiss installation comment
https://api.github.com/repos/huggingface/datasets/issues/4220/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36671559?v=4", "events_url": "https://api.github.com/users/vishalsrao/events{/privacy}", "followers_url": "https://api.github.com/users/vishalsrao/followers", "following_url": "https://api.github.com/users/vishalsrao/following{/other_user}", "gists_url": "https://api.github.com/users/vishalsrao/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vishalsrao", "id": 36671559, "login": "vishalsrao", "node_id": "MDQ6VXNlcjM2NjcxNTU5", "organizations_url": "https://api.github.com/users/vishalsrao/orgs", "received_events_url": "https://api.github.com/users/vishalsrao/received_events", "repos_url": "https://api.github.com/users/vishalsrao/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vishalsrao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vishalsrao/subscriptions", "type": "User", "url": "https://api.github.com/users/vishalsrao" }
[]
null
null
CONTRIBUTOR
2022-05-09T17:22:09Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4220.diff", "html_url": "https://github.com/huggingface/datasets/pull/4220", "merged_at": "2022-05-09T17:22:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/4220.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4220" }
PR_kwDODunzps42w5YO
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hi ! Can you explain why this change is needed ?", "Facebook recommends installing FAISS using conda (https://github.com/facebookresearch/faiss/blob/main/INSTALL.md). pip does not seem to have the latest version of FAISS. The latest version of faiss is 1.7.2 (https://anaconda.org/conda-forge/faiss), but the latest one available through pip is 1.5.3 (https://pypi.org/project/faiss/). " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4220/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4220
https://github.com/huggingface/datasets/pull/4220
true
1,214,934,025
https://api.github.com/repos/huggingface/datasets/issues/4219/labels{/name}
null
2022-04-26T20:44:18Z
4,219
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-25T19:14:56Z
https://api.github.com/repos/huggingface/datasets/issues/4219/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4219/timeline
Add F1 Metric Card
https://api.github.com/repos/huggingface/datasets/issues/4219/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
null
null
CONTRIBUTOR
2022-04-26T20:37:46Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4219.diff", "html_url": "https://github.com/huggingface/datasets/pull/4219", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4219.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4219" }
PR_kwDODunzps42v6rE
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4219/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4219
https://github.com/huggingface/datasets/pull/4219
true
1,214,748,226
https://api.github.com/repos/huggingface/datasets/issues/4218/labels{/name}
Fix #4199
2022-04-26T17:00:24Z
4,218
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-25T16:17:59Z
https://api.github.com/repos/huggingface/datasets/issues/4218/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4218/timeline
Make code for image downloading from image urls cacheable
https://api.github.com/repos/huggingface/datasets/issues/4218/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-04-26T13:38:26Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4218.diff", "html_url": "https://github.com/huggingface/datasets/pull/4218", "merged_at": "2022-04-26T13:38:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/4218.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4218" }
PR_kwDODunzps42vTA0
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4218/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4218
https://github.com/huggingface/datasets/pull/4218
true
1,214,688,141
https://api.github.com/repos/huggingface/datasets/issues/4217/labels{/name}
## Dataset viewer issue for '*big_patent*' **Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/big_patent/viewer/all/train)* *Unable to view because it says FileNotFound, also cannot download it through the python API* Am I the one who added this dataset ? No
2022-05-26T06:29:43Z
4,217
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "8B51EF", "default": false, "description": "", "id": 4069435429, "name": "hosted-on-google-drive", "node_id": "LA_kwDODunzps7yjqgl", "url": "https://api.github.com/repos/huggingface/datasets/labels/hosted-on-google-drive" } ]
2022-04-25T15:31:45Z
https://api.github.com/repos/huggingface/datasets/issues/4217/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4217/timeline
Big_Patent dataset broken
https://api.github.com/repos/huggingface/datasets/issues/4217/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/54189843?v=4", "events_url": "https://api.github.com/users/Matthew-Larsen/events{/privacy}", "followers_url": "https://api.github.com/users/Matthew-Larsen/followers", "following_url": "https://api.github.com/users/Matthew-Larsen/following{/other_user}", "gists_url": "https://api.github.com/users/Matthew-Larsen/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Matthew-Larsen", "id": 54189843, "login": "Matthew-Larsen", "node_id": "MDQ6VXNlcjU0MTg5ODQz", "organizations_url": "https://api.github.com/users/Matthew-Larsen/orgs", "received_events_url": "https://api.github.com/users/Matthew-Larsen/received_events", "repos_url": "https://api.github.com/users/Matthew-Larsen/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Matthew-Larsen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Matthew-Larsen/subscriptions", "type": "User", "url": "https://api.github.com/users/Matthew-Larsen" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
completed
NONE
2022-05-02T18:21:15Z
null
I_kwDODunzps5IZquN
[ "Thanks for reporting. The issue seems not to be directly related to the dataset viewer or the `datasets` library, but instead to it being hosted on Google Drive.\r\n\r\nSee related issues: https://github.com/huggingface/datasets/issues?q=is%3Aissue+is%3Aopen+drive.google.com\r\n\r\nTo quote [@lhoestq](https://github.com/huggingface/datasets/issues/4075#issuecomment-1087362551):\r\n\r\n> PS: if possible, please try to not use Google Drive links in your dataset script, since Google Drive has download quotas and is not always reliable.\r\n\r\n", "We should find out if the dataset license allows redistribution and contact the data owners to propose them to host their data on our Hub.", "The data owners have agreed on hosting their data on the Hub." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4217/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4217
https://github.com/huggingface/datasets/issues/4217
false
1,214,614,029
https://api.github.com/repos/huggingface/datasets/issues/4216/labels{/name}
I noticed this bug while answering [this question](https://discuss.huggingface.co/t/correct-way-to-create-a-dataset-from-a-csv-file/15686/11?u=mariosasko). This code replicates the bug: ```python from datasets import Dataset dset = Dataset.from_dict({"en": ["aa", "bb"], "fr": ["cc", "dd"]}) dset.map(lambda ex: {"translation": ex}) ``` and this is the fix for it (before this PR): ```python from datasets import Dataset dset = Dataset.from_dict({"en": ["aa", "bb"], "fr": ["cc", "dd"]}) dset.map(lambda ex: {"translation": dict(ex)}) ``` Internally, this can be fixed by merging two dicts via dict unpacking (instead of `dict.update) `in `Dataset.map`, which avoids creating recursive dictionaries. P.S. `{**a, **b}` is slightly more performant than `a.update(b)` in my bencmarks.
2022-05-04T17:20:06Z
4,216
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-25T14:40:32Z
https://api.github.com/repos/huggingface/datasets/issues/4216/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4216/timeline
Avoid recursion error in map if example is returned as dict value
https://api.github.com/repos/huggingface/datasets/issues/4216/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-05-04T17:12:52Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4216.diff", "html_url": "https://github.com/huggingface/datasets/pull/4216", "merged_at": "2022-05-04T17:12:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/4216.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4216" }
PR_kwDODunzps42u1_w
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4216/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4216
https://github.com/huggingface/datasets/pull/4216
true
1,214,579,162
https://api.github.com/repos/huggingface/datasets/issues/4215/labels{/name}
Addresses this comment: https://github.com/huggingface/datasets/pull/3801#pullrequestreview-901736921
2022-05-03T15:56:07Z
4,215
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-25T14:15:19Z
https://api.github.com/repos/huggingface/datasets/issues/4215/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4215/timeline
Add `drop_last_batch` to `IterableDataset.map`
https://api.github.com/repos/huggingface/datasets/issues/4215/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-05-03T15:48:54Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4215.diff", "html_url": "https://github.com/huggingface/datasets/pull/4215", "merged_at": "2022-05-03T15:48:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/4215.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4215" }
PR_kwDODunzps42uuhY
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4215/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4215
https://github.com/huggingface/datasets/pull/4215
true
1,214,572,430
https://api.github.com/repos/huggingface/datasets/issues/4214/labels{/name}
Avoids having to set `ignore_verifications=True` in `load_dataset("imagefolder", ...)` to skip checksum verification and speed up loading. The user can still pass `DownloadConfig(record_checksums=True)` to not skip this part.
2022-05-03T15:28:32Z
4,214
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-25T14:10:41Z
https://api.github.com/repos/huggingface/datasets/issues/4214/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4214/timeline
Skip checksum computation in Imagefolder by default
https://api.github.com/repos/huggingface/datasets/issues/4214/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-05-03T15:21:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4214.diff", "html_url": "https://github.com/huggingface/datasets/pull/4214", "merged_at": "2022-05-03T15:21:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/4214.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4214" }
PR_kwDODunzps42utC5
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4214/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4214
https://github.com/huggingface/datasets/pull/4214
true
1,214,510,010
https://api.github.com/repos/huggingface/datasets/issues/4213/labels{/name}
Ready for review.
2022-05-05T12:19:21Z
4,213
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-25T13:26:18Z
https://api.github.com/repos/huggingface/datasets/issues/4213/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4213/timeline
ETT time series dataset
https://api.github.com/repos/huggingface/datasets/issues/4213/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8100?v=4", "events_url": "https://api.github.com/users/kashif/events{/privacy}", "followers_url": "https://api.github.com/users/kashif/followers", "following_url": "https://api.github.com/users/kashif/following{/other_user}", "gists_url": "https://api.github.com/users/kashif/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kashif", "id": 8100, "login": "kashif", "node_id": "MDQ6VXNlcjgxMDA=", "organizations_url": "https://api.github.com/users/kashif/orgs", "received_events_url": "https://api.github.com/users/kashif/received_events", "repos_url": "https://api.github.com/users/kashif/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kashif/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kashif/subscriptions", "type": "User", "url": "https://api.github.com/users/kashif" }
[]
null
null
CONTRIBUTOR
2022-05-05T12:10:35Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4213.diff", "html_url": "https://github.com/huggingface/datasets/pull/4213", "merged_at": "2022-05-05T12:10:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/4213.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4213" }
PR_kwDODunzps42uft_
[ "_The documentation is not available anymore as the PR was closed or merged._", "thank you!\r\n" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4213/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4213
https://github.com/huggingface/datasets/pull/4213
true
1,214,498,582
https://api.github.com/repos/huggingface/datasets/issues/4212/labels{/name}
`path` should be set to local path inside audio feature if exist so that bytes can correctly be deleted.
2022-04-26T22:54:28Z
4,212
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-25T13:18:26Z
https://api.github.com/repos/huggingface/datasets/issues/4212/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4212/timeline
[Common Voice] Make sure bytes are correctly deleted if `path` exists
https://api.github.com/repos/huggingface/datasets/issues/4212/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[]
null
null
CONTRIBUTOR
2022-04-26T22:48:27Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4212.diff", "html_url": "https://github.com/huggingface/datasets/pull/4212", "merged_at": "2022-04-26T22:48:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/4212.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4212" }
PR_kwDODunzps42udRf
[ "_The documentation is not available anymore as the PR was closed or merged._", "cool that you noticed that we store unnecessary bytes again :D " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4212/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4212
https://github.com/huggingface/datasets/pull/4212
true
1,214,361,837
https://api.github.com/repos/huggingface/datasets/issues/4211/labels{/name}
Hi there, I am trying to load a dataset to the Hub. This dataset is a `DatasetDict` composed of various splits. Some splits have a different `Feature` mapping. Locally, the DatasetDict preserves the individual features but if I `push_to_hub` and then `load_dataset`, the features are all the same. Dataset and code to reproduce available [here](https://huggingface.co/datasets/pietrolesci/robust_nli). In short: I have 3 feature mapping ```python Tri_features = Features( { "idx": Value(dtype="int64"), "premise": Value(dtype="string"), "hypothesis": Value(dtype="string"), "label": ClassLabel(num_classes=3, names=["entailment", "neutral", "contradiction"]), } ) Ent_features = Features( { "idx": Value(dtype="int64"), "premise": Value(dtype="string"), "hypothesis": Value(dtype="string"), "label": ClassLabel(num_classes=2, names=["non-entailment", "entailment"]), } ) Con_features = Features( { "idx": Value(dtype="int64"), "premise": Value(dtype="string"), "hypothesis": Value(dtype="string"), "label": ClassLabel(num_classes=2, names=["non-contradiction", "contradiction"]), } ) ``` Then I create different datasets ```python dataset_splits = {} for split in df["split"].unique(): print(split) df_split = df.loc[df["split"] == split].copy() if split in Tri_dataset: df_split["label"] = df_split["label"].map({"entailment": 0, "neutral": 1, "contradiction": 2}) ds = Dataset.from_pandas(df_split, features=Tri_features) elif split in Ent_bin_dataset: df_split["label"] = df_split["label"].map({"non-entailment": 0, "entailment": 1}) ds = Dataset.from_pandas(df_split, features=Ent_features) elif split in Con_bin_dataset: df_split["label"] = df_split["label"].map({"non-contradiction": 0, "contradiction": 1}) ds = Dataset.from_pandas(df_split, features=Con_features) else: print("ERROR:", split) dataset_splits[split] = ds datasets = DatasetDict(dataset_splits) ``` I then push to hub ```python datasets.push_to_hub("pietrolesci/robust_nli", token="<token>") ``` Finally, I load it from the hub ```python datasets_loaded_from_hub = load_dataset("pietrolesci/robust_nli") ``` And I get that ```python datasets["LI_TS"].features != datasets_loaded_from_hub["LI_TS"].features ``` since ```python "label": ClassLabel(num_classes=2, names=["non-contradiction", "contradiction"]) ``` gets remapped to ```python "label": ClassLabel(num_classes=3, names=["entailment", "neutral", "contradiction"]) ```
2023-04-06T19:25:50Z
4,211
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-25T11:22:54Z
https://api.github.com/repos/huggingface/datasets/issues/4211/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
https://api.github.com/repos/huggingface/datasets/issues/4211/timeline
DatasetDict containing Datasets with different features when pushed to hub gets remapped features
https://api.github.com/repos/huggingface/datasets/issues/4211/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/61748653?v=4", "events_url": "https://api.github.com/users/pietrolesci/events{/privacy}", "followers_url": "https://api.github.com/users/pietrolesci/followers", "following_url": "https://api.github.com/users/pietrolesci/following{/other_user}", "gists_url": "https://api.github.com/users/pietrolesci/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/pietrolesci", "id": 61748653, "login": "pietrolesci", "node_id": "MDQ6VXNlcjYxNzQ4NjUz", "organizations_url": "https://api.github.com/users/pietrolesci/orgs", "received_events_url": "https://api.github.com/users/pietrolesci/received_events", "repos_url": "https://api.github.com/users/pietrolesci/repos", "site_admin": false, "starred_url": "https://api.github.com/users/pietrolesci/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pietrolesci/subscriptions", "type": "User", "url": "https://api.github.com/users/pietrolesci" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" } ]
null
completed
NONE
2022-05-20T15:15:30Z
null
I_kwDODunzps5IYbDt
[ "Hi @pietrolesci, thanks for reporting.\r\n\r\nPlease note that this is a design purpose: a `DatasetDict` has the same features for all its datasets. Normally, a `DatasetDict` is composed of several sub-datasets each corresponding to a different **split**.\r\n\r\nTo handle sub-datasets with different features, we use another approach: use different **configurations** instead of **splits**.\r\n\r\nHowever, for the moment `push_to_hub` does not support specifying different configurations. IMHO, we should implement this.", "Hi @albertvillanova,\r\n\r\nThanks a lot for your reply! I got it now. The strange thing for me was to have it correctly working (i.e., DatasetDict with different features in some datasets) locally and not on the Hub. It would be great to have configuration supported by `push_to_hub`. Personally, this latter functionality allowed me to iterate rather quickly on dataset curation.\r\n\r\nAgain, thanks for your time @albertvillanova!\r\n\r\nBest,\r\nPietro", "Hi! Yes, we should override `DatasetDict.__setitem__` and throw an error if features dictionaries are different. `DatasetDict` is a subclass of `dict`, so `DatasetDict.{update/setdefault}` need to be overridden as well. We could avoid this by subclassing `UserDict`, but then we would get the name collision - `DatasetDict.data` vs. `UserDict.data`. This makes me think we should rename the `data` attribute of `DatasetDict`/`Dataset` for easier dict subclassing (would also simplify https://github.com/huggingface/datasets/pull/3997) and to follow good Python practices. Another option is to have a custom `UserDict` class in `py_utils`, but it can be hard to keep this class consistent with the built-in `UserDict`. \r\n\r\n@albertvillanova @lhoestq wdyt?", "I would keep things simple and keep subclassing dict. Regarding the features check, I guess this can be done only for `push_to_hub` right ? It is the only function right now that requires the underlying datasets to be splits (e.g. train/test) and have the same features.\r\n\r\nNote that later you will be able to push datasets with different features as different dataset **configurations** (similarly to the [GLUE subsets](https://huggingface.co/datasets/glue) for example). We will work on this soon", "Hi @lhoestq,\r\n\r\nReturning to this thread to ask whether the possibility to create `DatasetDict` with different configurations will be supported in the future.\r\n\r\nBest,\r\nPietro", "DatasetDict is likely to always require the datasets to have the same columns and types, while different configurations may have different columns and types.\r\n\r\nWhy would you like to see that ?\r\nIf it's related to push_to_hub, we plan to allow pushing several configs, but not using DatasetDict", "Hi @lhoestq and @pietrolesci,\r\n\r\nI have been curious about this question as well. I don't have experience working with different configurations, but I can give a bit more detail on the work flow that I have been using with `Dataset_dict`.\r\n\r\nAs @pietrolesci mentions, I have been using `push_to_hub` to quickly iterate on dataset curation for different ML experiments - locally I create a set of dataset splits e.g. `train/val/test/inference`, then convert them to `HF_Datasets` and finally a to `Dataset_Dict` to `push_to_hub`. Where I have run into issues is when I want to include different metadata for different splits. For example, I have situations where I only have meta-data for one of the splits (e.g. test) or situations where I am working with `inference` data that does not have labels. Currently I use a rather hacky work around by adding \"dummy\" columns for missing columns to avoid the error:\r\n\r\n```\r\nValueError: All datasets in `DatasetDict` should have the same features\r\n```\r\n\r\nI am curious why `DatasetDict` will likely not support this functionality? I don't know much about working with different configurations, but allowing for different columns between datasets / splits would be a very helpful use-case for me. Are there any docs for using different configuration OR a more info about incorporating it with `push_to_hub`.\r\n\r\nBest wishes,\r\nJonathan\r\n\r\n", "+1", "> I am curious why DatasetDict will likely not support this functionality?\r\n\r\nThere's a possibility we may merge the Dataset and DatasetDict classes. The DatasetDict purpose was to define a way to get the train/test splits of a dataset.\r\n\r\nsee the discussions at https://github.com/huggingface/datasets/issues/5189\r\n\r\n> Are there any docs for using different configuration OR a more info about incorporating it with push_to_hub.\r\n\r\nThere's a PR open to allow to upload a dataset with a certain configuration name. Then later you can reload this specific configuration using `load_dataset(ds_name, config_name)`\r\n\r\nsee the PR at https://github.com/huggingface/datasets/pull/5213", "Hi, regarding the following information:\r\n\r\n> Please note that this is a design purpose: a `DatasetDict` has the same features for all its datasets. Normally, a `DatasetDict` is composed of several sub-datasets each corresponding to a different **split**.\r\n> \r\n> To handle sub-datasets with different features, we use another approach: use different **configurations** instead of **splits**.\r\n\r\nAltough this is often implied (such as how else would `DatasetDict` be able to process multiple splits in the same way?), I would expect it to be written somewhere in the docs plainly and maybe even in bold. Also I would expect to see it in multiple places such as:\r\n\r\n- in docstring of `DatasetDict`\r\n- in nlp/image/audio guides on how to create a dataset\r\n- [in conceptual guide on how to create a loading script](https://huggingface.co/docs/datasets/main/en/about_dataset_load)\r\n\r\n\r\nI think this addition would benefit the docs, especially when you guide a newbie (such as me) through the process of creating a dataset. As I said, you somehow suspect that this is in fact the case, but without reading it in the docs you cannot be sure." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4211/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4211
https://github.com/huggingface/datasets/issues/4211
false
1,214,089,130
https://api.github.com/repos/huggingface/datasets/issues/4210/labels{/name}
### System Info ```shell - `transformers` version: 4.18.0 - Platform: Linux-5.4.144+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - Huggingface_hub version: 0.5.1 - PyTorch version (GPU?): 1.10.0+cu111 (True) - Tensorflow version (GPU?): 2.8.0 (True) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using GPU in script?: <fill in> - Using distributed or parallel set-up in script?: <fill in> ``` ### Who can help? @LysandreJik ### Information - [ ] The official example scripts - [X] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [X] My own task or dataset (give details below) ### Reproduction ```python from datasets import load_dataset,Features,Value,ClassLabel class_names = ["cmn","deu","rus","fra","eng","jpn","spa","ita","kor","vie","nld","epo","por","tur","heb","hun","ell","ind","ara","arz","fin","bul","yue","swe","ukr","bel","que","ces","swh","nno","wuu","nob","zsm","est","kat","pol","lat","urd","sqi","isl","fry","afr","ron","fao","san","bre","tat","yid","uig","uzb","srp","qya","dan","pes","slk","eus","cycl","acm","tgl","lvs","kaz","hye","hin","lit","ben","cat","bos","hrv","tha","orv","cha","mon","lzh","scn","gle","mkd","slv","frm","glg","vol","ain","jbo","tok","ina","nds","mal","tlh","roh","ltz","oss","ido","gla","mlt","sco","ast","jav","oci","ile","ota","xal","tel","sjn","nov","khm","tpi","ang","aze","tgk","tuk","chv","hsb","dsb","bod","sme","cym","mri","ksh","kmr","ewe","kab","ber","tpw","udm","lld","pms","lad","grn","mlg","xho","pnb","grc","hat","lao","npi","cor","nah","avk","mar","guj","pan","kir","myv","prg","sux","crs","ckt","bak","zlm","hil","cbk","chr","nav","lkt","enm","arq","lin","abk","pcd","rom","gsw","tam","zul","awa","wln","amh","bar","hbo","mhr","bho","mrj","ckb","osx","pfl","mgm","sna","mah","hau","kan","nog","sin","glv","dng","kal","liv","vro","apc","jdt","fur","che","haw","yor","crh","pdc","ppl","kin","shs","mnw","tet","sah","kum","ngt","nya","pus","hif","mya","moh","wol","tir","ton","lzz","oar","lug","brx","non","mww","hak","nlv","ngu","bua","aym","vec","ibo","tkl","bam","kha","ceb","lou","fuc","smo","gag","lfn","arg","umb","tyv","kjh","oji","cyo","urh","kzj","pam","srd","lmo","swg","mdf","gil","snd","tso","sot","zza","tsn","pau","som","egl","ady","asm","ori","dtp","cho","max","kam","niu","sag","ilo","kaa","fuv","nch","hoc","iba","gbm","sun","war","mvv","pap","ary","kxi","csb","pag","cos","rif","kek","krc","aii","ban","ssw","tvl","mfe","tah","bvy","bcl","hnj","nau","nst","afb","quc","min","tmw","mad","bjn","mai","cjy","got","hsn","gan","tzl","dws","ldn","afh","sgs","krl","vep","rue","tly","mic","ext","izh","sma","jam","cmo","mwl","kpv","koi","bis","ike","run","evn","ryu","mnc","aoz","otk","kas","aln","akl","yua","shy","fkv","gos","fij","thv","zgh","gcf","cay","xmf","tig","div","lij","rap","hrx","cpi","tts","gaa","tmr","iii","ltg","bzt","syc","emx","gom","chg","osp","stq","frr","fro","nys","toi","new","phn","jpa","rel","drt","chn","pli","laa","bal","hdn","hax","mik","ajp","xqa","pal","crk","mni","lut","ayl","ood","sdh","ofs","nus","kiu","diq","qxq","alt","bfz","klj","mus","srn","guc","lim","zea","shi","mnr","bom","sat","szl"] features = Features({ 'label': ClassLabel(names=class_names), 'text': Value('string')}) num_labels = features['label'].num_classes data_files = { "train": "train.csv", "test": "test.csv" } sentences = load_dataset("loretoparisi/tatoeba-sentences", data_files=data_files, delimiter='\t', column_names=['label', 'text'], features = features ``` ERROR: ``` ClassLabel(num_classes=403, names=['cmn', 'deu', 'rus', 'fra', 'eng', 'jpn', 'spa', 'ita', 'kor', 'vie', 'nld', 'epo', 'por', 'tur', 'heb', 'hun', 'ell', 'ind', 'ara', 'arz', 'fin', 'bul', 'yue', 'swe', 'ukr', 'bel', 'que', 'ces', 'swh', 'nno', 'wuu', 'nob', 'zsm', 'est', 'kat', 'pol', 'lat', 'urd', 'sqi', 'isl', 'fry', 'afr', 'ron', 'fao', 'san', 'bre', 'tat', 'yid', 'uig', 'uzb', 'srp', 'qya', 'dan', 'pes', 'slk', 'eus', 'cycl', 'acm', 'tgl', 'lvs', 'kaz', 'hye', 'hin', 'lit', 'ben', 'cat', 'bos', 'hrv', 'tha', 'orv', 'cha', 'mon', 'lzh', 'scn', 'gle', 'mkd', 'slv', 'frm', 'glg', 'vol', 'ain', 'jbo', 'tok', 'ina', 'nds', 'mal', 'tlh', 'roh', 'ltz', 'oss', 'ido', 'gla', 'mlt', 'sco', 'ast', 'jav', 'oci', 'ile', 'ota', 'xal', 'tel', 'sjn', 'nov', 'khm', 'tpi', 'ang', 'aze', 'tgk', 'tuk', 'chv', 'hsb', 'dsb', 'bod', 'sme', 'cym', 'mri', 'ksh', 'kmr', 'ewe', 'kab', 'ber', 'tpw', 'udm', 'lld', 'pms', 'lad', 'grn', 'mlg', 'xho', 'pnb', 'grc', 'hat', 'lao', 'npi', 'cor', 'nah', 'avk', 'mar', 'guj', 'pan', 'kir', 'myv', 'prg', 'sux', 'crs', 'ckt', 'bak', 'zlm', 'hil', 'cbk', 'chr', 'nav', 'lkt', 'enm', 'arq', 'lin', 'abk', 'pcd', 'rom', 'gsw', 'tam', 'zul', 'awa', 'wln', 'amh', 'bar', 'hbo', 'mhr', 'bho', 'mrj', 'ckb', 'osx', 'pfl', 'mgm', 'sna', 'mah', 'hau', 'kan', 'nog', 'sin', 'glv', 'dng', 'kal', 'liv', 'vro', 'apc', 'jdt', 'fur', 'che', 'haw', 'yor', 'crh', 'pdc', 'ppl', 'kin', 'shs', 'mnw', 'tet', 'sah', 'kum', 'ngt', 'nya', 'pus', 'hif', 'mya', 'moh', 'wol', 'tir', 'ton', 'lzz', 'oar', 'lug', 'brx', 'non', 'mww', 'hak', 'nlv', 'ngu', 'bua', 'aym', 'vec', 'ibo', 'tkl', 'bam', 'kha', 'ceb', 'lou', 'fuc', 'smo', 'gag', 'lfn', 'arg', 'umb', 'tyv', 'kjh', 'oji', 'cyo', 'urh', 'kzj', 'pam', 'srd', 'lmo', 'swg', 'mdf', 'gil', 'snd', 'tso', 'sot', 'zza', 'tsn', 'pau', 'som', 'egl', 'ady', 'asm', 'ori', 'dtp', 'cho', 'max', 'kam', 'niu', 'sag', 'ilo', 'kaa', 'fuv', 'nch', 'hoc', 'iba', 'gbm', 'sun', 'war', 'mvv', 'pap', 'ary', 'kxi', 'csb', 'pag', 'cos', 'rif', 'kek', 'krc', 'aii', 'ban', 'ssw', 'tvl', 'mfe', 'tah', 'bvy', 'bcl', 'hnj', 'nau', 'nst', 'afb', 'quc', 'min', 'tmw', 'mad', 'bjn', 'mai', 'cjy', 'got', 'hsn', 'gan', 'tzl', 'dws', 'ldn', 'afh', 'sgs', 'krl', 'vep', 'rue', 'tly', 'mic', 'ext', 'izh', 'sma', 'jam', 'cmo', 'mwl', 'kpv', 'koi', 'bis', 'ike', 'run', 'evn', 'ryu', 'mnc', 'aoz', 'otk', 'kas', 'aln', 'akl', 'yua', 'shy', 'fkv', 'gos', 'fij', 'thv', 'zgh', 'gcf', 'cay', 'xmf', 'tig', 'div', 'lij', 'rap', 'hrx', 'cpi', 'tts', 'gaa', 'tmr', 'iii', 'ltg', 'bzt', 'syc', 'emx', 'gom', 'chg', 'osp', 'stq', 'frr', 'fro', 'nys', 'toi', 'new', 'phn', 'jpa', 'rel', 'drt', 'chn', 'pli', 'laa', 'bal', 'hdn', 'hax', 'mik', 'ajp', 'xqa', 'pal', 'crk', 'mni', 'lut', 'ayl', 'ood', 'sdh', 'ofs', 'nus', 'kiu', 'diq', 'qxq', 'alt', 'bfz', 'klj', 'mus', 'srn', 'guc', 'lim', 'zea', 'shi', 'mnr', 'bom', 'sat', 'szl'], id=None) Value(dtype='string', id=None) Using custom data configuration loretoparisi--tatoeba-sentences-7b2c5e991f398f39 Downloading and preparing dataset csv/loretoparisi--tatoeba-sentences to /root/.cache/huggingface/datasets/csv/loretoparisi--tatoeba-sentences-7b2c5e991f398f39/0.0.0/433e0ccc46f9880962cc2b12065189766fbb2bee57a221866138fb9203c83519... Downloading data files: 100% 2/2 [00:18<00:00, 8.06s/it] Downloading data: 100% 391M/391M [00:13<00:00, 35.3MB/s] Downloading data: 100% 92.4M/92.4M [00:02<00:00, 36.5MB/s] Failed to read file '/root/.cache/huggingface/datasets/downloads/933132df9905194ea9faeb30cabca8c49318795612f6495fcb941a290191dd5d' with error <class 'ValueError'>: invalid literal for int() with base 10: 'cmn' --------------------------------------------------------------------------- TypeError Traceback (most recent call last) /usr/local/lib/python3.7/dist-packages/pandas/_libs/parsers.pyx in pandas._libs.parsers.TextReader._convert_tokens() TypeError: Cannot cast array data from dtype('O') to dtype('int64') according to the rule 'safe' During handling of the above exception, another exception occurred: ValueError Traceback (most recent call last) 15 frames /usr/local/lib/python3.7/dist-packages/pandas/_libs/parsers.pyx in pandas._libs.parsers.TextReader._convert_tokens() ValueError: invalid literal for int() with base 10: 'cmn' ``` while loading without `features` it loads without errors ``` sentences = load_dataset("loretoparisi/tatoeba-sentences", data_files=data_files, delimiter='\t', column_names=['label', 'text'] ) ``` but the `label` col seems to be wrong (without the `ClassLabel` object): ``` sentences['train'].features {'label': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None)} ``` The dataset was https://huggingface.co/datasets/loretoparisi/tatoeba-sentences Dataset format is: ``` ces Nechci vědět, co je tam uvnitř. ces Kdo o tom chce slyšet? deu Tom sagte, er fühle sich nicht wohl. ber Mel-iyi-d anida-t tura ? hun Gondom lesz rá rögtön. ber Mel-iyi-d anida-tt tura ? deu Ich will dich nicht reden hören. ``` ### Expected behavior ```shell correctly load train and test files. ```
2022-05-31T12:16:31Z
4,210
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-25T07:28:42Z
https://api.github.com/repos/huggingface/datasets/issues/4210/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4210/timeline
TypeError: Cannot cast array data from dtype('O') to dtype('int64') according to the rule 'safe'
https://api.github.com/repos/huggingface/datasets/issues/4210/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/163333?v=4", "events_url": "https://api.github.com/users/loretoparisi/events{/privacy}", "followers_url": "https://api.github.com/users/loretoparisi/followers", "following_url": "https://api.github.com/users/loretoparisi/following{/other_user}", "gists_url": "https://api.github.com/users/loretoparisi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/loretoparisi", "id": 163333, "login": "loretoparisi", "node_id": "MDQ6VXNlcjE2MzMzMw==", "organizations_url": "https://api.github.com/users/loretoparisi/orgs", "received_events_url": "https://api.github.com/users/loretoparisi/received_events", "repos_url": "https://api.github.com/users/loretoparisi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/loretoparisi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/loretoparisi/subscriptions", "type": "User", "url": "https://api.github.com/users/loretoparisi" }
[]
null
completed
NONE
2022-05-31T12:16:31Z
null
I_kwDODunzps5IXYeq
[ "Hi! Casting class labels from strings is currently not supported in the CSV loader, but you can get the same result with an additional map as follows:\r\n```python\r\nfrom datasets import load_dataset,Features,Value,ClassLabel\r\nclass_names = [\"cmn\",\"deu\",\"rus\",\"fra\",\"eng\",\"jpn\",\"spa\",\"ita\",\"kor\",\"vie\",\"nld\",\"epo\",\"por\",\"tur\",\"heb\",\"hun\",\"ell\",\"ind\",\"ara\",\"arz\",\"fin\",\"bul\",\"yue\",\"swe\",\"ukr\",\"bel\",\"que\",\"ces\",\"swh\",\"nno\",\"wuu\",\"nob\",\"zsm\",\"est\",\"kat\",\"pol\",\"lat\",\"urd\",\"sqi\",\"isl\",\"fry\",\"afr\",\"ron\",\"fao\",\"san\",\"bre\",\"tat\",\"yid\",\"uig\",\"uzb\",\"srp\",\"qya\",\"dan\",\"pes\",\"slk\",\"eus\",\"cycl\",\"acm\",\"tgl\",\"lvs\",\"kaz\",\"hye\",\"hin\",\"lit\",\"ben\",\"cat\",\"bos\",\"hrv\",\"tha\",\"orv\",\"cha\",\"mon\",\"lzh\",\"scn\",\"gle\",\"mkd\",\"slv\",\"frm\",\"glg\",\"vol\",\"ain\",\"jbo\",\"tok\",\"ina\",\"nds\",\"mal\",\"tlh\",\"roh\",\"ltz\",\"oss\",\"ido\",\"gla\",\"mlt\",\"sco\",\"ast\",\"jav\",\"oci\",\"ile\",\"ota\",\"xal\",\"tel\",\"sjn\",\"nov\",\"khm\",\"tpi\",\"ang\",\"aze\",\"tgk\",\"tuk\",\"chv\",\"hsb\",\"dsb\",\"bod\",\"sme\",\"cym\",\"mri\",\"ksh\",\"kmr\",\"ewe\",\"kab\",\"ber\",\"tpw\",\"udm\",\"lld\",\"pms\",\"lad\",\"grn\",\"mlg\",\"xho\",\"pnb\",\"grc\",\"hat\",\"lao\",\"npi\",\"cor\",\"nah\",\"avk\",\"mar\",\"guj\",\"pan\",\"kir\",\"myv\",\"prg\",\"sux\",\"crs\",\"ckt\",\"bak\",\"zlm\",\"hil\",\"cbk\",\"chr\",\"nav\",\"lkt\",\"enm\",\"arq\",\"lin\",\"abk\",\"pcd\",\"rom\",\"gsw\",\"tam\",\"zul\",\"awa\",\"wln\",\"amh\",\"bar\",\"hbo\",\"mhr\",\"bho\",\"mrj\",\"ckb\",\"osx\",\"pfl\",\"mgm\",\"sna\",\"mah\",\"hau\",\"kan\",\"nog\",\"sin\",\"glv\",\"dng\",\"kal\",\"liv\",\"vro\",\"apc\",\"jdt\",\"fur\",\"che\",\"haw\",\"yor\",\"crh\",\"pdc\",\"ppl\",\"kin\",\"shs\",\"mnw\",\"tet\",\"sah\",\"kum\",\"ngt\",\"nya\",\"pus\",\"hif\",\"mya\",\"moh\",\"wol\",\"tir\",\"ton\",\"lzz\",\"oar\",\"lug\",\"brx\",\"non\",\"mww\",\"hak\",\"nlv\",\"ngu\",\"bua\",\"aym\",\"vec\",\"ibo\",\"tkl\",\"bam\",\"kha\",\"ceb\",\"lou\",\"fuc\",\"smo\",\"gag\",\"lfn\",\"arg\",\"umb\",\"tyv\",\"kjh\",\"oji\",\"cyo\",\"urh\",\"kzj\",\"pam\",\"srd\",\"lmo\",\"swg\",\"mdf\",\"gil\",\"snd\",\"tso\",\"sot\",\"zza\",\"tsn\",\"pau\",\"som\",\"egl\",\"ady\",\"asm\",\"ori\",\"dtp\",\"cho\",\"max\",\"kam\",\"niu\",\"sag\",\"ilo\",\"kaa\",\"fuv\",\"nch\",\"hoc\",\"iba\",\"gbm\",\"sun\",\"war\",\"mvv\",\"pap\",\"ary\",\"kxi\",\"csb\",\"pag\",\"cos\",\"rif\",\"kek\",\"krc\",\"aii\",\"ban\",\"ssw\",\"tvl\",\"mfe\",\"tah\",\"bvy\",\"bcl\",\"hnj\",\"nau\",\"nst\",\"afb\",\"quc\",\"min\",\"tmw\",\"mad\",\"bjn\",\"mai\",\"cjy\",\"got\",\"hsn\",\"gan\",\"tzl\",\"dws\",\"ldn\",\"afh\",\"sgs\",\"krl\",\"vep\",\"rue\",\"tly\",\"mic\",\"ext\",\"izh\",\"sma\",\"jam\",\"cmo\",\"mwl\",\"kpv\",\"koi\",\"bis\",\"ike\",\"run\",\"evn\",\"ryu\",\"mnc\",\"aoz\",\"otk\",\"kas\",\"aln\",\"akl\",\"yua\",\"shy\",\"fkv\",\"gos\",\"fij\",\"thv\",\"zgh\",\"gcf\",\"cay\",\"xmf\",\"tig\",\"div\",\"lij\",\"rap\",\"hrx\",\"cpi\",\"tts\",\"gaa\",\"tmr\",\"iii\",\"ltg\",\"bzt\",\"syc\",\"emx\",\"gom\",\"chg\",\"osp\",\"stq\",\"frr\",\"fro\",\"nys\",\"toi\",\"new\",\"phn\",\"jpa\",\"rel\",\"drt\",\"chn\",\"pli\",\"laa\",\"bal\",\"hdn\",\"hax\",\"mik\",\"ajp\",\"xqa\",\"pal\",\"crk\",\"mni\",\"lut\",\"ayl\",\"ood\",\"sdh\",\"ofs\",\"nus\",\"kiu\",\"diq\",\"qxq\",\"alt\",\"bfz\",\"klj\",\"mus\",\"srn\",\"guc\",\"lim\",\"zea\",\"shi\",\"mnr\",\"bom\",\"sat\",\"szl\"]\r\nfeatures = Features({ 'label': ClassLabel(names=class_names), 'text': Value('string')})\r\nnum_labels = features['label'].num_classes\r\ndata_files = { \"train\": \"train.csv\", \"test\": \"test.csv\" }\r\nsentences = load_dataset(\r\n \"loretoparisi/tatoeba-sentences\",\r\n data_files=data_files,\r\n delimiter='\\t', \r\n column_names=['label', 'text'],\r\n)\r\n# You can make this part faster with num_proc=<some int>\r\nsentences = sentences.map(lambda ex: features[\"label\"].str2int(ex[\"label\"]) if ex[\"label\"] is not None else None, features=features)\r\n```\r\n\r\n@lhoestq IIRC, I suggested adding `cast_to_storage` to `ClassLabel` + `table_cast` to the packaged loaders if the `ClassLabel`/`Image`/`Audio` type is present in `features` to avoid this kind of error, but your concern was speed. IMO shouldn't be a problem if we do `table_cast` only when these features are present.", "I agree packaged loaders should support `ClassLabel` feature without throwing an error.", "@albertvillanova @mariosasko thank you, with that change now I get\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError Traceback (most recent call last)\r\n[<ipython-input-9-eeb68eeb9bec>](https://localhost:8080/#) in <module>()\r\n 11 )\r\n 12 # You can make this part faster with num_proc=<some int>\r\n---> 13 sentences = sentences.map(lambda ex: features[\"label\"].str2int(ex[\"label\"]) if ex[\"label\"] is not None else None, features=features)\r\n 14 sentences = sentences.shuffle()\r\n\r\n8 frames\r\n[/usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in validate_function_output(processed_inputs, indices)\r\n 2193 if processed_inputs is not None and not isinstance(processed_inputs, (Mapping, pa.Table)):\r\n 2194 raise TypeError(\r\n-> 2195 f\"Provided `function` which is applied to all elements of table returns a variable of type {type(processed_inputs)}. Make sure provided `function` returns a variable of type `dict` (or a pyarrow table) to update the dataset or `None` if you are only interested in side effects.\"\r\n 2196 )\r\n 2197 elif isinstance(indices, list) and isinstance(processed_inputs, Mapping):\r\n\r\nTypeError: Provided `function` which is applied to all elements of table returns a variable of type <class 'int'>. Make sure provided `function` returns a variable of type `dict` (or a pyarrow table) to update the dataset or `None` if you are only interested in side effects.\r\n```\r\n\r\nthe error is raised by [this](https://github.com/huggingface/datasets/blob/master/src/datasets/arrow_dataset.py#L2221)\r\n\r\n```\r\n[/usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in validate_function_output(processed_inputs, indices)\r\n```", "@mariosasko changed it like\r\n\r\n```python\r\nsentences = sentences.map(lambda ex: {\"label\" : features[\"label\"].str2int(ex[\"label\"]) if ex[\"label\"] is not None else None}, features=features)\r\n```\r\n\r\nto avoid the above errorr.", "Any update on this? Is this correct ?\r\n> @mariosasko changed it like\r\n> \r\n> ```python\r\n> sentences = sentences.map(lambda ex: {\"label\" : features[\"label\"].str2int(ex[\"label\"]) if ex[\"label\"] is not None else None}, features=features)\r\n> ```\r\n> \r\n> to avoid the above errorr.\r\n\r\n" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4210/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4210
https://github.com/huggingface/datasets/issues/4210
false
1,213,716,426
https://api.github.com/repos/huggingface/datasets/issues/4208/labels{/name}
Resolves #3457 Dataset Request : Add CMU Graphics Lab Motion Capture dataset [#3457](https://github.com/huggingface/datasets/issues/3457) This PR adds the CMU MoCap Dataset. The authors didn't respond even after multiple follow ups, so I ended up crawling the website to get categories, subcategories and description information. Some of the subjects do not have category/subcategory/description as well. I am using a subject to categories, subcategories and description map (metadata file). Currently the loading of the dataset works for "asf/amc" and "avi" formats since they have a single download link. But "c3d" and "mpg" have multiple download links (part archives) and dl_manager.download_and_extract() extracts the files to multiple paths, is there a way to extract these multiple archives into one folder ? Any other way to go about this ? Any suggestions/inputs on this would be helpful. Thank you.
2022-10-03T09:38:24Z
4,208
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-04-24T17:31:08Z
https://api.github.com/repos/huggingface/datasets/issues/4208/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4208/timeline
Add CMU MoCap Dataset
https://api.github.com/repos/huggingface/datasets/issues/4208/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/17746528?v=4", "events_url": "https://api.github.com/users/dnaveenr/events{/privacy}", "followers_url": "https://api.github.com/users/dnaveenr/followers", "following_url": "https://api.github.com/users/dnaveenr/following{/other_user}", "gists_url": "https://api.github.com/users/dnaveenr/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dnaveenr", "id": 17746528, "login": "dnaveenr", "node_id": "MDQ6VXNlcjE3NzQ2NTI4", "organizations_url": "https://api.github.com/users/dnaveenr/orgs", "received_events_url": "https://api.github.com/users/dnaveenr/received_events", "repos_url": "https://api.github.com/users/dnaveenr/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dnaveenr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dnaveenr/subscriptions", "type": "User", "url": "https://api.github.com/users/dnaveenr" }
[]
null
null
CONTRIBUTOR
2022-10-03T09:36:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4208.diff", "html_url": "https://github.com/huggingface/datasets/pull/4208", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4208.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4208" }
PR_kwDODunzps42r7bW
[ "_The documentation is not available anymore as the PR was closed or merged._", "- Updated the readme.\r\n- Added dummy_data.zip and ran the all the tests.\r\n\r\nThe dataset works for \"asf/amc\" and \"avi\" formats which have a single download link for the complete dataset. But \"c3d\" and \"mpg\" have multiple download links, can we combine and host these links on the Hub since the dataset is free to use ?", "\"c3d\" and \"mpg\" have multiple download links (part archives) and dl_manager.download_and_extract() extracts the files to multiple paths, is there a way to extract these multiple archives into one folder ? Any other way to go about this ?\r\nCan we combine and host these links on the Hub since the dataset is free to use ?", "> \"c3d\" and \"mpg\" have multiple download links (part archives) and dl_manager.download_and_extract() extracts the files to multiple paths, is there a way to extract these multiple archives into one folder ? Any other way to go about this ?\r\n\r\nWe store downloaded data under `~/.cache/huggingface/datasets/downloads` (by default), so these downloads are \"hidden\" and won't clutter one's filesystem in an \"obvious way\".", "> We store downloaded data under ~/.cache/huggingface/datasets/downloads (by default), so these downloads are \"hidden\" and won't clutter one's filesystem in an \"obvious way\".\r\n\r\nYes, the filesystem won't be clustered, but the problem is processing the dataset becomes cumbersome. For eg, for the c3d format has 5 part-downloads, so the folders will be as follows : \r\n```\r\n['~/.cache/huggingface/datasets/downloads/extracted/0e6bf028f490bf18c23ce572d1437c4ef32a74f630e33c26a806250d35cfcdd1', '~/.cache/huggingface/datasets/downloads/extracted/1b44fc5c7a6e031c904545422d449fd964f8ee795b9d1dcb0b6a76d03b50ebe6', '~/.cache/huggingface/datasets/downloads/extracted/137595188e96187c24ce1aa5c78200c7f78816fbd9d6c62354c01b3e6ec550c7', '~/.cache/huggingface/datasets/downloads/extracted/6c0c893e435f36fd79aa0f199f58fe16f01985f039644a7cb094a8c43a15ffd4', '~/.cache/huggingface/datasets/downloads/extracted/45e4703354cbc975e6add66f1b17b716c882b56f44575b033c5926aa5fcfb17f']\r\n```\r\nEach of these folders have a given set of subjects, so we'll be need to write extra code to fetch data from each of these folders, and the mpg format has 12 part-downloads which will lead to 12 folders having certain set of subjects, so it is cumbersome to process them.", "I have added all the changes that were suggested. We just need to handle the multi-part download for c3d and mpg formats. Easiest way would be to have just one zip for these formats.", "But we can handle this with a simple mapping that stores the id ranges (for each config), no? And an actual file path is not important during processing.", "I have added code to handle c3d, mpg formats as well. The data for the mpg format seems incomplete as it contains only 53 rows. I have added a note regarding this in the Data Splits section.", "The real data test works fine and dummy_data test work fine. There were few missing files which was causing issues, I have fixed it now.\r\n", "- Reduced the dummy_data size.\r\n- Added sample dataset preprocessing code, it is not complete though.\r\n- Added all changes suggested.\r\n\r\nLet me know if anything else is required. Thank you. :)", "Thanks for your contribution, @dnaveenr.\r\n\r\nWe are removing the dataset scripts from this GitHub repo and moving them to the Hugging Face Hub: https://huggingface.co/datasets\r\n\r\nWe would suggest you create this dataset there. Please, feel free to tell us if you need some help." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4208/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4208
https://github.com/huggingface/datasets/pull/4208
true
1,213,604,615
https://api.github.com/repos/huggingface/datasets/issues/4207/labels{/name}
Typo: `datasets.DatsetDict` -> `datasets.DatasetDict`
2022-05-05T13:17:47Z
4,207
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-24T09:49:37Z
https://api.github.com/repos/huggingface/datasets/issues/4207/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4207/timeline
[Minor edit] Fix typo in class name
https://api.github.com/repos/huggingface/datasets/issues/4207/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3664563?v=4", "events_url": "https://api.github.com/users/cakiki/events{/privacy}", "followers_url": "https://api.github.com/users/cakiki/followers", "following_url": "https://api.github.com/users/cakiki/following{/other_user}", "gists_url": "https://api.github.com/users/cakiki/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cakiki", "id": 3664563, "login": "cakiki", "node_id": "MDQ6VXNlcjM2NjQ1NjM=", "organizations_url": "https://api.github.com/users/cakiki/orgs", "received_events_url": "https://api.github.com/users/cakiki/received_events", "repos_url": "https://api.github.com/users/cakiki/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cakiki/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cakiki/subscriptions", "type": "User", "url": "https://api.github.com/users/cakiki" }
[]
null
null
CONTRIBUTOR
2022-05-05T13:17:47Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4207.diff", "html_url": "https://github.com/huggingface/datasets/pull/4207", "merged_at": "2022-05-05T13:17:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/4207.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4207" }
PR_kwDODunzps42rmbK
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4207/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4207
https://github.com/huggingface/datasets/pull/4207
true
1,212,715,581
https://api.github.com/repos/huggingface/datasets/issues/4206/labels{/name}
This PR adds readme.md and ner_val.py to metrics. Nerval is a python package that helps evaluate NER models. It creates classification report and confusion matrix at entity level.
2023-07-11T09:34:56Z
4,206
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "E3165C", "default": false, "description": "", "id": 4190228726, "name": "transfer-to-evaluate", "node_id": "LA_kwDODunzps75wdD2", "url": "https://api.github.com/repos/huggingface/datasets/labels/transfer-to-evaluate" } ]
2022-04-22T19:45:00Z
https://api.github.com/repos/huggingface/datasets/issues/4206/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4206/timeline
Add Nerval Metric
https://api.github.com/repos/huggingface/datasets/issues/4206/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/49372461?v=4", "events_url": "https://api.github.com/users/mdadda/events{/privacy}", "followers_url": "https://api.github.com/users/mdadda/followers", "following_url": "https://api.github.com/users/mdadda/following{/other_user}", "gists_url": "https://api.github.com/users/mdadda/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mdadda", "id": 49372461, "login": "mdadda", "node_id": "MDQ6VXNlcjQ5MzcyNDYx", "organizations_url": "https://api.github.com/users/mdadda/orgs", "received_events_url": "https://api.github.com/users/mdadda/received_events", "repos_url": "https://api.github.com/users/mdadda/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mdadda/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mdadda/subscriptions", "type": "User", "url": "https://api.github.com/users/mdadda" }
[]
null
null
NONE
2023-07-11T09:34:55Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4206.diff", "html_url": "https://github.com/huggingface/datasets/pull/4206", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4206.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4206" }
PR_kwDODunzps42pJQW
[ "Metrics are deprecated in `datasets` and `evaluate` should be used instead: https://github.com/huggingface/evaluate" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4206/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4206
https://github.com/huggingface/datasets/pull/4206
true
1,212,466,138
https://api.github.com/repos/huggingface/datasets/issues/4205/labels{/name}
Minor change to fully align this function with the recent change in Transformers (https://github.com/huggingface/transformers/pull/16891)
2022-05-03T15:28:42Z
4,205
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-22T14:56:21Z
https://api.github.com/repos/huggingface/datasets/issues/4205/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4205/timeline
Fix `convert_file_size_to_int` for kilobits and megabits
https://api.github.com/repos/huggingface/datasets/issues/4205/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-05-03T15:21:48Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4205.diff", "html_url": "https://github.com/huggingface/datasets/pull/4205", "merged_at": "2022-05-03T15:21:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/4205.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4205" }
PR_kwDODunzps42oVFE
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4205/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4205
https://github.com/huggingface/datasets/pull/4205
true
1,212,431,764
https://api.github.com/repos/huggingface/datasets/issues/4204/labels{/name}
What this PR mainly does: - add metric card for recall metric - update docs in recall python file Note: I've also included a .json file with all of the metric card information. I've started compiling the relevant information in this type of .json files, and then using a script I wrote to generate the formatted metric card, as well as the docs to go in the .py file. I figured I'd upload the .json because it could be useful, especially if I also make a PR with the script I'm using (let me know if that's something you think would be beneficial!)
2022-05-03T13:23:23Z
4,204
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-22T14:24:26Z
https://api.github.com/repos/huggingface/datasets/issues/4204/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4204/timeline
Add Recall Metric Card
https://api.github.com/repos/huggingface/datasets/issues/4204/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
null
null
CONTRIBUTOR
2022-05-03T13:16:24Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4204.diff", "html_url": "https://github.com/huggingface/datasets/pull/4204", "merged_at": "2022-05-03T13:16:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/4204.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4204" }
PR_kwDODunzps42oN0j
[ "_The documentation is not available anymore as the PR was closed or merged._", "This looks good to me! " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4204/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4204
https://github.com/huggingface/datasets/pull/4204
true
1,212,431,067
https://api.github.com/repos/huggingface/datasets/issues/4203/labels{/name}
What this PR mainly does: - add metric card for precision metric - update docs in precision python file Note: I've also included a .json file with all of the metric card information. I've started compiling the relevant information in this type of .json files, and then using a script I wrote to generate the formatted metric card, as well as the docs to go in the .py file. I figured I'd upload the .json because it could be useful, especially if I also make a PR with the script I'm using (let me know if that's something you think would be beneficial!)
2022-05-03T14:23:40Z
4,203
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-22T14:23:48Z
https://api.github.com/repos/huggingface/datasets/issues/4203/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4203/timeline
Add Precision Metric Card
https://api.github.com/repos/huggingface/datasets/issues/4203/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
null
null
CONTRIBUTOR
2022-05-03T14:16:46Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4203.diff", "html_url": "https://github.com/huggingface/datasets/pull/4203", "merged_at": "2022-05-03T14:16:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/4203.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4203" }
PR_kwDODunzps42oNrS
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4203/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4203
https://github.com/huggingface/datasets/pull/4203
true
1,212,326,288
https://api.github.com/repos/huggingface/datasets/issues/4202/labels{/name}
null
2022-04-22T15:03:00Z
4,202
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-22T12:53:31Z
https://api.github.com/repos/huggingface/datasets/issues/4202/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4202/timeline
Fix some type annotation in doc
https://api.github.com/repos/huggingface/datasets/issues/4202/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
[]
null
null
CONTRIBUTOR
2022-04-22T14:56:43Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4202.diff", "html_url": "https://github.com/huggingface/datasets/pull/4202", "merged_at": "2022-04-22T14:56:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/4202.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4202" }
PR_kwDODunzps42n278
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4202/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4202
https://github.com/huggingface/datasets/pull/4202
true
1,212,086,420
https://api.github.com/repos/huggingface/datasets/issues/4201/labels{/name}
Update template to use new issue forms instead. With this PR we can check if this new feature is useful for us. Once validated, we can update the other templates. CC: @severo
2022-05-06T08:38:43Z
4,201
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-22T09:34:44Z
https://api.github.com/repos/huggingface/datasets/issues/4201/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4201/timeline
Update GH template for dataset viewer issues
https://api.github.com/repos/huggingface/datasets/issues/4201/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-04-26T08:45:55Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4201.diff", "html_url": "https://github.com/huggingface/datasets/pull/4201", "merged_at": "2022-04-26T08:45:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/4201.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4201" }
PR_kwDODunzps42nIRm
[ "_The documentation is not available anymore as the PR was closed or merged._", "You can see rendering at: https://github.com/huggingface/datasets/blob/6b48fedbdafe12a42c7b6edcecc32820af1a4822/.github/ISSUE_TEMPLATE/dataset-viewer.yml" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4201/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4201
https://github.com/huggingface/datasets/pull/4201
true
1,211,980,110
https://api.github.com/repos/huggingface/datasets/issues/4200/labels{/name}
This option was missing from the docs guide (it was only explained in the docstring of `load_dataset`). Although this is an infrequent use case, there might be some users interested in it. Related to #4192 CC: @stevhliu
2022-05-06T08:39:25Z
4,200
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-22T08:08:25Z
https://api.github.com/repos/huggingface/datasets/issues/4200/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4200/timeline
Add to docs how to load from local script
https://api.github.com/repos/huggingface/datasets/issues/4200/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-04-23T05:47:25Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4200.diff", "html_url": "https://github.com/huggingface/datasets/pull/4200", "merged_at": "2022-04-23T05:47:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/4200.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4200" }
PR_kwDODunzps42mz0w
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4200/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4200
https://github.com/huggingface/datasets/pull/4200
true
1,211,953,308
https://api.github.com/repos/huggingface/datasets/issues/4199/labels{/name}
## Describe the bug It looks like that result of `.map` operation dataset are missing the cache when you reload the script and always run from scratch. In same interpretor session, they are able to find the cache and reload it. But, when you exit the interpretor and reload it, the downloading starts from scratch. ## Steps to reproduce the bug Using the example provided in `red_caps` dataset. ```python from concurrent.futures import ThreadPoolExecutor from functools import partial import io import urllib import PIL.Image import datasets from datasets import load_dataset from datasets.utils.file_utils import get_datasets_user_agent def fetch_single_image(image_url, timeout=None, retries=0): for _ in range(retries + 1): try: request = urllib.request.Request( image_url, data=None, headers={"user-agent": get_datasets_user_agent()}, ) with urllib.request.urlopen(request, timeout=timeout) as req: image = PIL.Image.open(io.BytesIO(req.read())) break except Exception: image = None return image def fetch_images(batch, num_threads, timeout=None, retries=0): fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries) with ThreadPoolExecutor(max_workers=num_threads) as executor: batch["image"] = list(executor.map(lambda image_urls: [fetch_single_image_with_args(image_url) for image_url in image_urls], batch["image_url"])) return batch def process_image_urls(batch): processed_batch_image_urls = [] for image_url in batch["image_url"]: processed_example_image_urls = [] image_url_splits = re.findall(r"http\S+", image_url) for image_url_split in image_url_splits: if "imgur" in image_url_split and "," in image_url_split: for image_url_part in image_url_split.split(","): if not image_url_part: continue image_url_part = image_url_part.strip() root, ext = os.path.splitext(image_url_part) if not root.startswith("http"): root = "http://i.imgur.com/" + root root = root.split("#")[0] if not ext: ext = ".jpg" ext = re.split(r"[?%]", ext)[0] image_url_part = root + ext processed_example_image_urls.append(image_url_part) else: processed_example_image_urls.append(image_url_split) processed_batch_image_urls.append(processed_example_image_urls) batch["image_url"] = processed_batch_image_urls return batch dset = load_dataset("red_caps", "jellyfish") dset = dset.map(process_image_urls, batched=True, num_proc=4) features = dset["train"].features.copy() features["image"] = datasets.Sequence(datasets.Image()) num_threads = 5 dset = dset.map(fetch_images, batched=True, batch_size=50, features=features, fn_kwargs={"num_threads": num_threads}) ``` Run this in an interpretor or as a script twice and see that the cache is missed the second time. ## Expected results At reload there should not be any cache miss ## Actual results Every time script is run, cache is missed and dataset is built from scratch. ## Environment info - `datasets` version: 2.1.1.dev0 - Platform: Linux-4.19.0-20-cloud-amd64-x86_64-with-glibc2.10 - Python version: 3.8.13 - PyArrow version: 7.0.0 - Pandas version: 1.4.1
2022-04-26T17:00:32Z
4,199
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-22T07:47:08Z
https://api.github.com/repos/huggingface/datasets/issues/4199/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
https://api.github.com/repos/huggingface/datasets/issues/4199/timeline
Cache miss during reload for datasets using image fetch utilities through map
https://api.github.com/repos/huggingface/datasets/issues/4199/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3616806?v=4", "events_url": "https://api.github.com/users/apsdehal/events{/privacy}", "followers_url": "https://api.github.com/users/apsdehal/followers", "following_url": "https://api.github.com/users/apsdehal/following{/other_user}", "gists_url": "https://api.github.com/users/apsdehal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/apsdehal", "id": 3616806, "login": "apsdehal", "node_id": "MDQ6VXNlcjM2MTY4MDY=", "organizations_url": "https://api.github.com/users/apsdehal/orgs", "received_events_url": "https://api.github.com/users/apsdehal/received_events", "repos_url": "https://api.github.com/users/apsdehal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/apsdehal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/apsdehal/subscriptions", "type": "User", "url": "https://api.github.com/users/apsdehal" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" } ]
null
completed
CONTRIBUTOR
2022-04-26T13:38:26Z
null
I_kwDODunzps5IPPCc
[ "Hi ! Maybe one of the objects in the function is not deterministic across sessions ? You can read more about it and how to investigate here: https://huggingface.co/docs/datasets/about_cache", "Hi @apsdehal! Can you verify that replacing\r\n```python\r\ndef fetch_single_image(image_url, timeout=None, retries=0):\r\n for _ in range(retries + 1):\r\n try:\r\n request = urllib.request.Request(\r\n image_url,\r\n data=None,\r\n headers={\"user-agent\": get_datasets_user_agent()},\r\n )\r\n with urllib.request.urlopen(request, timeout=timeout) as req:\r\n image = PIL.Image.open(io.BytesIO(req.read()))\r\n break\r\n except Exception:\r\n image = None\r\n return image\r\n```\r\nwith \r\n```python\r\nUSER_AGENT = get_datasets_user_agent()\r\n\r\ndef fetch_single_image(image_url, timeout=None, retries=0):\r\n for _ in range(retries + 1):\r\n try:\r\n request = urllib.request.Request(\r\n image_url,\r\n data=None,\r\n headers={\"user-agent\": USER_AGENT},\r\n )\r\n with urllib.request.urlopen(request, timeout=timeout) as req:\r\n image = PIL.Image.open(io.BytesIO(req.read()))\r\n break\r\n except Exception:\r\n image = None\r\n return image\r\n```\r\nfixes the issue?", "Thanks @mariosasko. That does fix the issue. In general, I think these image downloading utilities since they are being used by a lot of image dataset should be provided as a part of `datasets` library right to keep the logic consistent and READMEs smaller? If they already exists, that is also great, please point me to those. I saw that `http_get` does exist.", "You can find my rationale (and a proposed solution) for why these utilities are not a part of `datasets` here: https://github.com/huggingface/datasets/pull/4100#issuecomment-1097994003.", "Makes sense. But, I think as the number of image datasets as grow, more people are copying pasting original code from docs to work as it is while we make fixes to them later. I think we do need a central place for these to avoid that confusion as well as more easier access to image datasets. Should we restart that discussion, possible on slack?" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4199/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4199
https://github.com/huggingface/datasets/issues/4199
false
1,211,456,559
https://api.github.com/repos/huggingface/datasets/issues/4198/labels{/name}
## Dataset viewer issue for '*name of the dataset*' **Link:** *link to the dataset viewer page* *short description of the issue* Am I the one who added this dataset ? Yes-No
2022-05-03T11:29:05Z
4,198
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-04-21T19:19:26Z
https://api.github.com/repos/huggingface/datasets/issues/4198/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4198/timeline
There is no dataset
https://api.github.com/repos/huggingface/datasets/issues/4198/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1625647?v=4", "events_url": "https://api.github.com/users/wilfoderek/events{/privacy}", "followers_url": "https://api.github.com/users/wilfoderek/followers", "following_url": "https://api.github.com/users/wilfoderek/following{/other_user}", "gists_url": "https://api.github.com/users/wilfoderek/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wilfoderek", "id": 1625647, "login": "wilfoderek", "node_id": "MDQ6VXNlcjE2MjU2NDc=", "organizations_url": "https://api.github.com/users/wilfoderek/orgs", "received_events_url": "https://api.github.com/users/wilfoderek/received_events", "repos_url": "https://api.github.com/users/wilfoderek/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wilfoderek/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wilfoderek/subscriptions", "type": "User", "url": "https://api.github.com/users/wilfoderek" }
[]
null
completed
NONE
2022-04-22T06:12:25Z
null
I_kwDODunzps5INVwv
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4198/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4198
https://github.com/huggingface/datasets/issues/4198
false
1,211,342,558
https://api.github.com/repos/huggingface/datasets/issues/4197/labels{/name}
This should fix all the issue we have with in place operations in mapping functions. This is crucial as where we do some weird things like: ``` def apply(batch): batch_size = len(batch["id"]) batch["text"] = ["potato" for _ range(batch_size)] return {} # Columns are: {"id": int} dset.map(apply, batched=True, remove_columns="text") # crashes because `text` is not in the original columns dset.map(apply, batched=True) # mapped datasets has `text` column ``` In this PR we suggest to have `remove_columns=True` so that we ignore the input completely, and just use the output to generate mapped dataset. This means that inplace operations won't have any effects anymore.
2023-09-24T10:02:32Z
4,197
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-21T17:28:13Z
https://api.github.com/repos/huggingface/datasets/issues/4197/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4197/timeline
Add remove_columns=True
https://api.github.com/repos/huggingface/datasets/issues/4197/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
[]
null
null
CONTRIBUTOR
2022-04-22T14:45:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4197.diff", "html_url": "https://github.com/huggingface/datasets/pull/4197", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4197.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4197" }
PR_kwDODunzps42kyXD
[ "_The documentation is not available anymore as the PR was closed or merged._", "Any reason why we can't just do `[inputs.copy()]` in this line for in-place operations to not have effects anymore:\r\nhttps://github.com/huggingface/datasets/blob/bf432011ff9155a5bc16c03956bc63e514baf80d/src/datasets/arrow_dataset.py#L2232.\r\n\r\n(in the `batched` case, we can also copy the inputs' values (list objects) to ignore in-place modifications to the inputs' columns)\r\n\r\nI think `remove_columns=True` has no meaning, so I'm not a fan of this change.", "@mariosasko copy does have a cost associated with it ... and plus you'll have to consider `deepcopy` Imagine columnds that are list of list of list of list .... Though I have to agree that `remove_columns=True` doesn't make sense (but, IMO, neither does it in its current use-case as it should refer to `input_columns`) ", "Okay closing this PR for the following reasons:\r\n - `remove_columns=True` was expected to keep the `.update`-like operator for `.map`. I initially thought it would be a good way to ignore function side effects and only keep output of that function (cf. PR description).\r\n - expected `remove_columns=True` is a bad API according to @mariosasko and introduces unecessary changes for little gain (strictly equivalent to `remove_columns=dset.column_names`)" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4197/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4197
https://github.com/huggingface/datasets/pull/4197
true
1,211,271,261
https://api.github.com/repos/huggingface/datasets/issues/4196/labels{/name}
Following https://github.com/huggingface/datasets/pull/4184, currently a dataset saved using `save_to_disk` doesn't actually contain the bytes of the image or audio files. Instead it stores the path to your local files. Adding `embed_external_files` and set it to True by default to save_to_disk would be kind of a breaking change since some users will get bigger Arrow files when updating the lib, but the advantages are nice: - the resulting dataset is self contained, in case you want to delete your cache for example or share it with someone else - users also upload these Arrow files to cloud storage via the fs parameter, and in this case they would expect to upload a self-contained dataset - consistency with push_to_hub This can be implemented at the same time as sharding for `save_to_disk` for efficiency, and reuse the helpers from `push_to_hub` to embed the external files. cc @mariosasko
2022-12-14T18:22:59Z
4,196
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-04-21T16:25:18Z
https://api.github.com/repos/huggingface/datasets/issues/4196/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4196/timeline
Embed image and audio files in `save_to_disk`
https://api.github.com/repos/huggingface/datasets/issues/4196/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
completed
MEMBER
2022-12-14T18:22:59Z
null
I_kwDODunzps5IMohd
[]
{ "+1": 6, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 6, "url": "https://api.github.com/repos/huggingface/datasets/issues/4196/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4196
https://github.com/huggingface/datasets/issues/4196
false
1,210,958,602
https://api.github.com/repos/huggingface/datasets/issues/4194/labels{/name}
Fix #4191. CC: @SaulLu
2022-05-12T15:16:34Z
4,194
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-21T12:22:26Z
https://api.github.com/repos/huggingface/datasets/issues/4194/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4194/timeline
Support lists of multi-dimensional numpy arrays
https://api.github.com/repos/huggingface/datasets/issues/4194/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-05-12T15:08:40Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4194.diff", "html_url": "https://github.com/huggingface/datasets/pull/4194", "merged_at": "2022-05-12T15:08:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/4194.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4194" }
PR_kwDODunzps42jjD3
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4194/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4194
https://github.com/huggingface/datasets/pull/4194
true
1,210,734,701
https://api.github.com/repos/huggingface/datasets/issues/4193/labels{/name}
Following https://github.com/huggingface/datasets/pull/4187, I explained in the documentation of `save_to_disk` and `push_to_hub` how they handle image and audio data.
2022-04-22T09:55:55Z
4,193
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-21T09:04:36Z
https://api.github.com/repos/huggingface/datasets/issues/4193/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4193/timeline
Document save_to_disk and push_to_hub on images and audio files
https://api.github.com/repos/huggingface/datasets/issues/4193/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-04-22T09:49:31Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4193.diff", "html_url": "https://github.com/huggingface/datasets/pull/4193", "merged_at": "2022-04-22T09:49:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/4193.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4193" }
PR_kwDODunzps42izQG
[ "_The documentation is not available anymore as the PR was closed or merged._", "Good catch, I updated the docstrings" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4193/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4193
https://github.com/huggingface/datasets/pull/4193
true
1,210,692,554
https://api.github.com/repos/huggingface/datasets/issues/4192/labels{/name}
Traceback (most recent call last): File "/home/gs603/ahf/pretrained/model.py", line 48, in <module> dataset = load_dataset("json",data_files="dataset/dataset_infos.json") File "/home/gs603/miniconda3/envs/coderepair/lib/python3.7/site-packages/datasets/load.py", line 1675, in load_dataset **config_kwargs, File "/home/gs603/miniconda3/envs/coderepair/lib/python3.7/site-packages/datasets/load.py", line 1496, in load_dataset_builder data_files=data_files, File "/home/gs603/miniconda3/envs/coderepair/lib/python3.7/site-packages/datasets/load.py", line 1155, in dataset_module_factory download_mode=download_mode, File "/home/gs603/miniconda3/envs/coderepair/lib/python3.7/site-packages/datasets/load.py", line 800, in get_module data_files = DataFilesDict.from_local_or_remote(patterns, use_auth_token=self.downnload_config.use_auth_token) File "/home/gs603/miniconda3/envs/coderepair/lib/python3.7/site-packages/datasets/data_files.py", line 582, in from_local_or_remote if not isinstance(patterns_for_key, DataFilesList) File "/home/gs603/miniconda3/envs/coderepair/lib/python3.7/site-packages/datasets/data_files.py", line 544, in from_local_or_remote data_files = resolve_patterns_locally_or_by_urls(base_path, patterns, allowed_extensions) File "/home/gs603/miniconda3/envs/coderepair/lib/python3.7/site-packages/datasets/data_files.py", line 194, in resolve_patterns_locally_or_by_urls for path in _resolve_single_pattern_locally(base_path, pattern, allowed_extensions): File "/home/gs603/miniconda3/envs/coderepair/lib/python3.7/site-packages/datasets/data_files.py", line 144, in _resolve_single_pattern_locally raise FileNotFoundError(error_msg) FileNotFoundError: Unable to find '/home/gs603/ahf/pretrained/dataset/dataset_infos.json' at /home/gs603/ahf/pretrained ![image](https://user-images.githubusercontent.com/33253979/164413285-84ea65ac-9126-408f-9cd2-ce4751a5dd73.png) ![image](https://user-images.githubusercontent.com/33253979/164413338-4735142f-408b-41d9-ab87-8484de2be54f.png) the code is in the model.py,why I can't use the load_dataset function to load my local dataset?
2022-04-25T16:51:57Z
4,192
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-21T08:28:58Z
https://api.github.com/repos/huggingface/datasets/issues/4192/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4192/timeline
load_dataset can't load local dataset,Unable to find ...
https://api.github.com/repos/huggingface/datasets/issues/4192/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/33253979?v=4", "events_url": "https://api.github.com/users/ahf876828330/events{/privacy}", "followers_url": "https://api.github.com/users/ahf876828330/followers", "following_url": "https://api.github.com/users/ahf876828330/following{/other_user}", "gists_url": "https://api.github.com/users/ahf876828330/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ahf876828330", "id": 33253979, "login": "ahf876828330", "node_id": "MDQ6VXNlcjMzMjUzOTc5", "organizations_url": "https://api.github.com/users/ahf876828330/orgs", "received_events_url": "https://api.github.com/users/ahf876828330/received_events", "repos_url": "https://api.github.com/users/ahf876828330/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ahf876828330/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ahf876828330/subscriptions", "type": "User", "url": "https://api.github.com/users/ahf876828330" }
[]
null
completed
NONE
2022-04-22T07:39:53Z
null
I_kwDODunzps5IKbPK
[ "Hi! :)\r\n\r\nI believe that should work unless `dataset_infos.json` isn't actually a dataset. For Hugging Face datasets, there is usually a file named `dataset_infos.json` which contains metadata about the dataset (eg. the dataset citation, license, description, etc). Can you double-check that `dataset_infos.json` isn't just metadata please?", "Hi @ahf876828330, \r\n\r\nAs @stevhliu pointed out, the proper way to load a dataset is not trying to load its metadata file.\r\n\r\nIn your case, as the dataset script is local, you should better point to your local loading script:\r\n```python\r\ndataset = load_dataset(\"dataset/opus_books.py\")\r\n```\r\n\r\nPlease, feel free to re-open this issue if the previous code snippet does not work for you.", "> Hi! :)\r\n> \r\n> I believe that should work unless `dataset_infos.json` isn't actually a dataset. For Hugging Face datasets, there is usually a file named `dataset_infos.json` which contains metadata about the dataset (eg. the dataset citation, license, description, etc). Can you double-check that `dataset_infos.json` isn't just metadata please?\r\n\r\nYes,you are right!So if I have a metadata dataset local,How can I turn it to a dataset that can be used by the load_dataset() function?Are there some examples?", "The metadata file isn't a dataset so you can't turn it into one. You should try @albertvillanova's code snippet above (now merged in the docs [here](https://huggingface.co/docs/datasets/master/en/loading#local-loading-script)), which uses your local loading script `opus_books.py` to:\r\n\r\n1. Download the actual dataset. \r\n2. Once the dataset is downloaded, `load_dataset` will load it for you." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4192/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4192
https://github.com/huggingface/datasets/issues/4192
false
1,210,028,090
https://api.github.com/repos/huggingface/datasets/issues/4191/labels{/name}
**Is your feature request related to a problem? Please describe.** It is possible to create an `Array2D` column from a list of arrays of dimension 1. Similarly, I think it might be nice to be able to create a `Array3D` column from a list of lists of arrays of dimension 1. To illustrate my proposal, let's take the following toy dataset t: ```python import numpy as np from datasets import Dataset, features data_map = { 1: np.array([[0.2, 0,4],[0.19, 0,3]]), 2: np.array([[0.1, 0,4],[0.19, 0,3]]), } def create_toy_ds(): my_dict = {"id":[1, 2]} return Dataset.from_dict(my_dict) ds = create_toy_ds() ``` The following 2D processing works without any errors raised: ```python def prepare_dataset_2D(batch): batch["pixel_values"] = [data_map[index] for index in batch["id"]] return batch ds_2D = ds.map( prepare_dataset_2D, batched=True, remove_columns=ds.column_names, features=features.Features({"pixel_values": features.Array2D(shape=(2, 3), dtype="float32")}) ) ``` The following 3D processing doesn't work: ```python def prepare_dataset_3D(batch): batch["pixel_values"] = [[data_map[index]] for index in batch["id"]] return batch ds_3D = ds.map( prepare_dataset_3D, batched=True, remove_columns=ds.column_names, features=features.Features({"pixel_values": features.Array3D(shape=(1, 2, 3, dtype="float32")}) ) ``` The error raised is: ``` --------------------------------------------------------------------------- ArrowInvalid Traceback (most recent call last) [<ipython-input-6-676547e4cd41>](https://localhost:8080/#) in <module>() 3 batched=True, 4 remove_columns=ds.column_names, ----> 5 features=features.Features({"pixel_values": features.Array3D(shape=(1, 2, 3), dtype="float32")}) 6 ) 12 frames [/usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in map(self, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 1971 new_fingerprint=new_fingerprint, 1972 disable_tqdm=disable_tqdm, -> 1973 desc=desc, 1974 ) 1975 else: [/usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in wrapper(*args, **kwargs) 518 self: "Dataset" = kwargs.pop("self") 519 # apply actual function --> 520 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 521 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 522 for dataset in datasets: [/usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in wrapper(*args, **kwargs) 485 } 486 # apply actual function --> 487 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 488 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 489 # re-apply format to the output [/usr/local/lib/python3.7/dist-packages/datasets/fingerprint.py](https://localhost:8080/#) in wrapper(*args, **kwargs) 456 # Call actual function 457 --> 458 out = func(self, *args, **kwargs) 459 460 # Update fingerprint of in-place transforms + update in-place history of transforms [/usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py](https://localhost:8080/#) in _map_single(self, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, disable_tqdm, desc, cache_only) 2354 writer.write_table(batch) 2355 else: -> 2356 writer.write_batch(batch) 2357 if update_data and writer is not None: 2358 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file [/usr/local/lib/python3.7/dist-packages/datasets/arrow_writer.py](https://localhost:8080/#) in write_batch(self, batch_examples, writer_batch_size) 505 col_try_type = try_features[col] if try_features is not None and col in try_features else None 506 typed_sequence = OptimizedTypedSequence(batch_examples[col], type=col_type, try_type=col_try_type, col=col) --> 507 arrays.append(pa.array(typed_sequence)) 508 inferred_features[col] = typed_sequence.get_inferred_type() 509 schema = inferred_features.arrow_schema if self.pa_writer is None else self.schema /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib.array() /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib._handle_arrow_array_protocol() [/usr/local/lib/python3.7/dist-packages/datasets/arrow_writer.py](https://localhost:8080/#) in __arrow_array__(self, type) 175 storage = list_of_np_array_to_pyarrow_listarray(data, type=pa_type.value_type) 176 else: --> 177 storage = pa.array(data, pa_type.storage_dtype) 178 return pa.ExtensionArray.from_storage(pa_type, storage) 179 /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib.array() /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib._sequence_to_array() /usr/local/lib/python3.7/dist-packages/pyarrow/error.pxi in pyarrow.lib.pyarrow_internal_check_status() /usr/local/lib/python3.7/dist-packages/pyarrow/error.pxi in pyarrow.lib.check_status() ArrowInvalid: Can only convert 1-dimensional array values ``` **Describe the solution you'd like** No error in the second scenario and an identical result to the following snippets. **Describe alternatives you've considered** There are other alternatives that work such as: ```python def prepare_dataset_3D_bis(batch): batch["pixel_values"] = [[data_map[index].tolist()] for index in batch["id"]] return batch ds_3D_bis = ds.map( prepare_dataset_3D_bis, batched=True, remove_columns=ds.column_names, features=features.Features({"pixel_values": features.Array3D(shape=(1, 2, 3), dtype="float32")}) ) ``` or ```python def prepare_dataset_3D_ter(batch): batch["pixel_values"] = [data_map[index][np.newaxis, :, :] for index in batch["id"]] return batch ds_3D_ter = ds.map( prepare_dataset_3D_ter, batched=True, remove_columns=ds.column_names, features=features.Features({"pixel_values": features.Array3D(shape=(1, 2, 3), dtype="float32")}) ) ``` But both solutions require the user to be aware that `data_map[index]` is an `np.array` type. cc @lhoestq as we discuss this offline :smile:
2022-05-12T15:08:40Z
4,191
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-04-20T18:04:32Z
https://api.github.com/repos/huggingface/datasets/issues/4191/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4191/timeline
feat: create an `Array3D` column from a list of arrays of dimension 2
https://api.github.com/repos/huggingface/datasets/issues/4191/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/55560583?v=4", "events_url": "https://api.github.com/users/SaulLu/events{/privacy}", "followers_url": "https://api.github.com/users/SaulLu/followers", "following_url": "https://api.github.com/users/SaulLu/following{/other_user}", "gists_url": "https://api.github.com/users/SaulLu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SaulLu", "id": 55560583, "login": "SaulLu", "node_id": "MDQ6VXNlcjU1NTYwNTgz", "organizations_url": "https://api.github.com/users/SaulLu/orgs", "received_events_url": "https://api.github.com/users/SaulLu/received_events", "repos_url": "https://api.github.com/users/SaulLu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SaulLu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SaulLu/subscriptions", "type": "User", "url": "https://api.github.com/users/SaulLu" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
completed
CONTRIBUTOR
2022-05-12T15:08:40Z
null
I_kwDODunzps5IH5A6
[ "Hi @SaulLu, thanks for your proposal.\r\n\r\nJust I got a bit confused about the dimensions...\r\n- For the 2D case, you mention it is possible to create an `Array2D` from a list of arrays of dimension 1\r\n- However, you give an example of creating an `Array2D` from arrays of dimension 2:\r\n - the values of `data_map` are arrays of dimension 2\r\n - the outer list in `prepare_dataset_2D` should not be taken into account in the dimension counting, as it is used because in `map` you pass `batched=True`\r\n\r\nNote that for the 3D alternatives you mention:\r\n- In `prepare_dataset_3D_ter`, you create an `Array3D` from arrays of dimension 3:\r\n - the array `data_map[index][np.newaxis, :, :]` has dimension 3\r\n - the outer list in `prepare_dataset_3D_ter` is the one used by `batched=True`\r\n- In `prepare_dataset_3D_bis`, you create an `Array3D` from a list of list of lists:\r\n - the value of `data_map[index].tolist()` is a list of lists\r\n - it is enclosed by another list `[data_map[index].tolist()]`, thus giving a list of list of lists\r\n - the outer list is the one used by `batched=True`\r\n\r\nTherefore, if I understand correctly, your request would be to be able to create an `Array3D` from a list of an array of dimension 2:\r\n- In `prepare_dataset_3D`, `data_map[index]` is an array of dimension 2\r\n- it is enclosed by a list `[data_map[index]]`, thus giving a list of an array of dimension 2\r\n- the outer list is the one used by `batched=True`\r\n\r\nPlease, feel free to tell me if I did not understand you correctly.", "Hi @albertvillanova ,\r\n\r\nIndeed my message was confusing and you guessed right :smile: : I think would be interesting to be able to create an Array3D from a list of an array of dimension 2. \r\n\r\nFor the 2D case I should have given as a \"similar\" example:\r\n```python\r\n\r\ndata_map_1D = {\r\n 1: np.array([0.2, 0.4]),\r\n 2: np.array([0.1, 0.4]),\r\n}\r\n\r\ndef prepare_dataset_2D(batch):\r\n batch[\"pixel_values\"] = [[data_map_1D[index]] for index in batch[\"id\"]]\r\n return batch\r\n \r\nds_2D = ds.map(\r\n prepare_dataset_2D, \r\n batched=True, \r\n remove_columns=ds.column_names, \r\n features=features.Features({\"pixel_values\": features.Array2D(shape=(1, 2), dtype=\"float32\")})\r\n)\r\n```" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4191/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4191
https://github.com/huggingface/datasets/issues/4191
false
1,209,901,677
https://api.github.com/repos/huggingface/datasets/issues/4190/labels{/name}
This PR adds a `max_shard_size` param to `push_to_hub` and deprecates `shard_size` in favor of this new param to have a more descriptive name (a shard has at most the `shard_size` bytes in `push_to_hub`) for the param and to align the API with [Transformers](https://github.com/huggingface/transformers/blob/ff06b177917384137af2d9585697d2d76c40cdfc/src/transformers/modeling_utils.py#L1350).
2022-04-22T13:58:25Z
4,190
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-20T16:08:01Z
https://api.github.com/repos/huggingface/datasets/issues/4190/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4190/timeline
Deprecate `shard_size` in `push_to_hub` in favor of `max_shard_size`
https://api.github.com/repos/huggingface/datasets/issues/4190/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-04-22T13:52:00Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4190.diff", "html_url": "https://github.com/huggingface/datasets/pull/4190", "merged_at": "2022-04-22T13:52:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/4190.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4190" }
PR_kwDODunzps42gK3y
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4190/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4190
https://github.com/huggingface/datasets/pull/4190
true
1,209,881,351
https://api.github.com/repos/huggingface/datasets/issues/4189/labels{/name}
Document how to use FAISS index for special operations, by accessing the index itself. Close #4029.
2022-05-06T08:43:10Z
4,189
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-20T15:51:56Z
https://api.github.com/repos/huggingface/datasets/issues/4189/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4189/timeline
Document how to use FAISS index for special operations
https://api.github.com/repos/huggingface/datasets/issues/4189/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-05-06T08:35:52Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4189.diff", "html_url": "https://github.com/huggingface/datasets/pull/4189", "merged_at": "2022-05-06T08:35:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/4189.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4189" }
PR_kwDODunzps42gGv5
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4189/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4189
https://github.com/huggingface/datasets/pull/4189
true
1,209,740,957
https://api.github.com/repos/huggingface/datasets/issues/4188/labels{/name}
Support streaming cnn_dailymail dataset. Fix #3969. CC: @severo
2022-05-11T13:39:06Z
4,188
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-20T14:04:36Z
https://api.github.com/repos/huggingface/datasets/issues/4188/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4188/timeline
Support streaming cnn_dailymail dataset
https://api.github.com/repos/huggingface/datasets/issues/4188/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-04-20T15:52:49Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4188.diff", "html_url": "https://github.com/huggingface/datasets/pull/4188", "merged_at": "2022-04-20T15:52:49Z", "patch_url": "https://github.com/huggingface/datasets/pull/4188.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4188" }
PR_kwDODunzps42fpMv
[ "_The documentation is not available anymore as the PR was closed or merged._", "Did you run the `datasets-cli` command before merging to make sure you generate all the examples ?" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4188/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4188
https://github.com/huggingface/datasets/pull/4188
true
1,209,721,532
https://api.github.com/repos/huggingface/datasets/issues/4187/labels{/name}
Right now if you pass both the `bytes` and a local `path` for audio or image data, then the `bytes` are unnecessarily written in the Arrow file, while we could just keep the local `path`. This PR discards the `bytes` when the audio or image file exists locally. In particular it's common for audio datasets builders to provide both the bytes and the local path in order to work for both streaming (using the bytes) and non-streaming mode (using a local file - which is often required for audio). cc @patrickvonplaten
2022-04-21T09:17:00Z
4,187
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-20T13:50:37Z
https://api.github.com/repos/huggingface/datasets/issues/4187/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4187/timeline
Don't duplicate data when encoding audio or image
https://api.github.com/repos/huggingface/datasets/issues/4187/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-04-21T09:10:47Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4187.diff", "html_url": "https://github.com/huggingface/datasets/pull/4187", "merged_at": "2022-04-21T09:10:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/4187.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4187" }
PR_kwDODunzps42flGp
[ "_The documentation is not available anymore as the PR was closed or merged._", "I'm not familiar with the concept of streaming vs non-streaming in HF datasets. I just wonder that you have the distinction here. Why doesn't it work to always make use of `bytes`? \"using a local file - which is often required for audio\" - why would that be?\r\n\r\nThe `path` would always point to some location in the `cache_dir`? I think this can be problematic. I would have expected that after I did `dataset.save_to_disk(...)` that I can remove the cache dir. But maybe just because I'm not familiar with HF. Or maybe the docs can be improved to clarify this.\r\n", "We could always load every data file into `bytes` and save it this way the audio as bytes in `arrow` format, but the problem then would be that it makes the `file` column useless, *i.e.* people cannot inspect the audio file locally anymore or else they would need to first save bytes as a file which is not evident. This either breaks backwards compatibility or forces the user to stored 2x the required size locally. There was a longer discussion here: https://github.com/huggingface/datasets/issues/3663\r\n\r\nIt's a good argument though that `dataset.save_to_disk(...)` should save everything that is needed to the disk and should be independent of other folders, but I do think the arguments of #3663 to not break backwards compatibility and to allow people to inspect the downloaded audio files locally are a bit more important here. \r\n\r\nBut maybe, we could add a flag, `save_files_as_bytes` or `make_independent`, `make_self_contained` or a better name to `save_to_disk(...)` and `push_to_hub(...)` that would allow to make the resulting folder completely independent. ", "What do you think @mariosasko @lhoestq @polinaeterna @anton-l ?\r\n", "For context: you can either store the path to local images or audio files, or the bytes of those files.\r\n\r\nIf your images and audio files are local files, then the arrow file from `save_to_disk` will store paths to these files.\r\nIf you want to include the bytes or your images or audio files instead, you must `read()` those files first.\r\nThis can be done by storing the \"bytes\" instead of the \"path\" of the images or audio files.\r\n\r\nOn the other hand, the resulting Parquet files from `push_to_hub` are self-contained, so that anyone can reload the dataset from the Hub. If your dataset contains image or audio data, the Parquet files will store the bytes of your images or audio files.\r\n\r\nFor now I just updated the documentation: https://github.com/huggingface/datasets/pull/4193. Maybe we can also embed the image and audio bytes in `save_to_disk` when we implement sharding, so that is can be done as efficiently as `push_to_hub`.\r\n\r\nAnyway, merging this one :)" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4187/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4187
https://github.com/huggingface/datasets/pull/4187
true
1,209,463,599
https://api.github.com/repos/huggingface/datasets/issues/4186/labels{/name}
null
2022-04-22T12:54:44Z
4,186
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-20T10:04:51Z
https://api.github.com/repos/huggingface/datasets/issues/4186/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4186/timeline
Fix outdated docstring about default dataset config
https://api.github.com/repos/huggingface/datasets/issues/4186/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-04-22T12:48:31Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4186.diff", "html_url": "https://github.com/huggingface/datasets/pull/4186", "merged_at": "2022-04-22T12:48:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/4186.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4186" }
PR_kwDODunzps42evF5
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4186/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4186
https://github.com/huggingface/datasets/pull/4186
true
1,209,429,743
https://api.github.com/repos/huggingface/datasets/issues/4185/labels{/name}
https://github.com/huggingface/datasets/blob/cd3ce34ab1604118351e1978d26402de57188901/datasets/librispeech_asr/librispeech_asr.py#L53 > Note that in order to limit the required storage for preparing this dataset, the audio > is stored in the .flac format and is not converted to a float32 array. To convert, the audio > file to a float32 array, please make use of the `.map()` function as follows: > > ```python > import soundfile as sf > def map_to_array(batch): > speech_array, _ = sf.read(batch["file"]) > batch["speech"] = speech_array > return batch > dataset = dataset.map(map_to_array, remove_columns=["file"]) > ``` Is this still true? In my case, `ds["train.100"]` returns: ``` Dataset({ features: ['file', 'audio', 'text', 'speaker_id', 'chapter_id', 'id'], num_rows: 28539 }) ``` and taking the first instance yields: ``` {'file': '374-180298-0000.flac', 'audio': {'path': '374-180298-0000.flac', 'array': array([ 7.01904297e-04, 7.32421875e-04, 7.32421875e-04, ..., -2.74658203e-04, -1.83105469e-04, -3.05175781e-05]), 'sampling_rate': 16000}, 'text': 'CHAPTER SIXTEEN I MIGHT HAVE TOLD YOU OF THE BEGINNING OF THIS LIAISON IN A FEW LINES BUT I WANTED YOU TO SEE EVERY STEP BY WHICH WE CAME I TO AGREE TO WHATEVER MARGUERITE WISHED', 'speaker_id': 374, 'chapter_id': 180298, 'id': '374-180298-0000'} ``` The `audio` `array` seems to be already decoded. So such convert/decode code as mentioned in the doc is wrong? But I wonder, is it actually stored as flac on disk, and the decoding is done on-the-fly? Or was it decoded already during the preparation and is stored as raw samples on disk? Note that I also used `datasets.load_dataset("librispeech_asr", "clean").save_to_disk(...)` and then `datasets.load_from_disk(...)` in this example. Does this change anything on how it is stored on disk? A small related question: Actually I would prefer to even store it as mp3 or ogg on disk. Is this easy to convert?
2022-04-21T11:00:53Z
4,185
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-04-20T09:35:55Z
https://api.github.com/repos/huggingface/datasets/issues/4185/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4185/timeline
Librispeech documentation, clarification on format
https://api.github.com/repos/huggingface/datasets/issues/4185/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59132?v=4", "events_url": "https://api.github.com/users/albertz/events{/privacy}", "followers_url": "https://api.github.com/users/albertz/followers", "following_url": "https://api.github.com/users/albertz/following{/other_user}", "gists_url": "https://api.github.com/users/albertz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertz", "id": 59132, "login": "albertz", "node_id": "MDQ6VXNlcjU5MTMy", "organizations_url": "https://api.github.com/users/albertz/orgs", "received_events_url": "https://api.github.com/users/albertz/received_events", "repos_url": "https://api.github.com/users/albertz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertz/subscriptions", "type": "User", "url": "https://api.github.com/users/albertz" }
[]
null
null
NONE
null
null
I_kwDODunzps5IFm7v
[ "(@patrickvonplaten )", "Also cc @lhoestq here", "The documentation in the code is definitely outdated - thanks for letting me know, I'll remove it in https://github.com/huggingface/datasets/pull/4184 .\r\n\r\nYou're exactly right `audio` `array` already decodes the audio file to the correct waveform. This is done on the fly, which is also why one should **not** do `ds[\"audio\"][\"array\"][0]` as this will decode all dataset samples, but instead `ds[0][\"audio\"][\"array\"]` see: https://huggingface.co/docs/datasets/audio_process#audio-datasets\r\n\r\n", "So, again to clarify: On disk, only the raw flac file content is stored? Is this also the case after `save_to_disk`?\r\n\r\nAnd is it simple to also store it re-encoded as ogg or mp3 instead?\r\n", "Hey, \r\n\r\nSorry yeah I was just about to look into this! We actually had an outdated version of Librispeech ASR that didn't save any files, but instead converted the audio files to a byte string, then was then decoded on-the-fly. This however is not very user-friendly so we recently decided to instead show the full path of the audio files with the `path` parameter.\r\n\r\nI'm currently changing this for Librispeech here: https://github.com/huggingface/datasets/pull/4184 .\r\nYou should be able to see the audio file in the original `flac` format under `path` then. I don't think it's a good idea to convert to MP3 out-of-the-box, but we could maybe think about some kind of convert function for audio datasets cc @lhoestq ? ", "> I don't think it's a good idea to convert to MP3 out-of-the-box, but we could maybe think about some kind of convert function for audio datasets cc @lhoestq ?\r\n\r\nSure, I would expect that `load_dataset(\"librispeech_asr\")` would give you the original (not re-encoded) data (flac or already decoded). So such re-encoding logic would be some separate generic function. So I could do sth like `dataset.reencode_as_ogg(**ogg_encode_opts).save_to_disk(...)` or so.\r\n", "A follow-up question: I wonder whether a Parquet dataset is maybe more what we actually want to have? (Following also my comment here: https://github.com/huggingface/datasets/pull/4184#issuecomment-1105045491.) Because I think we actually would prefer to embed the data content in the dataset.\r\n\r\nSo, instead of `save_to_disk`/`load_from_disk`, we would use `to_parquet`,`from_parquet`? Is there any downside? Are arrow files more efficient?\r\n\r\nRelated is also the doc update in #4193.\r\n", "`save_to_disk` saves the dataset as an Arrow file, which is the format we use to load a dataset using memory mapping. This way the dataset does not fill your RAM, but is read from your disk instead.\r\n\r\nTherefore you can directly reload a dataset saved with `save_to_disk` using `load_from_disk`.\r\n\r\nParquet files are used for cold storage: to use memory mapping on a Parquet dataset, you first have to convert it to Arrow. We use Parquet to reduce the I/O when pushing/downloading data from the Hugging face Hub. When you load a Parquet file from the Hub, it is converted to Arrow on the fly during the download." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4185/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4185
https://github.com/huggingface/datasets/issues/4185
false
1,208,592,669
https://api.github.com/repos/huggingface/datasets/issues/4184/labels{/name}
Add `"all"` config to Librispeech Closed #4179
2022-08-29T06:35:57Z
4,184
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-19T16:27:56Z
https://api.github.com/repos/huggingface/datasets/issues/4184/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4184/timeline
[Librispeech] Add 'all' config
https://api.github.com/repos/huggingface/datasets/issues/4184/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[]
null
null
CONTRIBUTOR
2022-04-22T09:45:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4184.diff", "html_url": "https://github.com/huggingface/datasets/pull/4184", "merged_at": "2022-04-22T09:45:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/4184.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4184" }
PR_kwDODunzps42cB2j
[ "Fix https://github.com/huggingface/datasets/issues/4179", "_The documentation is not available anymore as the PR was closed or merged._", "Just that I understand: With this change, simply doing `load_dataset(\"librispeech_asr\")` is possible and returns the whole dataset?\r\n\r\nAnd to get the subsets, I do sth like:\r\n```python\r\nds = load_dataset(\"librispeech_asr\")\r\ntrain_ds = ds[\"train\"]\r\ndev_clean_ds = ds[\"dev-clean\"]\r\ndev_other_ds = ds[\"dev-other\"]\r\ntest_clean_ds = ds[\"test-clean\"]\r\ntest_other_ds = ds[\"test-other\"]\r\n```\r\n?\r\n", "> Just that I understand: With this change, simply doing `load_dataset(\"librispeech_asr\")` is possible and returns the whole dataset?\r\n> \r\n> And to get the subsets, I do sth like:\r\n> \r\n> ```python\r\n> ds = load_dataset(\"librispeech_asr\")\r\n> train_ds = ds[\"train\"]\r\n> dev_clean_ds = ds[\"dev-clean\"]\r\n> dev_other_ds = ds[\"dev-other\"]\r\n> test_clean_ds = ds[\"test-clean\"]\r\n> test_other_ds = ds[\"test-other\"]\r\n> ```\r\n> \r\n> ?\r\n\r\nYou could do:\r\n\r\n\r\n```python\r\nds = load_dataset(\"librispeech_asr\", \"all\") # <- note that we have to pass a config\r\ntrain_ds = ds[\"train\"]\r\ndev_clean_ds = ds[\"dev-clean\"]\r\ndev_other_ds = ds[\"dev-other\"]\r\ntest_clean_ds = ds[\"test-clean\"]\r\ntest_other_ds = ds[\"test-other\"]\r\n```", "So, `load_dataset(\"librispeech_asr\")` is not possible, it must be `load_dataset(\"librispeech_asr\", \"all\")`?\r\n\r\nWhy is that?\r\n\r\nThe docs say:\r\n```\r\nname: `str` name, optional configuration for the dataset that affects the data generated on disk. Different\r\n `builder_config`s will have their own subdirectories and versions.\r\n If not provided, uses the first configuration in self.BUILDER_CONFIGS\r\n```\r\nhttps://github.com/huggingface/datasets/blob/cd3ce34ab1604118351e1978d26402de57188901/src/datasets/builder.py#L228\r\n\r\nOr maybe you could just define `DEFAULT_CONFIG_NAME`?\r\n", "> If not provided, uses the first configuration in self.BUILDER_CONFIGS\r\n\r\nOh crap this is outdated documentation. No it doesn't take the first config by default.\r\n\r\nEDIT: opened a PR to fix this: https://github.com/huggingface/datasets/pull/4186", "> No it doesn't take the first config by default.\r\n\r\nBut defining `DEFAULT_CONFIG_NAME` would work?\r\n\r\nSo should we define `DEFAULT_CONFIG_NAME = \"all\"` here as well? I think this is a reasonable default config.\r\n\r\nDon't most datasets have some default config?\r\n", "> But defining DEFAULT_CONFIG_NAME would work?\r\n>\r\n> So should we define DEFAULT_CONFIG_NAME = \"all\" here as well? I think this is a reasonable default config.\r\n\r\nYes that would work, and I also find it reasonable to do it :)\r\n\r\n> Don't most datasets have some default config?\r\n\r\nMost datasets only have one configuration, so the single configuration is the default one. Then other datasets gave several configurations, and whether they have a default one is decided case-by-case.\r\n\r\ne.g. `glue` is a benchmark and doesn't have a default task, one must choose which task of `glue` they want to use explicitely.", "Thanks a lot for the feedback! \r\n\r\nUsing `\"all\"` now as the default config. I changed the layout a bit so that there is not a single \"train\", but instead we have multiple \"train.clean.100\", \"train.clean.360\", \"train.other.500\". This way we don't even need to do filtering and it's also cleaner IMO.\r\n\r\n@albertz - you should now be able to do the following:\r\n\r\n```python\r\nload_dataset(\"librispeech_asr\") # <- run this once to download, prepare dataset and cache everything\r\n\r\n# The following operations will be very fast since all the downloading and processing is already cached\r\ntrain_1 = load_dataset(\"librispeech_asr\", split=\"train.clean.100\")\r\nprint(train_1)\r\ntrain_2 = load_dataset(\"librispeech_asr\", split=\"train.clean.100+train.clean.360\")\r\nprint(train_2)\r\ntrain_full = load_dataset(\"librispeech_asr\", split=\"train.clean.100+train.clean.360+train.other.500\")\r\nprint(train_full)\r\ndev_clean_ds = load_dataset(\"librispeech_asr\", split=\"validation.clean\")\r\nprint(dev_clean_ds)\r\ndev_other_ds = load_dataset(\"librispeech_asr\", split=\"validation.other\")\r\nprint(dev_other_ds)\r\ntest_clean_ds = load_dataset(\"librispeech_asr\", split=\"test.clean\")\r\nprint(test_clean_ds)\r\ntest_other_ds = load_dataset(\"librispeech_asr\", split=\"test.other\")\r\nprint(test_other_ds)\r\n```\r\n\r\n\r\n", "Think this way we have the best of both worlds. Also @lhoestq, I think we could highlight better in the docs that it's possible to combine different splits. We do this actually quite a lot for speech. For Common Voice many people include \"validation\" in the training if the data is too small, e.g.: https://github.com/huggingface/transformers/blob/ff06b177917384137af2d9585697d2d76c40cdfc/examples/pytorch/speech-recognition/run_speech_recognition_ctc.py#L147\r\n\r\nShould we maybe add a short section to the loading tutorial here: https://huggingface.co/docs/datasets/v2.1.0/en/loading#hugging-face-hub ? (Happy to do it)", "Is there any advantage or difference in calling `load_dataset` multiple times for each split? Or why not just call `load_dataset` once and then access each split?\r\n\r\nNote in our case, we cannot really use the caching mechanism because we have a recipe pipeline used by multiple users (and I think a common cache dir for all users might end up in problems) and we basically would use `load_dataset(\"librispeech_asr\").save_to_disk(...)` and then later `load_from_disk(...)`. (See here: https://github.com/rwth-i6/i6_core/pull/253)\r\n\r\nSo with `load_from_disk`, we cannot really provide the split this way, so we anyway would do sth like:\r\n```python\r\nds = datasets.load_from_disk(...)\r\ntrain = ds[\"train\"]\r\n```\r\nOr with your latest proposal, it would look like:\r\n```python\r\nds = datasets.load_from_disk(...)\r\ntrain_ds = datasets.concatenate_datasets(\r\n [ds[\"train.clean.100\"], ds[\"train.clean.360\"], ds[\"train.other.500\"]])\r\n```\r\nright?\r\n", "> Is there any advantage or difference in calling `load_dataset` multiple times for each split? Or why not just call `load_dataset` once and then access each split?\r\n> \r\n> Note in our case, we cannot really use the caching mechanism because we have a recipe pipeline used by multiple users (and I think a common cache dir for all users might end up in problems) and we basically would use `load_dataset(\"librispeech_asr\").save_to_disk(...)` and then later `load_from_disk(...)`. (See here: [rwth-i6/i6_core#253](https://github.com/rwth-i6/i6_core/pull/253))\r\n> \r\n> So with `load_from_disk`, we cannot really provide the split this way, so we anyway would do sth like:\r\n> \r\n> ```python\r\n> ds = datasets.load_from_disk(...)\r\n> train = ds[\"train\"]\r\n> ```\r\n> \r\n> Or with your latest proposal, it would look like:\r\n> \r\n> ```python\r\n> ds = datasets.load_from_disk(...)\r\n> train_ds = datasets.concatenate_datasets(\r\n> [ds[\"train.clean.100\"], ds[\"train.clean.360\"], ds[\"train.other.500\"]])\r\n> ```\r\n> \r\n> right?\r\n\r\nI see the use case! The only advantage by calling `datasets` multiple times is that one can easily \"merge\" splits with `\"+\"`, but yeah you can do the exact same with `concatenate`.\r\n\r\n@lhoestq what do you think is the best approach with `load_from_disk`? \r\n\r\n@albertz, you could also define the `cache_dir` when doing `load_dataset(...)` which will then put all the relevant `arrow` files int the cache dir that you defined, e.g.:\r\n\r\n```python\r\nload_dataset(\"librispeech_asr\", cache_dir=\"/easy/to/access/directory\")\r\n```", "@albertz, I took a read through https://github.com/rwth-i6/i6_core/pull/253 . \r\n\r\nI think the best would be the following:\r\n\r\n1. Do `ds = load_dataset(..., cache_dir=\"/dir/that/is/easy/to/access\")` <- having merged this PR, this will save all the original `.flac` files in the `cache_dir`\r\n2. Do `ds.save_to_disk(\"local/path\")` this should then only save the `arrow.format` with a `path` string to the audio files which are located in `cache_dir` <- this won't require a lot of memory after https://github.com/huggingface/datasets/pull/4184#discussion_r854132740 is fixed and can be done for each person individually.\r\n3. `ds = datasets.load_from_disk(\"local/path\")` can the be used. An object of `ds` will then have a `path` variable that links to the original audio files in the `cache_dir`. You can change these audio files then easily to `.mp3. You could do this with the `.map(...)` function, e.g. define a function that maps through all audio files, load them and then save them on disk afterward.", "@lhoestq - I think this one is good to go", "> @albertz, I took a read through [rwth-i6/i6_core#253](https://github.com/rwth-i6/i6_core/pull/253) .\r\n> \r\n> I think the best would be the following:\r\n> \r\n> 1. Do `ds = load_dataset(..., cache_dir=\"/dir/that/is/easy/to/access\")` <- having merged this PR, this will save all the original `.flac` files in the `cache_dir`\r\n> 2. Do `ds.save_to_disk(\"local/path\")` this should then only save the `arrow.format` with a `path` string to the audio files which are located in `cache_dir` <- this won't require a lot of memory after [[Librispeech] Add 'all' config #4184 (comment)](https://github.com/huggingface/datasets/pull/4184#discussion_r854132740) is fixed and can be done for each person individually.\r\n> 3. `ds = datasets.load_from_disk(\"local/path\")` can the be used. An object of `ds` will then have a `path` variable that links to the original audio files in the `cache_dir`. You can change these audio files then easily to `.mp3. You could do this with the `.map(...)` function, e.g. define a function that maps through all audio files, load them and then save them on disk afterward.\r\n\r\nOh, so you say that our current implementation in https://github.com/rwth-i6/i6_core/pull/253 is broken? Because our cache dir is just some temp directory which will be removed afterwards, and we just store what we get out of `save_to_disk`. I think it would be good to clarify that in the doc of `save_to_disk`, that this is not enough and can depend on files from the cache dir. (@dthulke)\r\n\r\nSo, you say we anyway need to share the cache dir among users? But we would want to make sure that after the initial download and preparation of the data, this is set to readonly, because we want to make sure that other people will not modify the data in any way. Right?\r\n\r\nBut then, we don't really need the `save_to_disk` and `load_from_disk` at all, right?\r\n", "@albertz \r\n\r\n> Oh, so you say that our current implementation in https://github.com/rwth-i6/i6_core/pull/253 is broken? Because our cache dir is just some temp directory which will be removed afterwards, and we just store what we get out of save_to_disk. I think it would be good to clarify that in the doc of save_to_disk, that this is not enough and can depend on files from the cache dir. (@dthulke)\r\n\r\nOh, I wasn't aware that audio files are handled this way. Then we should have the cache directory as an additional job output, so that we keep the audio files. \r\n\r\n> So, you say we anyway need to share the cache dir among users?\r\n\r\nNo, the cache dir can still be a directory in the job output folder. Then the audio paths in the corresponding dataset column correspond to the flac files in that directory. This way the \"output\" of the job is contained into the job directory and we don't write files to a global cache directory that is independent of the sisyphus graph.\r\n\r\nIf we want to share the audio data between different users, we can just link to a central instance of the job (similar to how we do it with the `DownloadLibriSpeechCorpusJob`).", "@dthulke - that's a good point actually! So you can do both things:\r\n\r\n1. Convert all audio files to bytes. Bytes can be saved by `arrow` so in this case you can do `save_to_disk(...)`, but then you cannot really inspect the audio files locally as they'll just be saved within a large arrow file (this actually used to be the default case but we're changing this now). The problem of this is summarized here a bit: https://github.com/huggingface/datasets/issues/3663 . You can still do this if you'd like, e.g. you could do:\r\n\r\n```python\r\nds = load_dataset(\"librispeech_asr\")\r\n\r\ndef read_file(batch):\r\n with open(batch[\"file\"], \"r\") as f:\r\n batch[\"bytes\"] = f.read() \r\n return batch\r\n\r\nds = ds.map(read_file)\r\nds.save_to_disk(\"/path\") <- the saved arrow object will now contain everything you need\r\n```\r\n\r\nhowever this is not recommend - it's should be much easier to just save the path to the downloaded audio files.\r\n\r\n2. Not convert audio files to bytes, but just leave them in their original file format. Then only the path to the original files will be save in arrow. This will be the default case. This means that when you do `load_dataset(...)` both the orginal audio data and the arrow file will be saved in the `cache_dir` (which can be saved locally for every user or in a shared cache - we actually use a shared cache quite a bit at Hugging Face). When do you do `save_to_disk(...)` now only the `path` will be saved in `arrow` format (after this PR is merged, you'll see that the `arrow files should be very light weight` meaning that `save_to_disk(...)` can be done for every user, but has a dependency on the `cache_dir` (because the audio files live there).\r\n\r\n=> Now what you could do as well would be to simply move all the audio files to the folder you want (the `save_to_disk(...)` folder) and then change the path of every sample to this folder (maybe with `map(...)`) and then this folder would be self contained. I do however think it's better to just specific a `cache_dir` and re-use `load_dataset(...)` every time instead of `load_from_disk` or `save_to_disk(...)`. Note that you can even pass the relevant cache files to `load_dataset(...)` here: https://huggingface.co/docs/datasets/v2.1.0/en/package_reference/loading_methods#datasets.load_dataset.data_files in which case you can be 100% sure that nothing is redownloaded. \r\n\r\nWe discussed storing audio files quite a bit, e.g. see: https://github.com/huggingface/datasets/issues/3663 and had (too many) changes around this topic recently, but we've come to the conclusion that the best is to leave the audio format in the format it was originally (`.flac` for Librispeech) so that the user can easily inspect it / understand the data. Arrow cannot save data is `.flac` so we'll just save a path to the original data. Curious to hear your guys opinion on this as well.", "So what I would suggest here is to do the following:\r\n\r\n1. Do `load_dataset(..., cache_dir=/a/read-only/folder)`\r\n2. \r\n- Either just re-use `load_dataset(..., cache_dir=...)` which should always re-use the data in the `cache_dir` since the hash of the url matches - so there should never be any duplicated downloading \r\n\r\nor \r\n\r\n- If you want to store the files in MP3 locally, first convert the files to MP3 in the read-only folder, then take do `ds.save_to_disk(/some/path)` which will save the correct path to the read-only folder to MP3 and then you can easily re-use the small arrow dataset that is saved in `/some/path`", "> So what I would suggest here is to do the following:\r\n> \r\n> 1. Do `load_dataset(..., cache_dir=/a/read-only/folder)`\r\n> \r\n> * Either just re-use `load_dataset(..., cache_dir=...)` which should always re-use the data in the `cache_dir` since the hash of the url matches - so there should never be any duplicated downloading\r\n> \r\n> or\r\n> \r\n> * If you want to store the files in MP3 locally, first convert the files to MP3 in the read-only folder, then take do `ds.save_to_disk(/some/path)` which will save the correct path to the read-only folder to MP3 and then you can easily re-use the small arrow dataset that is saved in `/some/path`\r\n\r\nAlso relevant here: https://github.com/huggingface/datasets/issues/3663", "I also added some documentation about how `save_to_disk` handles audio files here: https://github.com/huggingface/datasets/pull/4193", "> > So, you say we anyway need to share the cache dir among users?\r\n> \r\n> No, the cache dir can still be a directory in the job output folder.\r\n\r\n@dthulke But this is what I mean. When we share the job output folder, it means we share the cache dir among users.\r\n\r\nI wonder if `load_dataset(..., cache_dir=job_output_cache_dir)` is always save to do then, that it really would not modify the `job_output_cache_dir`.\r\n\r\nWe could enforce that by making the `job_output_cache_dir` read-only afterwards. We currently don't do this.\r\n\r\n@patrickvonplaten @dthulke But in any case, we actually prefer the data content to be inside the dataset (the arrow files). Lots of small files would be very problematic for our cache manager. We have one main copy of the data on NFS, but accessing the NFS directly by all computing nodes is not feasible, so the cache manager will have copies of the files on the nodes. So it means, whenever we access some file, we query the cache manager DB whether the file is already cached somewhere (some other computing node) and if so, it copies it from the other computing node and not from NFS. This works very well when there are not too many files (but the files can be big). So, we want to have only a few but big files. Even for NFS access this is much better.\r\n\r\nI also commented in #3663.\r\n", "Hey @albertz @dthulke,\r\n\r\nThanks a lot for your input! \r\n\r\nWe've discussed quite a bit with @lhoestq and we think the best approach is the following:\r\n\r\n\r\na)\r\n`load_dataset(...)` will not store both bytes and the files because this would mean that 3x the size of the dataset would often be needed (1. the compressed `tar.gz` file, 2. the extracted file b, 3. the raw bytes in arrow format). \r\n\r\nFor canonical datasets like librispeech and common voice I think we want to keep the dataset filenames because of i) no breaking changes and ii) reasons explained in #3663\r\n\r\nHowever it's also trivial to write your own datasetset downloading script of librispeech and just not extract the folder e.g. this line: https://huggingface.co/datasets/common_voice/blob/main/common_voice.py#L671\r\n\r\nAnd then it'll be allowed to save the bytes and the dataset will be self-contained out-of-the-box when using `load_dataset(...)`\r\n\r\nb) Now, one major problem that you guys uncovered is that `save_to_disk(...)` is currently not necessarily saving a dataset to be self-contained. We will change that asap. This means that after we've corrected this when you do download the canonical librispeech dataset the following will work:\r\n\r\n```python\r\nds = load_dataset(\"....\") # <- here we have a dependency on the filepathes\r\nds[0][\"audio\"][\"bytes\"] # <- will not work\r\n\r\nds.save_to_disk(\"/local/path\") # <- now we want to have a self-contained dataset in arrow format, so we load the files into bytes and save it in arrow format\r\n\r\n# now you can delete everything besides \"/local/path\"\r\n\r\nds = load_from_disk(\"/local/path\") # <- this will work\r\n```\r\n\r\nSo either option a) where you define your own librispeech data downloading script (you guys could just sign up here: https://huggingface.co/join) and upload a dataset loading script in private mode so that no one can see it and you would always store the audio as bytes or b) where you first load then save to disk then delete cache would work. \r\n\r\nHope that fits in your vision :-)\r\n\r\ncc @lhoestq @mariosasko ", "@patrickvonplaten sounds like a good approach to me. For b) this could even be configurable with a parameter like `embed_external_files` as you have for `push_to_hub` (if people prefer to keep separate audio files).\r\n", "> However it's also trivial to write your own datasetset downloading script of librispeech and just not extract the folder\r\n\r\nI don't exactly understand. In all cases, we need to extract it to prepare the dataset, or not? No matter if we want to store the raw bytes inside the dataset or leaving them as local files. Just in the first case, we can safely delete the extracted files after the dataset preparation.\r\n\r\n> `save_to_disk(...)` is currently not necessarily saving a dataset to be self-contained. We will change that asap.\r\n\r\nFor us, this sounds exactly like what we want.\r\n\r\nBut regarding not introducing breaking changes, wouldn't this maybe also break some setups for users who don't expect this new behavior?\r\n", "@albertz I would suggest to move the discussion on implementation details on our side to the following issue: rwth-i6/i6_core/issues/257", "I like the idea of adding `embed_external_files` and set it to True by default to `save_to_disk`.\r\nIt's indeed a kind of breaking change since some users will get bigger Arrow files when updating the lib, but the advantages are nice:\r\n1. I like the idea of having it self contained, in case you want to delete your cache\r\n2. users also upload these Arrow files to cloud storage via the `fs` parameter, and in this case they would expect to upload a self-contained dataset\r\n3. consistency with `push_to_hub`\r\n\r\nIf it sounds good to you I'll open an issue to discuss this and track the advancements", "Closed #4179." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4184/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4184
https://github.com/huggingface/datasets/pull/4184
true
1,208,449,335
https://api.github.com/repos/huggingface/datasets/issues/4183/labels{/name}
Added an example of how to load one config or the other
2023-09-24T10:02:24Z
4,183
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-19T14:26:59Z
https://api.github.com/repos/huggingface/datasets/issues/4183/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4183/timeline
Document librispeech configs
https://api.github.com/repos/huggingface/datasets/issues/4183/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-04-19T15:15:20Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4183.diff", "html_url": "https://github.com/huggingface/datasets/pull/4183", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4183.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4183" }
PR_kwDODunzps42bjXn
[ "I think the main purpose of #4179 was how to be able to load both configs into one, so should we maybe add this part of the code: https://github.com/huggingface/datasets/issues/4179#issuecomment-1102383717 \r\n\r\nto the doc? \r\n\r\nActually @lhoestq would this work given that they have different split names: https://huggingface.co/datasets/librispeech_asr#data-splits ? ", "This doc extension does not explain why I can't simply load the whole dataset. Or what workaround I need to get the whole dataset, which is what people usually want for Librispeech.", "_The documentation is not available anymore as the PR was closed or merged._", "@lhoestq, I can add a `\"all\"` config to Librispeech have the datasets already cached somewhere ", "I'm closing this PR then, feel free to continue the discussion in https://github.com/huggingface/datasets/issues/4179\r\n" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4183/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4183
https://github.com/huggingface/datasets/pull/4183
true
1,208,285,235
https://api.github.com/repos/huggingface/datasets/issues/4182/labels{/name}
## Describe the bug Source download_url from zenodo.org does not respond. `_DOWNLOAD_URL = "https://zenodo.org/record/2787612/files/SICK.zip?download=1"` Other datasets also use zenodo.org to store data and they cannot be downloaded as well. It would be better to actually use more reliable way to store original data like s3 bucket. ## Steps to reproduce the bug ```python load_dataset("sick") ``` ## Expected results Dataset should be downloaded. ## Actual results ConnectionError: Couldn't reach https://zenodo.org/record/2787612/files/SICK.zip?download=1 (ReadTimeout(ReadTimeoutError("HTTPSConnectionPool(host='zenodo.org', port=443): Read timed out. (read timeout=100)"))) ## Environment info - `datasets` version: 2.1.0 - Platform: Darwin-21.4.0-x86_64-i386-64bit - Python version: 3.7.11 - PyArrow version: 7.0.0 - Pandas version: 1.3.5
2022-04-20T07:11:05Z
4,182
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-19T12:26:57Z
https://api.github.com/repos/huggingface/datasets/issues/4182/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4182/timeline
Zenodo.org download is not responding
https://api.github.com/repos/huggingface/datasets/issues/4182/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/32985207?v=4", "events_url": "https://api.github.com/users/dkajtoch/events{/privacy}", "followers_url": "https://api.github.com/users/dkajtoch/followers", "following_url": "https://api.github.com/users/dkajtoch/following{/other_user}", "gists_url": "https://api.github.com/users/dkajtoch/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dkajtoch", "id": 32985207, "login": "dkajtoch", "node_id": "MDQ6VXNlcjMyOTg1MjA3", "organizations_url": "https://api.github.com/users/dkajtoch/orgs", "received_events_url": "https://api.github.com/users/dkajtoch/received_events", "repos_url": "https://api.github.com/users/dkajtoch/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dkajtoch/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dkajtoch/subscriptions", "type": "User", "url": "https://api.github.com/users/dkajtoch" }
[]
null
completed
CONTRIBUTOR
2022-04-20T07:11:05Z
null
I_kwDODunzps5IBPgz
[ "[Off topic but related: Is the uptime of S3 provably better than Zenodo's?]", "Hi @dkajtoch, please note that at HuggingFace we are not hosting this dataset: we are just using a script to download their data file and create a dataset from it.\r\n\r\nIt was the dataset owners decision to host their data at Zenodo. You can see this on their website: https://marcobaroni.org/composes/sick.html\r\n\r\nAnd yes, you are right: Zenodo is currently having some incidents and people are reporting problems from it.\r\n\r\nOn the other hand, we could contact the data owners and propose them to host their data at our Hugging Face Hub.\r\n\r\n@julien-c I guess so.\r\n", "Thanks @albertvillanova. I know that the problem lies in the source data. I just wanted to point out that these kind of problems are unavoidable without having one place where data sources are cached. Websites may go down or data sources may move. Having a copy in Hugging Face Hub would be a great solution. ", "Definitely, @dkajtoch! But we have to ask permission to the data owners. And many dataset licenses directly forbid data redistribution: in those cases we are not allowed to host their data on our Hub.", "Ahhh good point! License is the problem :(" ]
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4182/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4182
https://github.com/huggingface/datasets/issues/4182
false
1,208,194,805
https://api.github.com/repos/huggingface/datasets/issues/4181/labels{/name}
## Dataset viewer issue for '*name of the dataset*' https://huggingface.co/datasets/google/fleurs ``` Status code: 400 Exception: NotImplementedError Message: Extraction protocol for TAR archives like 'https://storage.googleapis.com/xtreme_translations/FLEURS/af_za.tar.gz' is not implemented in streaming mode. Please use `dl_manager.iter_archive` instead. ``` Am I the one who added this dataset ? Yes Can I fix this somehow in the script? @lhoestq @severo
2022-07-25T11:44:02Z
4,181
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
2022-04-19T11:09:56Z
https://api.github.com/repos/huggingface/datasets/issues/4181/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4181/timeline
Support streaming FLEURS dataset
https://api.github.com/repos/huggingface/datasets/issues/4181/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[]
null
completed
CONTRIBUTOR
2022-07-25T11:44:02Z
null
I_kwDODunzps5IA5b1
[ "Yes, you just have to use `dl_manager.iter_archive` instead of `dl_manager.download_and_extract`.\r\n\r\nThat's because `download_and_extract` doesn't support TAR archives in streaming mode.", "Tried to make it streamable, but I don't think it's really possible. @lhoestq @polinaeterna maybe you guys can check: \r\nhttps://huggingface.co/datasets/google/fleurs/commit/dcf80160cd77977490a8d32b370c027107f2407b \r\n\r\nreal quick. \r\n\r\nI think the problem is that we cannot ensure that the metadata file is found before the audio. Or is this possible somehow @lhoestq ? ", "@patrickvonplaten I think the metadata file should be found first because the audio files are contained in a folder next to the metadata files (just as in common voice), so the metadata files should be \"on top of the list\" as they are closer to the root in the directories hierarchy ", "@patrickvonplaten but apparently it doesn't... I don't really know why.", "Yeah! Any ideas what could be the reason here? cc @lhoestq ?", "The order of the files is determined when the TAR archive is created, depending on the commands the creator ran.\r\nIf the metadata file is not at the beginning of the file, that makes streaming completely inefficient. In this case the TAR archive needs to be recreated in an appropriate order.", "Actually we could maybe just host the metadata file ourselves and then stream the audio data only. Don't think that this would be a problem for the FLEURS authors (I can ask them :-)) ", "I made a PR to their repo to support streaming (by uploading the metadata file to the Hub). See:\r\n- https://huggingface.co/datasets/google/fleurs/discussions/4", "I'm closing this issue as the PR above has been merged." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4181/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4181
https://github.com/huggingface/datasets/issues/4181
false
1,208,042,320
https://api.github.com/repos/huggingface/datasets/issues/4180/labels{/name}
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Currently, `dataset["audio"]` will load EVERY element in the dataset in RAM, which can be quite big for an audio dataset. Having an iterator (or sequence) type of object, would make inference with `transformers` 's `pipeline` easier to use and not so memory hungry. **Describe the solution you'd like** A clear and concise description of what you want to happen. For a non breaking change: ```python for audio in dataset.iterate("audio"): # {"array": np.array(...), "sampling_rate":...} ``` For a breaking change solution (not necessary), changing the type of `dataset["audio"]` to a sequence type so that ```python pipe = pipeline(model="...") for out in pipe(dataset["audio"]): # {"text":....} ``` could work **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. ```python def iterate(dataset, key): for item in dataset: yield dataset[key] for out in pipeline(iterate(dataset, "audio")): # {"array": ...} ``` This works but requires the helper function which feels slightly clunky. **Additional context** Add any other context about the feature request here. The context is actually to showcase better integration between `pipeline` and `datasets` in the Quicktour demo: https://github.com/huggingface/transformers/pull/16723/files @lhoestq
2022-04-21T10:30:58Z
4,180
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-04-19T09:15:45Z
https://api.github.com/repos/huggingface/datasets/issues/4180/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4180/timeline
Add some iteration method on a dataset column (specific for inference)
https://api.github.com/repos/huggingface/datasets/issues/4180/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "events_url": "https://api.github.com/users/Narsil/events{/privacy}", "followers_url": "https://api.github.com/users/Narsil/followers", "following_url": "https://api.github.com/users/Narsil/following{/other_user}", "gists_url": "https://api.github.com/users/Narsil/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Narsil", "id": 204321, "login": "Narsil", "node_id": "MDQ6VXNlcjIwNDMyMQ==", "organizations_url": "https://api.github.com/users/Narsil/orgs", "received_events_url": "https://api.github.com/users/Narsil/received_events", "repos_url": "https://api.github.com/users/Narsil/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Narsil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Narsil/subscriptions", "type": "User", "url": "https://api.github.com/users/Narsil" }
[]
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5IAUNQ
[ "Thanks for the suggestion ! I agree it would be nice to have something directly in `datasets` to do something as simple as that\r\n\r\ncc @albertvillanova @mariosasko @polinaeterna What do you think if we have something similar to pandas `Series` that wouldn't bring everything in memory when doing `dataset[\"audio\"]` ? Currently it returns a list with all the decoded audio data in memory.\r\n\r\nIt would be a breaking change though, since `isinstance(dataset[\"audio\"], list)` wouldn't work anymore, but we could implement a `Sequence` so that `dataset[\"audio\"][0]` still works and only loads one item in memory.\r\n\r\nYour alternative suggestion with `iterate` is also sensible, though maybe less satisfactory in terms of experience IMO", "I agree that current behavior (decoding all audio file sin the dataset when accessing `dataset[\"audio\"]`) is not useful, IMHO. Indeed in our docs, we are constantly warning our collaborators not to do that.\r\n\r\nTherefore I upvote for a \"useful\" behavior of `dataset[\"audio\"]`. I don't think the breaking change is important in this case, as I guess no many people use it with its current behavior. Therefore, for me it seems reasonable to return a generator (instead of an in-memeory list) for \"special\" features, like Audio/Image.\r\n\r\n@lhoestq on the other hand I don't understand your proposal about Pandas-like... ", "I recall I had the same idea while working on the `Image` feature, so I agree implementing something similar to `pd.Series` that lazily brings elements in memory would be beneficial.", "@lhoestq @mariosasko Could you please give a link to that new feature of `pandas.Series`? As far as I remember since I worked with pandas for more than 6 years, there was no lazy in-memory feature; it was everything in-memory; that was the reason why other frameworks were created, like Vaex or Dask, e.g. ", "Yea pandas doesn't do lazy loading. I was referring to pandas.Series to say that they have a dedicated class to represent a column ;)" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4180/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4180
https://github.com/huggingface/datasets/issues/4180
false
1,208,001,118
https://api.github.com/repos/huggingface/datasets/issues/4179/labels{/name}
## Describe the bug The dataset librispeech_asr (standard Librispeech) fails to load. ## Steps to reproduce the bug ```python datasets.load_dataset("librispeech_asr") ``` ## Expected results It should download and prepare the whole dataset (all subsets). In [the doc](https://huggingface.co/datasets/librispeech_asr), it says it has two configurations (clean and other). However, the dataset doc says that not specifying `split` should just load the whole dataset, which is what I want. Also, in case of this specific dataset, this is also the standard what the community uses. When you look at any publications with results on Librispeech, they always use the whole train dataset for training. ## Actual results ``` ... File "/home/az/.cache/huggingface/modules/datasets_modules/datasets/librispeech_asr/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/librispeech_asr.py", line 119, in LibrispeechASR._split_generators line: archive_path = dl_manager.download(_DL_URLS[self.config.name]) locals: archive_path = <not found> dl_manager = <local> <datasets.utils.download_manager.DownloadManager object at 0x7fc07b426160> dl_manager.download = <local> <bound method DownloadManager.download of <datasets.utils.download_manager.DownloadManager object at 0x7fc07b426160>> _DL_URLS = <global> {'clean': {'dev': 'http://www.openslr.org/resources/12/dev-clean.tar.gz', 'test': 'http://www.openslr.org/resources/12/test-clean.tar.gz', 'train.100': 'http://www.openslr.org/resources/12/train-clean-100.tar.gz', 'train.360': 'http://www.openslr.org/resources/12/train-clean-360.tar.gz'}, 'other'... self = <local> <datasets_modules.datasets.librispeech_asr.1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c.librispeech_asr.LibrispeechASR object at 0x7fc12a633310> self.config = <local> BuilderConfig(name='default', version=0.0.0, data_dir='/home/az/i6/setups/2022-03-20--sis/work/i6_core/datasets/huggingface/DownloadAndPrepareHuggingFaceDatasetJob.TV6Nwm6dFReF/output/data_dir', data_files=None, description=None) self.config.name = <local> 'default', len = 7 KeyError: 'default' ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.1.0 - Platform: Linux-5.4.0-107-generic-x86_64-with-glibc2.31 - Python version: 3.9.9 - PyArrow version: 6.0.1 - Pandas version: 1.4.2
2022-07-27T16:10:00Z
4,179
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-19T08:45:48Z
https://api.github.com/repos/huggingface/datasets/issues/4179/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4179/timeline
Dataset librispeech_asr fails to load
https://api.github.com/repos/huggingface/datasets/issues/4179/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59132?v=4", "events_url": "https://api.github.com/users/albertz/events{/privacy}", "followers_url": "https://api.github.com/users/albertz/followers", "following_url": "https://api.github.com/users/albertz/following{/other_user}", "gists_url": "https://api.github.com/users/albertz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertz", "id": 59132, "login": "albertz", "node_id": "MDQ6VXNlcjU5MTMy", "organizations_url": "https://api.github.com/users/albertz/orgs", "received_events_url": "https://api.github.com/users/albertz/received_events", "repos_url": "https://api.github.com/users/albertz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertz/subscriptions", "type": "User", "url": "https://api.github.com/users/albertz" }
[]
null
completed
NONE
2022-07-27T16:10:00Z
null
I_kwDODunzps5IAKJe
[ "@patrickvonplaten Hi! I saw that you prepared this? :)", "Another thing, but maybe this should be a separate issue: As I see from the code, it would try to use up to 16 simultaneous downloads? This is problematic for Librispeech or anything on OpenSLR. On [the homepage](https://www.openslr.org/), it says:\r\n\r\n> If you want to download things from this site, please download them one at a time, and please don't use any fancy software-- just download things from your browser or use 'wget'. We have a firewall rule to drop connections from hosts with more than 5 simultaneous connections, and certain types of download software may activate this rule.\r\n\r\nRelated: https://github.com/tensorflow/datasets/issues/3885", "Hey @albertz,\r\n\r\nNice to see you here! It's been a while ;-) ", "Sorry maybe the docs haven't been super clear here. By `split` we mean one of `train.500`, `train.360`, `train.100`, `validation`, `test`. For Librispeech, you'll have to specific a config (either `other` or `clean`) though:\r\n\r\n```py\r\ndatasets.load_dataset(\"librispeech_asr\", \"clean\")\r\n```\r\n\r\nshould work and give you all splits (being \"train\", \"test\", ...) for the clean config of the dataset.\r\n", "If you need both `\"clean\"` and `\"other\"` I think you'll have to do concatenate them as follows: \r\n\r\n```py\r\nfrom datasets import concatenate_datasets, load_dataset\r\n\r\nother = load_dataset(\"librispeech_asr\", \"other\")\r\nclean = load_dataset(\"librispeech_asr\", \"clean\")\r\n\r\nlibrispeech = concatenate_datasets([other, clean])\r\n```\r\n\r\nSee https://huggingface.co/docs/datasets/v2.1.0/en/process#concatenate", "Downloading one split would be:\r\n\r\n```py\r\nfrom datasets import load_dataset\r\n\r\nother = load_dataset(\"librispeech_asr\", \"other\", split=\"train.500\")\r\n```\r\n\r\n\r\n", "cc @lhoestq FYI maybe the docs can be improved here", "Ah thanks. But wouldn't it be easier/nicer (and more canonical) to just make it in a way that simply `load_dataset(\"librispeech_asr\")` works?", "Pinging @lhoestq here, think this could make sense! Not sure however how the dictionary would then look like", "Would it make sense to have `clean` as the default config ?\r\n\r\nAlso I think `load_dataset(\"librispeech_asr\")` should have raised you an error that says that you need to specify a config\r\n\r\nI also opened a PR to improve the doc: https://github.com/huggingface/datasets/pull/4183", "> Would it make sense to have `clean` as the default config ?\r\n\r\nI think a user would expect that the default would give you the full dataset.\r\n\r\n> Also I think `load_dataset(\"librispeech_asr\")` should have raised you an error that says that you need to specify a config\r\n\r\nIt does raise an error, but this error confused me because I did not understand why I needed a config, or why I could not simply download the whole dataset, which is what people usually do with Librispeech.\r\n", "+1 for @albertz. Also think lots of people download the whole dataset (`\"clean\"` + `\"other\"`) for Librispeech.\r\n\r\nThink there are also some people though who:\r\n- a) Don't have the memory to store the whole dataset\r\n- b) Just want to evaluate on one of the two configs", "Ok ! Adding the \"all\" configuration would do the job then, thanks ! In the \"all\" configuration we can merge all the train.xxx splits into one \"train\" split, or keep them separate depending on what's the most practical to use (probably put everything in \"train\" no ?)", "I'm not too familiar with how to work with HuggingFace datasets, but people often do some curriculum learning scheme, where they start with train.100, later go over to train.100 + train.360, and then later use the whole train (960h). It would be good if this is easily possible.\r\n", "Hey @albertz, \r\n\r\nopened a PR here. Think by adding the \"subdataset\" class to each split \"train\", \"dev\", \"other\" as shown here: https://github.com/huggingface/datasets/pull/4184/files#r853272727 it should be easily possible (e.g. with the filter function https://huggingface.co/docs/datasets/v2.1.0/en/package_reference/main_classes#datasets.Dataset.filter )", "But also since everything is cached one could also just do:\r\n\r\n```python\r\nload_dataset(\"librispeech\", \"clean\", \"train.100\")\r\nload_dataset(\"librispeech\", \"clean\", \"train.100+train.360\")\r\nload_dataset(\"librispeech\" \"all\", \"train\") \r\n```", "Hi @patrickvonplaten ,\r\n\r\nload_dataset(\"librispeech_asr\", \"clean\", \"train.100\") actually downloads the whole dataset and not the 100 hr split, is this a bug?", "Hmm, I don't really see how that's possible: https://github.com/huggingface/datasets/blob/d22e39a0693d4be7410cf9a5d41fd5aac22be3cc/datasets/librispeech_asr/librispeech_asr.py#L51\r\n\r\nNote that all datasets related to `\"clean\"` are downloaded, but only `\"train.100\"` should be used. \r\n\r\ncc @lhoestq @albertvillanova @mariosasko can we do anything against download dataset links that are not related to the \"split\" that one actually needs. E.g. why should the split `\"train.360\"` be downloaded if for the user executes the above command:\r\n\r\n```py\r\nload_dataset(\"librispeech_asr\", \"clean\", \"train.100\")\r\n```", "@patrickvonplaten This problem is a bit harder than it may seem, and it has to do with how our scripts are structured - `_split_generators` downloads data for a split before its definition. There was an attempt to fix this in https://github.com/huggingface/datasets/pull/2249, but it wasn't flexible enough. Luckily, I have a plan of attack, and this issue is on our short-term roadmap, so I'll work on it soon.\r\n\r\nIn the meantime, one can use streaming or manually download a dataset script, remove unwanted splits and load a dataset via `load_dataset`.", "> load_dataset(\"librispeech_asr\", \"clean\", \"train.100\") actually downloads the whole dataset and not the 100 hr split, is this a bug?\r\n\r\nSince this bug is still there and google led me here when I was searching for a solution, I am writing down how to quickly fix it (as suggested by @mariosasko) for whoever else is not familiar with how the HF Hub works.\r\n\r\nDownload the [librispeech_asr.py](https://huggingface.co/datasets/librispeech_asr/blob/main/librispeech_asr.py) script and remove the unwanted splits both from the [`_DL_URLS` dictionary](https://huggingface.co/datasets/librispeech_asr/blob/main/librispeech_asr.py#L47-L68) and from the [`_split_generators` function](https://huggingface.co/datasets/librispeech_asr/blob/main/librispeech_asr.py#L121-L241).\r\n[Here ](https://huggingface.co/datasets/andreagasparini/librispeech_test_only) I made an example with only the test sets.\r\n\r\nThen either save the script locally and load the dataset via \r\n```python\r\nload_dataset(\"${local_path}/librispeech_asr.py\")\r\n```\r\n\r\nor [create a new dataset repo on the hub](https://huggingface.co/new-dataset) named \"librispeech_asr\" and upload the script there, then you can just run\r\n```python\r\nload_dataset(\"${hugging_face_username}/librispeech_asr\")\r\n```", "Fixed by https://github.com/huggingface/datasets/pull/4184" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4179/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4179
https://github.com/huggingface/datasets/issues/4179
false
1,207,787,073
https://api.github.com/repos/huggingface/datasets/issues/4178/labels{/name}
To use the dataset download the tar file [imagenet_object_localization_patched2019.tar.gz](https://www.kaggle.com/competitions/imagenet-object-localization-challenge/data?select=imagenet_object_localization_patched2019.tar.gz) from Kaggle and then point the datasets library to it by using: ```py from datasets import load_dataset dataset = load_dataset("imagenet", data_dir="/path/to/imagenet_object_localization_patched2019.tar.gz") ``` Currently train and validation splits are supported.
2022-04-29T21:43:59Z
4,178
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-19T06:01:35Z
https://api.github.com/repos/huggingface/datasets/issues/4178/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4178/timeline
[feat] Add ImageNet dataset
https://api.github.com/repos/huggingface/datasets/issues/4178/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3616806?v=4", "events_url": "https://api.github.com/users/apsdehal/events{/privacy}", "followers_url": "https://api.github.com/users/apsdehal/followers", "following_url": "https://api.github.com/users/apsdehal/following{/other_user}", "gists_url": "https://api.github.com/users/apsdehal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/apsdehal", "id": 3616806, "login": "apsdehal", "node_id": "MDQ6VXNlcjM2MTY4MDY=", "organizations_url": "https://api.github.com/users/apsdehal/orgs", "received_events_url": "https://api.github.com/users/apsdehal/received_events", "repos_url": "https://api.github.com/users/apsdehal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/apsdehal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/apsdehal/subscriptions", "type": "User", "url": "https://api.github.com/users/apsdehal" }
[]
null
null
CONTRIBUTOR
2022-04-29T21:37:08Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4178.diff", "html_url": "https://github.com/huggingface/datasets/pull/4178", "merged_at": "2022-04-29T21:37:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/4178.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4178" }
PR_kwDODunzps42ZfFN
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks for the comments. I believe I have addressed all of them and also decreased the size of the dummy data file, so it should be ready for a re-review. I also made a change to allow adding synset mapping and valprep script in config in case we add ImageNet 21k some time later. ", "@lhoestq I have updated the PR to address all of the review comments." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4178/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4178
https://github.com/huggingface/datasets/pull/4178
true
1,207,535,920
https://api.github.com/repos/huggingface/datasets/issues/4177/labels{/name}
This dataset for the [1st task of SemEval-2018](https://competitions.codalab.org/competitions/17751) competition was missing all subtasks except for subtask 5. I added another two subtasks (subtask 1 and 2), which are each comprised of 12 additional data subsets: for each language in En, Es, Ar, there are 4 datasets, broken down by emotions (anger, fear, joy, sadness). ## Remaining questions I wasn't able to find any documentation about how one should make PRs to modify datasets. Because of that, I just did my best to integrate the new data into the code, and tested locally that this worked. I'm sorry if I'm not respecting your contributing guidelines – if they are documented somewhere, I'd appreciate if you could send a pointer! Not sure how `dataset_infos.json` and `dummy` should be updated. My understanding is that they were automatically generated at the time of the original dataset creation?
2022-10-05T10:38:16Z
4,177
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-04-18T22:59:30Z
https://api.github.com/repos/huggingface/datasets/issues/4177/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4177/timeline
Adding missing subsets to the `SemEval-2018 Task 1` dataset
https://api.github.com/repos/huggingface/datasets/issues/4177/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11460267?v=4", "events_url": "https://api.github.com/users/micahcarroll/events{/privacy}", "followers_url": "https://api.github.com/users/micahcarroll/followers", "following_url": "https://api.github.com/users/micahcarroll/following{/other_user}", "gists_url": "https://api.github.com/users/micahcarroll/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/micahcarroll", "id": 11460267, "login": "micahcarroll", "node_id": "MDQ6VXNlcjExNDYwMjY3", "organizations_url": "https://api.github.com/users/micahcarroll/orgs", "received_events_url": "https://api.github.com/users/micahcarroll/received_events", "repos_url": "https://api.github.com/users/micahcarroll/repos", "site_admin": false, "starred_url": "https://api.github.com/users/micahcarroll/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/micahcarroll/subscriptions", "type": "User", "url": "https://api.github.com/users/micahcarroll" }
[]
null
null
NONE
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4177.diff", "html_url": "https://github.com/huggingface/datasets/pull/4177", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4177.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4177" }
PR_kwDODunzps42Yxca
[ "Datasets are not tracked in this repository anymore. You should move this PR to the [discussions page of this dataset](https://huggingface.co/datasets/sem_eval_2018_task_1/discussions)" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4177/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4177
https://github.com/huggingface/datasets/pull/4177
true
1,206,515,563
https://api.github.com/repos/huggingface/datasets/issues/4176/labels{/name}
Hello, in the processing stage, I use two operations. The first one : map + filter, is very fast and it uses the full cores, while the socond step is very slow and did not use full cores. Also, there is a significant lag between them. Am I missing something ? ``` raw_datasets = raw_datasets.map(split_func, batched=False, num_proc=args.preprocessing_num_workers, load_from_cache_file=not args.overwrite_cache, desc = "running split para ==>")\ .filter(lambda example: example['text1']!='' and example['text2']!='', num_proc=args.preprocessing_num_workers, desc="filtering ==>") processed_datasets = raw_datasets.map( preprocess_function, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc="Running tokenizer on dataset===>", ) ```
2022-04-18T00:03:00Z
4,176
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-04-17T23:52:29Z
https://api.github.com/repos/huggingface/datasets/issues/4176/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4176/timeline
Very slow between two operations
https://api.github.com/repos/huggingface/datasets/issues/4176/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/26405281?v=4", "events_url": "https://api.github.com/users/yananchen1989/events{/privacy}", "followers_url": "https://api.github.com/users/yananchen1989/followers", "following_url": "https://api.github.com/users/yananchen1989/following{/other_user}", "gists_url": "https://api.github.com/users/yananchen1989/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yananchen1989", "id": 26405281, "login": "yananchen1989", "node_id": "MDQ6VXNlcjI2NDA1Mjgx", "organizations_url": "https://api.github.com/users/yananchen1989/orgs", "received_events_url": "https://api.github.com/users/yananchen1989/received_events", "repos_url": "https://api.github.com/users/yananchen1989/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yananchen1989/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yananchen1989/subscriptions", "type": "User", "url": "https://api.github.com/users/yananchen1989" }
[]
null
completed
NONE
2022-04-18T00:03:00Z
null
I_kwDODunzps5H6fdr
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4176/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4176
https://github.com/huggingface/datasets/issues/4176
false
1,205,589,842
https://api.github.com/repos/huggingface/datasets/issues/4175/labels{/name}
closes #2981 #2810 @nateraw @hassiahk I've listed you guys as co-author as you've contributed previously to this dataset
2023-09-24T10:02:38Z
4,175
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-15T13:42:32Z
https://api.github.com/repos/huggingface/datasets/issues/4175/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4175/timeline
Add WIT Dataset
https://api.github.com/repos/huggingface/datasets/issues/4175/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
[]
null
null
CONTRIBUTOR
2022-05-02T14:26:41Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4175.diff", "html_url": "https://github.com/huggingface/datasets/pull/4175", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4175.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4175" }
PR_kwDODunzps42SqF-
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hi! Coming in late with some context.\r\n\r\nThere are two versions of the WIT dataset:\r\n1. The original source dataset managed by Wikimedia. It has more information, raw image representations, and each row corresponds to an image linked to all of its captions wherever it happens in Wikipedia (in multiple languages)\r\n2. The Google version, corresponding to the data script in this PR, which duplicates image instances and requires the user to download the images themselves from the provided URL (note that a basic implementation will have them download the same picture several time. @thomasw21 using our download manager instead of `urllib` could help with that, but it wouldn't be required if people had access to the first version)\r\n\r\nThe Wikimedia folks were really interested in us hosting a ready-to-go streaming version of this dataset where users don't have to download the version themselves, which is why we have the pre-processed versions on an HF bucket, with the raw images and a pre-computed embedding (don't remember the model, we can keep it ). That's the data script currently in https://github.com/huggingface/datasets/pull/2981 . It's nearly ready to go, the one thing we should do is move the raw data from our HF google Cloud bucket to the Hub.\r\n\r\nHow do you want to move forward? IMO the best way would be to have a WIT dataset under the Wikimedia org with both configurations, but it depends on everyone's timelines", "Okay after offline discussion. We'll improve this versions and push it to the hub under `google` namespace. \r\n\r\n> which duplicates image instances and requires the user to download the images themselves from the provided URL (note that a basic implementation will have them download the same picture several time. @thomasw21 using our download manager instead of urllib could help with that, but it wouldn't be required if people had access to the first version)\r\n\r\nAh interesting wasn't aware of this duplication issue, concretely it'll just mean that our dataset in bigger than expected ... I think this should be handled after this loading script (though I have to figure our how to spawn a dl_manager).\r\n\r\n> The Wikimedia folks were really interested in us hosting a ready-to-go streaming version of this dataset where users don't have to download the version themselves, which is why we have the pre-processed versions on an HF bucket, with the raw images and a pre-computed embedding (don't remember the model, we can keep it ). That's the data script currently in https://github.com/huggingface/datasets/pull/2981 . It's nearly ready to go, the one thing we should do is move the raw data from our HF google Cloud bucket to the Hub.\r\n\r\nSimilarly a script will be written and pushed to `wikimedia` organisation.", "@mariosasko can you make one last review concerning the text description changes? Then I'll handle putting it under `google` namespace and close this PR.", "Looks all good now. Great job! ", "Closing as this has been migrated to the hub under `google` namespace: https://huggingface.co/datasets/google/wit" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4175/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4175
https://github.com/huggingface/datasets/pull/4175
true
1,205,575,941
https://api.github.com/repos/huggingface/datasets/issues/4174/labels{/name}
When `function` modifies input in-place, the guarantee that columns in `remove_columns` are contained in `input` doesn't hold true anymore. Therefore we need to relax way we pop elements by checking if that column exists.
2022-04-15T14:52:07Z
4,174
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-15T13:23:15Z
https://api.github.com/repos/huggingface/datasets/issues/4174/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4174/timeline
Fix when map function modifies input in-place
https://api.github.com/repos/huggingface/datasets/issues/4174/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
[]
null
null
CONTRIBUTOR
2022-04-15T14:45:58Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4174.diff", "html_url": "https://github.com/huggingface/datasets/pull/4174", "merged_at": "2022-04-15T14:45:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/4174.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4174" }
PR_kwDODunzps42SnJS
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4174/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4174
https://github.com/huggingface/datasets/pull/4174
true
1,204,657,114
https://api.github.com/repos/huggingface/datasets/issues/4173/labels{/name}
As mentioned in https://github.com/huggingface/datasets/issues/4139 it's currently not possible to stream private/gated zipped images from the Hub. This is because `Image.decode_example` does not handle authentication. Indeed decoding requires to access and download the file from the private repository. In this PR I added authentication to `Image.decode_example` via a `token_per_repo_id` optional argument. I first wanted to just pass `use_auth_token` but a single `Image` instance can be responsible of decoding images from a combination of several datasets together (from `interleave_datasets` for example). Therefore I just used a dictionary `repo_id` -> `token` instead. I'm getting the `repo_id` from the dataset builder (I replaced the `namepace` attribute with `repo_id`) I did the same for `Audio.decode_example` cc @SBrandeis @severo
2022-05-05T14:05:54Z
4,173
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-14T15:15:07Z
https://api.github.com/repos/huggingface/datasets/issues/4173/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4173/timeline
Stream private zipped images
https://api.github.com/repos/huggingface/datasets/issues/4173/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-05-05T13:58:35Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4173.diff", "html_url": "https://github.com/huggingface/datasets/pull/4173", "merged_at": "2022-05-05T13:58:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/4173.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4173" }
PR_kwDODunzps42Ppnd
[ "_The documentation is not available anymore as the PR was closed or merged._", "oops looks like some tests are failing sorry, will fix them tomorrow\r\n\r\nEDIT: not today but asap hopefully", "cc @mariosasko this is ready for review, let me know what you think !" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4173/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4173
https://github.com/huggingface/datasets/pull/4173
true
1,204,433,160
https://api.github.com/repos/huggingface/datasets/issues/4172/labels{/name}
Following comments in https://github.com/huggingface/datasets/issues/4003 we found that it was outdated and casing an error when loading the dataset
2022-04-15T14:47:42Z
4,172
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-14T11:53:06Z
https://api.github.com/repos/huggingface/datasets/issues/4172/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4172/timeline
Update assin2 dataset_infos.json
https://api.github.com/repos/huggingface/datasets/issues/4172/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-04-15T14:41:22Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4172.diff", "html_url": "https://github.com/huggingface/datasets/pull/4172", "merged_at": "2022-04-15T14:41:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/4172.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4172" }
PR_kwDODunzps42O7LW
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4172/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4172
https://github.com/huggingface/datasets/pull/4172
true
1,204,413,620
https://api.github.com/repos/huggingface/datasets/issues/4170/labels{/name}
This PR rewrites almost all of `to_tf_dataset()`, which makes it kind of hard to list all the changes, but the most critical ones are: - Much better stability and no more dropping unexpected column names (Sorry @NielsRogge) - Doesn't clobber custom transforms on the data (Sorry @NielsRogge again) - Much better handling of the situation when the `collate_fn` adds columns that aren't in the dataset. - Better inference of shapes and data types - Lots of hacky special-casing code removed - Can return string columns (as `tf.String`) - Most arguments have default values, calling the method should be much simpler - ~~Can accept a `model` argument and only return columns that are valid inputs to that model~~ - Drops the `dummy_labels` argument - this was a workaround for Keras issues that have been resolved by changes in `transformers`. Also remove it from tests and the Overview notebook. I still have a couple of TODOs remaining and some testing to do, so don't merge yet, but it should be mostly ready for review at this point!
2022-06-06T14:31:12Z
4,170
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-14T11:30:58Z
https://api.github.com/repos/huggingface/datasets/issues/4170/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4170/timeline
to_tf_dataset rewrite
https://api.github.com/repos/huggingface/datasets/issues/4170/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4", "events_url": "https://api.github.com/users/Rocketknight1/events{/privacy}", "followers_url": "https://api.github.com/users/Rocketknight1/followers", "following_url": "https://api.github.com/users/Rocketknight1/following{/other_user}", "gists_url": "https://api.github.com/users/Rocketknight1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rocketknight1", "id": 12866554, "login": "Rocketknight1", "node_id": "MDQ6VXNlcjEyODY2NTU0", "organizations_url": "https://api.github.com/users/Rocketknight1/orgs", "received_events_url": "https://api.github.com/users/Rocketknight1/received_events", "repos_url": "https://api.github.com/users/Rocketknight1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rocketknight1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rocketknight1/subscriptions", "type": "User", "url": "https://api.github.com/users/Rocketknight1" }
[]
null
null
MEMBER
2022-06-06T14:22:09Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4170.diff", "html_url": "https://github.com/huggingface/datasets/pull/4170", "merged_at": "2022-06-06T14:22:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/4170.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4170" }
PR_kwDODunzps42O2-L
[ "_The documentation is not available anymore as the PR was closed or merged._", "[Magic is now banned](https://www.youtube.com/watch?v=WIn58XoY728#t=36s) by decree of @sgugger. This is honestly much cleaner, and the functionality will make much more sense in `transformers` anyway!", "@gante I renamed the default collator to `minimal_tf_collate_fn`!", "@lhoestq @sgugger @gante \r\n\r\nI think this should now be ready, it looks good in testing! I'll try a few more notebooks today and tomorrow to be sure before I merge. Key changes are:\r\n\r\n- No column autodetection magic (will make a separate PR to add this as a `transformers` function)\r\n- Drops non-numerical features automatically (this is more of a 'DataLoader' method, we'll have a separate method to expose 'raw' datasets to `tf.data`)\r\n- Better autodetection of numerical features.\r\n- Shouldn't randomly crash mid-function :skull: \r\n\r\nWe definitely have some questions still to resolve about how to handle making a 'DataLoader' dataset versus a 'raw' dataset - see [the Notion doc](https://www.notion.so/huggingface2/Splitting-to_tf_dataset-c2e0773c4bec484384064b30ed634383) if you're interested. Still, since this PR is just fixes/improvements to an existing method which never supported non-numerical features anyway, we can merge it before we've resolved those issues, and then think about how to name and split things afterwards.", "P.S. I'll take out the region comments at the end before I merge, I promise! They're just helpful while I'm editing it", "+1 for the tests\r\n\r\n> Drops non-numerical features automatically\r\n\r\nCan you give more details on how this work and the rationale as well ? This is not explained in the docs\r\n\r\nAlso why are you adding `error_on_missing` and `auto_fix_label_names ` ? The rationale is not clear to me. In particular I think it is sensible enough to expect users to not ask columns that don't exist, and to rename a label column when required.", "@lhoestq I rewrote those parts - they were causing some other issues too! `error_on_missing` and `auto_fix_label_names` have been removed. The new logic is to simply drop (before batch collation) all columns the user doesn't ask for, but not to raise errors if the user asked for columns not in the dataset, as they may be added by the collator. Hopefully this cleans it up and matches the documentation better!", "@lhoestq New tests are now in!", "Seeing some other random tests failing that don't look to be associated with this PR.", "@lhoestq I can't figure out these test failures! They don't seem related to this PR at all, but I rebased to the latest version and they keep happening, even though they're not visible on master.", "Thanks for the ping, will take a look tomorrow :)\r\n\r\nMaybe the rebase didn't go well for the code recently merged about label alignment from https://github.com/huggingface/datasets/pull/4277 ?", "It's very strange! The rebase looks fine to me. I might try to move my changes to a new branch from `master` and see if I can figure out which change causes this problem to appear.", "@lhoestq Got it! It was caused by a name collision - I was importing `typing.Sequence`, but the code also needed `features.Sequence`. The tests from that PR were expecting the latter but got the former, and then crashed.", "@lhoestq Thanks! Also, when you're ready, don't merge it immediately! I'd like to do a quick round of manual testing with the very final build once you're happy to make sure it still works in our notebooks and examples.", "@lhoestq Tests look good to me, merging now!" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4170/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4170
https://github.com/huggingface/datasets/pull/4170
true
1,203,995,869
https://api.github.com/repos/huggingface/datasets/issues/4169/labels{/name}
## Dataset viewer issue for '*timit_asr*' **Link:** *https://huggingface.co/datasets/timit_asr* Issue: The timit-asr dataset cannot be previewed recently. Am I the one who added this dataset ? Yes-No No
2023-02-03T04:54:57Z
4,169
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-04-14T03:28:31Z
https://api.github.com/repos/huggingface/datasets/issues/4169/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4169/timeline
Timit_asr dataset cannot be previewed recently
https://api.github.com/repos/huggingface/datasets/issues/4169/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/75192317?v=4", "events_url": "https://api.github.com/users/YingLi001/events{/privacy}", "followers_url": "https://api.github.com/users/YingLi001/followers", "following_url": "https://api.github.com/users/YingLi001/following{/other_user}", "gists_url": "https://api.github.com/users/YingLi001/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/YingLi001", "id": 75192317, "login": "YingLi001", "node_id": "MDQ6VXNlcjc1MTkyMzE3", "organizations_url": "https://api.github.com/users/YingLi001/orgs", "received_events_url": "https://api.github.com/users/YingLi001/received_events", "repos_url": "https://api.github.com/users/YingLi001/repos", "site_admin": false, "starred_url": "https://api.github.com/users/YingLi001/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/YingLi001/subscriptions", "type": "User", "url": "https://api.github.com/users/YingLi001" }
[]
null
completed
NONE
2022-05-06T16:06:51Z
null
I_kwDODunzps5Hw4Td
[ "Thanks for reporting. The bug has already been detected, and we hope to fix it soon.", "TIMIT is now a dataset that requires manual download, see #4145 \r\n\r\nTherefore it might take a bit more time to fix it", "> TIMIT is now a dataset that requires manual download, see #4145\r\n> \r\n> Therefore it might take a bit more time to fix it\r\n\r\nThank you for your quickly response. Exactly, I also found the manual download issue in the morning. But when I used *list_datasets()* to check the available datasets, *'timit_asr'* is still in the list. So I am a little bit confused. If *'timit_asr'* need to be manually downloaded, does that mean we can **not** automatically download it **any more** in the future?", "Yes exactly. If you try to load the dataset it will ask you to download it manually first, and to pass the downloaded and extracted data like `load_dataset(\"timir_asr\", data_dir=\"path/to/extracted/data\")`\r\n\r\nThe URL we were using was coming from a host that doesn't have the permission to redistribute the data, and the dataset owners (LDC) notified us about it.", "I downloaded the timit_asr data and unzipped. But I can't run my code. Could you resolve this problem for me? Thanks\r\n\r\n import soundfile as sf\r\n import torch\r\n from datasets import load_dataset\r\n dataset = load_dataset(\"timit_asr\", data_dir=\"/Users/nguyenvannham/Documents/test_case/data\")\r\n \r\n \r\n Generating train split: 0 examples [00:00, ? examples/s]\r\n\r\nGenerating train split: 0 examples [00:00, ? examples/s]Traceback (most recent call last):\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/site-packages/datasets/builder.py\", line 1571, in _prepare_split_single\r\n for key, record in generator:\r\n\r\n File \"/Users/nguyenvannham/.cache/huggingface/modules/datasets_modules/datasets/timit_asr/43f9448dd5db58e95ee48a277f466481b151f112ea53e27f8173784da9254fb2/timit_asr.py\", line 138, in _generate_examples\r\n with txt_path.open(encoding=\"utf-8\") as op:\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/pathlib.py\", line 1252, in open\r\n return io.open(self, mode, buffering, encoding, errors, newline,\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/pathlib.py\", line 1120, in _opener\r\n return self._accessor.open(self, flags, mode)\r\n\r\nFileNotFoundError: [Errno 2] No such file or directory: '/Users/nguyenvannham/Documents/test_case/data/train/DR1/FCJF0/SA1.WAV.TXT'\r\n\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n\r\n File \"/var/folders/t9/l8d3rwpn1k33_gjtqs732lzc0000gn/T/ipykernel_3891/1203313828.py\", line 1, in <module>\r\n dataset = load_dataset(\"timit_asr\", data_dir=\"/Users/nguyenvannham/Documents/test_case/data\")\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/site-packages/datasets/load.py\", line 1758, in load_dataset\r\n builder_instance.download_and_prepare(\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/site-packages/datasets/builder.py\", line 860, in download_and_prepare\r\n self._download_and_prepare(\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/site-packages/datasets/builder.py\", line 1612, in _download_and_prepare\r\n super()._download_and_prepare(\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/site-packages/datasets/builder.py\", line 953, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/site-packages/datasets/builder.py\", line 1450, in _prepare_split\r\n for job_id, done, content in self._prepare_split_single(\r\n\r\n File \"/opt/anaconda3/envs/audio/lib/python3.9/site-packages/datasets/builder.py\", line 1607, in _prepare_split_single\r\n raise DatasetGenerationError(\"An error occurred while generating the dataset\") from e\r\n\r\nDatasetGenerationError: An error occurred while generating the dataset" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4169/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4169
https://github.com/huggingface/datasets/issues/4169
false
1,203,867,540
https://api.github.com/repos/huggingface/datasets/issues/4168/labels{/name}
This PR adds code examples for functions related to the base Datasets class to highlight usage. Most of the examples use the `rotten_tomatoes` dataset since it is nice and small. Several things I would appreciate feedback on: - Do you think it is clearer to make every code example fully reproducible so when users copy the code they can actually run it and get an output? This seems quite repetitive - maybe even unnecessary - but it is definitely clearer. Personally, I think we might be able to get away with not including this since users probably want to try the function on their own dataset. For example: ```py >>> from datasets import load_dataset >>> ds = load_dataset("rotten_tomatoes", split="validation") >>> code example goes here ``` - Should we showcase a function with more than one parameter to highlight different use-cases (it's pretty basic right now, but I'd be happy to add more)? - For the `class_encode_column` function, let me know if there is a simpler dataset with fewer columns (currently using `winograd_wsc`) so it is easier for users to see what changed. - Where possible, I try to show the input before and the output after using a function like `flatten` for example. Do you think this is too much and just showing the usage (ie, `>>> ds.flatten()`) will be sufficient? Thanks :)
2022-04-27T18:53:37Z
4,168
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
2022-04-13T23:03:38Z
https://api.github.com/repos/huggingface/datasets/issues/4168/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4168/timeline
Add code examples to API docs
https://api.github.com/repos/huggingface/datasets/issues/4168/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2022-04-27T18:48:34Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4168.diff", "html_url": "https://github.com/huggingface/datasets/pull/4168", "merged_at": "2022-04-27T18:48:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/4168.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4168" }
PR_kwDODunzps42NL6F
[ "_The documentation is not available anymore as the PR was closed or merged._", "> Do you think it is clearer to make every code example fully reproducible so when users copy the code they can actually run it and get an output? This seems quite repetitive - maybe even unnecessary - but it is definitely clearer.\r\n\r\nI think it's ok to be repetitive to get more clarity. Many users come from `transformers` and may have little experience with some processing methods (especially torch users).\r\n\r\n> Should we showcase a function with more than one parameter to highlight different use-cases (it's pretty basic right now, but I'd be happy to add more)?\r\n\r\nMaybe let's do it case by case, depending on whether there are parameters that are likely to be used often ?\r\n\r\n> For the class_encode_column function, let me know if there is a simpler dataset with fewer columns (currently using winograd_wsc) so it is easier for users to see what changed.\r\n\r\nYou can try with `boolq`, it has a boolean column that can be converted to labels\r\n\r\n> Where possible, I try to show the input before and the output after using a function like flatten for example. Do you think this is too much and just showing the usage (ie, >>> ds.flatten()) will be sufficient?\r\n\r\nNo I don't think it's too much, it's nice this way thanks :)", "Updated each code example so they are fully reproducible (where applicable)! The next step will be to identify some functions where we can show off some parameters that are useful or commonly used. Some useful parameters can be:\r\n\r\n- use `map(batched=True)` to process batches of examples.\r\n- set a seed in `shuffle`.\r\n- set `shuffle` and `seed` in `train_test_split`.\r\n\r\nLet me know if you think of anything else related to the functions in `arrow_dataset.py`!", "Cool thanks ! I think you can also do `num_proc` for `map`" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4168/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4168
https://github.com/huggingface/datasets/pull/4168
true
1,203,761,614
https://api.github.com/repos/huggingface/datasets/issues/4167/labels{/name}
use http.extraHeader to avoid rate limit
2022-04-13T20:56:41Z
4,167
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-13T20:32:17Z
https://api.github.com/repos/huggingface/datasets/issues/4167/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4167/timeline
Avoid rate limit in update hub repositories
https://api.github.com/repos/huggingface/datasets/issues/4167/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-04-13T20:50:32Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4167.diff", "html_url": "https://github.com/huggingface/datasets/pull/4167", "merged_at": "2022-04-13T20:50:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/4167.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4167" }
PR_kwDODunzps42M1O5
[ "I also set GIT_LFS_SKIP_SMUDGE=1 to speed up git clones", "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4167/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4167
https://github.com/huggingface/datasets/pull/4167
true
1,203,758,004
https://api.github.com/repos/huggingface/datasets/issues/4166/labels{/name}
Clarify docs and add clarifying example to the exact_match metric
2022-05-03T12:23:31Z
4,166
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-04-13T20:28:06Z
https://api.github.com/repos/huggingface/datasets/issues/4166/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4166/timeline
Fix exact match
https://api.github.com/repos/huggingface/datasets/issues/4166/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
null
null
CONTRIBUTOR
2022-05-03T12:16:27Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4166.diff", "html_url": "https://github.com/huggingface/datasets/pull/4166", "merged_at": "2022-05-03T12:16:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/4166.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4166" }
PR_kwDODunzps42M0dS
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4166/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4166
https://github.com/huggingface/datasets/pull/4166
true