url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.5B
node_id
stringlengths
18
32
number
int64
1
5.38k
title
stringlengths
1
276
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
sequence
created_at
stringlengths
20
20
updated_at
stringlengths
20
20
closed_at
stringlengths
20
20
author_association
stringclasses
3 values
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
3 values
is_pull_request
bool
1 class
https://api.github.com/repos/huggingface/datasets/issues/4151
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4151/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4151/comments
https://api.github.com/repos/huggingface/datasets/issues/4151/events
https://github.com/huggingface/datasets/pull/4151
1,201,837,999
PR_kwDODunzps42GgLu
4,151
Add missing label for emotion description
{ "avatar_url": "https://avatars.githubusercontent.com/u/44396506?v=4", "events_url": "https://api.github.com/users/lijiazheng99/events{/privacy}", "followers_url": "https://api.github.com/users/lijiazheng99/followers", "following_url": "https://api.github.com/users/lijiazheng99/following{/other_user}", "gists_url": "https://api.github.com/users/lijiazheng99/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lijiazheng99", "id": 44396506, "login": "lijiazheng99", "node_id": "MDQ6VXNlcjQ0Mzk2NTA2", "organizations_url": "https://api.github.com/users/lijiazheng99/orgs", "received_events_url": "https://api.github.com/users/lijiazheng99/received_events", "repos_url": "https://api.github.com/users/lijiazheng99/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lijiazheng99/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lijiazheng99/subscriptions", "type": "User", "url": "https://api.github.com/users/lijiazheng99" }
[]
closed
false
null
[]
null
[]
2022-04-12T13:17:37Z
2022-04-12T13:58:50Z
2022-04-12T13:58:50Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4151.diff", "html_url": "https://github.com/huggingface/datasets/pull/4151", "merged_at": "2022-04-12T13:58:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/4151.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4151" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4151/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4151/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4150
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4150/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4150/comments
https://api.github.com/repos/huggingface/datasets/issues/4150/events
https://github.com/huggingface/datasets/issues/4150
1,201,689,730
I_kwDODunzps5HoFSC
4,150
Inconsistent splits generation for datasets without loading script (packaged dataset puts everything into a single split)
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-04-12T11:15:55Z
2022-04-28T21:02:44Z
2022-04-28T21:02:44Z
CONTRIBUTOR
null
null
null
## Describe the bug Splits for dataset loaders without scripts are prepared inconsistently. I think it might be confusing for users. ## Steps to reproduce the bug * If you load a packaged datasets from Hub, it infers splits from directory structure / filenames (check out the data [here](https://huggingface.co/datasets/nateraw/test-imagefolder-dataset)): ```python ds = load_dataset("nateraw/test-imagefolder-dataset") print(ds) ### Output: DatasetDict({ train: Dataset({ features: ['image', 'label'], num_rows: 6 }) test: Dataset({ features: ['image', 'label'], num_rows: 4 }) }) ``` * If you do the same from locally stored data specifying only directory path you'll get the same: ```python ds = load_dataset("/path/to/local/data/test-imagefolder-dataset") print(ds) ### Output: DatasetDict({ train: Dataset({ features: ['image', 'label'], num_rows: 6 }) test: Dataset({ features: ['image', 'label'], num_rows: 4 }) }) ``` * However, if you explicitely specify package name (like `imagefolder`, `csv`, `json`), all the data is put into a single split: ```python ds = load_dataset("imagefolder", data_dir="/path/to/local/data/test-imagefolder-dataset") print(ds) ### Output: DatasetDict({ train: Dataset({ features: ['image', 'label'], num_rows: 10 }) }) ``` ## Expected results For `load_dataset("imagefolder", data_dir="/path/to/local/data/test-imagefolder-dataset")` I expect the same output as of the two first options.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4150/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4150/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4149
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4149/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4149/comments
https://api.github.com/repos/huggingface/datasets/issues/4149/events
https://github.com/huggingface/datasets/issues/4149
1,201,389,221
I_kwDODunzps5Hm76l
4,149
load_dataset for winoground returning decoding error
{ "avatar_url": "https://avatars.githubusercontent.com/u/4686956?v=4", "events_url": "https://api.github.com/users/odellus/events{/privacy}", "followers_url": "https://api.github.com/users/odellus/followers", "following_url": "https://api.github.com/users/odellus/following{/other_user}", "gists_url": "https://api.github.com/users/odellus/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/odellus", "id": 4686956, "login": "odellus", "node_id": "MDQ6VXNlcjQ2ODY5NTY=", "organizations_url": "https://api.github.com/users/odellus/orgs", "received_events_url": "https://api.github.com/users/odellus/received_events", "repos_url": "https://api.github.com/users/odellus/repos", "site_admin": false, "starred_url": "https://api.github.com/users/odellus/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/odellus/subscriptions", "type": "User", "url": "https://api.github.com/users/odellus" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-04-12T08:16:16Z
2022-05-04T23:40:38Z
2022-05-04T23:40:38Z
CONTRIBUTOR
null
null
null
## Describe the bug I am trying to use datasets to load winoground and I'm getting a JSON decoding error. ## Steps to reproduce the bug ```python from datasets import load_dataset token = 'hf_XXXXX' # my HF access token datasets = load_dataset('facebook/winoground', use_auth_token=token) ``` ## Expected results I downloaded images.zip and examples.jsonl manually. I was expecting to have some trouble decoding json so I didn't use jsonlines but instead was able to get a complete set of 400 examples by doing ```python import json with open('examples.jsonl', 'r') as f: examples = f.read().split('\n') # Thinking this would error if the JSON is not utf-8 encoded json_data = [json.loads(x) for x in examples] print(json_data[-1]) ``` and I see ```python {'caption_0': 'someone is overdoing it', 'caption_1': 'someone is doing it over', 'collapsed_tag': 'Relation', 'id': 399, 'image_0': 'ex_399_img_0', 'image_1': 'ex_399_img_1', 'num_main_preds': 1, 'secondary_tag': 'Morpheme-Level', 'tag': 'Scope, Preposition'} ``` so I'm not sure what's going on here honestly. The file `examples.jsonl` doesn't have non-UTF-8 encoded text. ## Actual results During the split operation after downloading, datasets encounters an error in the JSON ([trace](https://gist.github.com/odellus/e55d390ca203386bf551f38e0c63a46b) abbreviated for brevity). ``` datasets/packaged_modules/json/json.py:144 in Json._generate_tables(self, files) ... UnicodeDecodeError: 'utf-8' codec can't decode byte 0xff in position 0: invalid start byte ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.18.4 - Platform: Linux-5.13.0-39-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 7.0.0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4149/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4149/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4148
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4148/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4148/comments
https://api.github.com/repos/huggingface/datasets/issues/4148/events
https://github.com/huggingface/datasets/issues/4148
1,201,169,242
I_kwDODunzps5HmGNa
4,148
fix confusing bleu metric example
{ "avatar_url": "https://avatars.githubusercontent.com/u/6253193?v=4", "events_url": "https://api.github.com/users/aizawa-naoki/events{/privacy}", "followers_url": "https://api.github.com/users/aizawa-naoki/followers", "following_url": "https://api.github.com/users/aizawa-naoki/following{/other_user}", "gists_url": "https://api.github.com/users/aizawa-naoki/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/aizawa-naoki", "id": 6253193, "login": "aizawa-naoki", "node_id": "MDQ6VXNlcjYyNTMxOTM=", "organizations_url": "https://api.github.com/users/aizawa-naoki/orgs", "received_events_url": "https://api.github.com/users/aizawa-naoki/received_events", "repos_url": "https://api.github.com/users/aizawa-naoki/repos", "site_admin": false, "starred_url": "https://api.github.com/users/aizawa-naoki/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aizawa-naoki/subscriptions", "type": "User", "url": "https://api.github.com/users/aizawa-naoki" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[]
2022-04-12T06:18:26Z
2022-04-13T14:16:34Z
2022-04-13T14:16:34Z
NONE
null
null
null
**Is your feature request related to a problem? Please describe.** I would like to see the example in "Metric Card for BLEU" changed. The 0th element in the predictions list is not closed in square brackets, and the 1st list is missing a comma. The BLEU score are calculated correctly, but it is difficult to understand, so it would be helpful if you could correct this. ``` >> predictions = [ ... ["hello", "there", "general", "kenobi", # <- no closing square bracket. ... ["foo", "bar" "foobar"] # <- no comma between "bar" and "foobar" ... ] >>> references = [ ... [["hello", "there", "general", "kenobi"]], ... [["foo", "bar", "foobar"]] ... ] >>> bleu = datasets.load_metric("bleu") >>> results = bleu.compute(predictions=predictions, references=references) >>> print(results) {'bleu': 0.6370964381207871, ... ``` **Describe the solution you'd like** ``` >> predictions = [ ... ["hello", "there", "general", "kenobi", # <- no closing square bracket. ... ["foo", "bar" "foobar"] # <- no comma between "bar" and "foobar" ... ] # and >>> print(results) {'bleu':1.0, ... ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4148/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4148/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4147
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4147/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4147/comments
https://api.github.com/repos/huggingface/datasets/issues/4147/events
https://github.com/huggingface/datasets/pull/4147
1,200,756,008
PR_kwDODunzps42CtPl
4,147
Adjust path to datasets tutorial in How-To
{ "avatar_url": "https://avatars.githubusercontent.com/u/6765188?v=4", "events_url": "https://api.github.com/users/NimaBoscarino/events{/privacy}", "followers_url": "https://api.github.com/users/NimaBoscarino/followers", "following_url": "https://api.github.com/users/NimaBoscarino/following{/other_user}", "gists_url": "https://api.github.com/users/NimaBoscarino/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NimaBoscarino", "id": 6765188, "login": "NimaBoscarino", "node_id": "MDQ6VXNlcjY3NjUxODg=", "organizations_url": "https://api.github.com/users/NimaBoscarino/orgs", "received_events_url": "https://api.github.com/users/NimaBoscarino/received_events", "repos_url": "https://api.github.com/users/NimaBoscarino/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NimaBoscarino/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NimaBoscarino/subscriptions", "type": "User", "url": "https://api.github.com/users/NimaBoscarino" }
[]
closed
false
null
[]
null
[]
2022-04-12T01:20:34Z
2022-04-12T08:32:24Z
2022-04-12T08:26:02Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4147.diff", "html_url": "https://github.com/huggingface/datasets/pull/4147", "merged_at": "2022-04-12T08:26:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/4147.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4147" }
The link in the How-To overview page to the Datasets tutorials is currently broken. This is just a small adjustment to make it match the format used in https://github.com/huggingface/datasets/blob/master/docs/source/tutorial.md. (Edit to add: The link in the PR deployment (https://moon-ci-docs.huggingface.co/docs/datasets/pr_4147/en/how_to) is also broken since it's actually hardcoded to `master` and not dynamic to the branch name, but other links seem to behave similarly.)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4147/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4147/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4146
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4146/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4146/comments
https://api.github.com/repos/huggingface/datasets/issues/4146/events
https://github.com/huggingface/datasets/issues/4146
1,200,215,789
I_kwDODunzps5Hidbt
4,146
SAMSum dataset viewer not working
{ "avatar_url": "https://avatars.githubusercontent.com/u/39906333?v=4", "events_url": "https://api.github.com/users/aakashnegi10/events{/privacy}", "followers_url": "https://api.github.com/users/aakashnegi10/followers", "following_url": "https://api.github.com/users/aakashnegi10/following{/other_user}", "gists_url": "https://api.github.com/users/aakashnegi10/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/aakashnegi10", "id": 39906333, "login": "aakashnegi10", "node_id": "MDQ6VXNlcjM5OTA2MzMz", "organizations_url": "https://api.github.com/users/aakashnegi10/orgs", "received_events_url": "https://api.github.com/users/aakashnegi10/received_events", "repos_url": "https://api.github.com/users/aakashnegi10/repos", "site_admin": false, "starred_url": "https://api.github.com/users/aakashnegi10/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aakashnegi10/subscriptions", "type": "User", "url": "https://api.github.com/users/aakashnegi10" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-11T16:22:57Z
2022-04-29T16:26:09Z
2022-04-29T16:26:09Z
NONE
null
null
null
## Dataset viewer issue for '*name of the dataset*' **Link:** *link to the dataset viewer page* *short description of the issue* Am I the one who added this dataset ? Yes-No
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4146/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4146/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4145
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4145/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4145/comments
https://api.github.com/repos/huggingface/datasets/issues/4145/events
https://github.com/huggingface/datasets/pull/4145
1,200,209,781
PR_kwDODunzps42A6Rt
4,145
Redirect TIMIT download from LDC
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2022-04-11T16:17:55Z
2022-04-13T15:39:31Z
2022-04-13T15:33:04Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4145.diff", "html_url": "https://github.com/huggingface/datasets/pull/4145", "merged_at": "2022-04-13T15:33:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/4145.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4145" }
LDC data is protected under US copyright laws and under various legal agreements between the Linguistic Data Consortium/the University of Pennsylvania and data providers which prohibit redistribution of that data by anyone other than LDC. Similarly, LDC's membership agreements, non-member user agreement and various corpus-specific license agreements specifically state that users cannot publish, retransmit, disclose, copy, reproduce or redistribute LDC databases to others outside their organizations. LDC explicitly asked us to remove the download script for the TIMIT dataset. In this PR I remove all means to download the dataset, and redirect users to download the data from https://catalog.ldc.upenn.edu/LDC93S1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4145/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4145/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4144
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4144/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4144/comments
https://api.github.com/repos/huggingface/datasets/issues/4144/events
https://github.com/huggingface/datasets/pull/4144
1,200,016,983
PR_kwDODunzps42ARmu
4,144
Fix splits in local packaged modules, local datasets without script and hub datasets without script
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
closed
false
null
[]
null
[]
2022-04-11T13:57:33Z
2022-04-29T09:12:14Z
2022-04-28T21:02:45Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4144.diff", "html_url": "https://github.com/huggingface/datasets/pull/4144", "merged_at": "2022-04-28T21:02:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/4144.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4144" }
fixes #4150 I suggest to infer splits structure from files when `data_dir` is passed with `get_patterns_locally`, analogous to what's done in `LocalDatasetModuleFactoryWithoutScript` with `self.path`, instead of generating files with `data_dir/**` patterns and putting them all into a single default (train) split. I would also suggest to align `HubDatasetModuleFactoryWithoutScript` and `LocalDatasetModuleFactoryWithoutScript` with this logic (remove `data_files = os.path.join(data_dir, "**")`). It's not reflected in the current code now as I'd like to discuss it cause I might be unaware of some use cases. @lhoestq @mariosasko @albertvillanova WDYT?
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4144/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4144/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4143
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4143/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4143/comments
https://api.github.com/repos/huggingface/datasets/issues/4143/events
https://github.com/huggingface/datasets/issues/4143
1,199,937,961
I_kwDODunzps5HhZmp
4,143
Unable to download `Wikepedia` 20220301.en version
{ "avatar_url": "https://avatars.githubusercontent.com/u/37113676?v=4", "events_url": "https://api.github.com/users/beyondguo/events{/privacy}", "followers_url": "https://api.github.com/users/beyondguo/followers", "following_url": "https://api.github.com/users/beyondguo/following{/other_user}", "gists_url": "https://api.github.com/users/beyondguo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/beyondguo", "id": 37113676, "login": "beyondguo", "node_id": "MDQ6VXNlcjM3MTEzNjc2", "organizations_url": "https://api.github.com/users/beyondguo/orgs", "received_events_url": "https://api.github.com/users/beyondguo/received_events", "repos_url": "https://api.github.com/users/beyondguo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/beyondguo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/beyondguo/subscriptions", "type": "User", "url": "https://api.github.com/users/beyondguo" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-04-11T13:00:14Z
2022-08-17T00:37:55Z
2022-04-21T17:04:14Z
NONE
null
null
null
## Describe the bug Unable to download `Wikepedia` dataset, 20220301.en version ## Steps to reproduce the bug ```python !pip install apache_beam mwparserfromhell dataset_wikipedia = load_dataset("wikipedia", "20220301.en") ``` ## Actual results ``` ValueError: BuilderConfig 20220301.en not found. Available: ['20200501.aa', '20200501.ab', '20200501.ace', '20200501.ady', '20200501.af', '20200501.ak', '20200501.als', '20200501.am', '20200501.an', '20200501.ang', '20200501.ar', '20200501.arc', '20200501.arz', '20200501.as', '20200501.ast', '20200501.atj', '20200501.av', '20200501.ay', '20200501.az', '20200501.azb', '20200501.ba', '20200501.bar', '20200501.bat-smg', '20200501.bcl', '20200501.be', '20200501.be-x-old', '20200501.bg', '20200501.bh', '20200501.bi', '20200501.bjn', '20200501.bm', '20200501.bn', '20200501.bo', '20200501.bpy', '20200501.br', '20200501.bs', '20200501.bug', '20200501.bxr', '20200501.ca', '20200501.cbk-zam', '20200501.cdo', '20200501.ce', '20200501.ceb', '20200501.ch', '20200501.cho', '20200501.chr', '20200501.chy', '20200501.ckb', '20200501.co', '20200501.cr', '20200501.crh', '20200501.cs', '20200501.csb', '20200501.cu', '20200501.cv', '20200501.cy', '20200501.da', '20200501.de', '20200501.din', '20200501.diq', '20200501.dsb', '20200501.dty', '20200501.dv', '20200501.dz', '20200501.ee', '20200501.el', '20200501.eml', '20200501.en', '20200501.eo', '20200501.es', '20200501.et', '20200501.eu', '20200501.ext', '20200501.fa', '20200501.ff', '20200501.fi', '20200501.fiu-vro', '20200501.fj', '20200501.fo', '20200501.fr', '20200501.frp', '20200501.frr', '20200501.fur', '20200501.fy', '20200501.ga', '20200501.gag', '20200501.gan', '20200501.gd', '20200501.gl', '20200501.glk', '20200501.gn', '20200501.gom', '20200501.gor', '20200501.got', '20200501.gu', '20200501.gv', '20200501.ha', '20200501.hak', '20200501.haw', '20200501.he', '20200501.hi', '20200501.hif', '20200501.ho', '20200501.hr', '20200501.hsb', '20200501.ht', '20200501.hu', '20200501.hy', '20200501.ia', '20200501.id', '20200501.ie', '20200501.ig', '20200501.ii', '20200501.ik', '20200501.ilo', '20200501.inh', '20200501.io', '20200501.is', '20200501.it', '20200501.iu', '20200501.ja', '20200501.jam', '20200501.jbo', '20200501.jv', '20200501.ka', '20200501.kaa', '20200501.kab', '20200501.kbd', '20200501.kbp', '20200501.kg', '20200501.ki', '20200501.kj', '20200501.kk', '20200501.kl', '20200501.km', '20200501.kn', '20200501.ko', '20200501.koi', '20200501.krc', '20200501.ks', '20200501.ksh', '20200501.ku', '20200501.kv', '20200501.kw', '20200501.ky', '20200501.la', '20200501.lad', '20200501.lb', '20200501.lbe', '20200501.lez', '20200501.lfn', '20200501.lg', '20200501.li', '20200501.lij', '20200501.lmo', '20200501.ln', '20200501.lo', '20200501.lrc', '20200501.lt', '20200501.ltg', '20200501.lv', '20200501.mai', '20200501.map-bms', '20200501.mdf', '20200501.mg', '20200501.mh', '20200501.mhr', '20200501.mi', '20200501.min', '20200501.mk', '20200501.ml', '20200501.mn', '20200501.mr', '20200501.mrj', '20200501.ms', '20200501.mt', '20200501.mus', '20200501.mwl', '20200501.my', '20200501.myv', '20200501.mzn', '20200501.na', '20200501.nah', '20200501.nap', '20200501.nds', '20200501.nds-nl', '20200501.ne', '20200501.new', '20200501.ng', '20200501.nl', '20200501.nn', '20200501.no', '20200501.nov', '20200501.nrm', '20200501.nso', '20200501.nv', '20200501.ny', '20200501.oc', '20200501.olo', '20200501.om', '20200501.or', '20200501.os', '20200501.pa', '20200501.pag', '20200501.pam', '20200501.pap', '20200501.pcd', '20200501.pdc', '20200501.pfl', '20200501.pi', '20200501.pih', '20200501.pl', '20200501.pms', '20200501.pnb', '20200501.pnt', '20200501.ps', '20200501.pt', '20200501.qu', '20200501.rm', '20200501.rmy', '20200501.rn', '20200501.ro', '20200501.roa-rup', '20200501.roa-tara', '20200501.ru', '20200501.rue', '20200501.rw', '20200501.sa', '20200501.sah', '20200501.sat', '20200501.sc', '20200501.scn', '20200501.sco', '20200501.sd', '20200501.se', '20200501.sg', '20200501.sh', '20200501.si', '20200501.simple', '20200501.sk', '20200501.sl', '20200501.sm', '20200501.sn', '20200501.so', '20200501.sq', '20200501.sr', '20200501.srn', '20200501.ss', '20200501.st', '20200501.stq', '20200501.su', '20200501.sv', '20200501.sw', '20200501.szl', '20200501.ta', '20200501.tcy', '20200501.te', '20200501.tet', '20200501.tg', '20200501.th', '20200501.ti', '20200501.tk', '20200501.tl', '20200501.tn', '20200501.to', '20200501.tpi', '20200501.tr', '20200501.ts', '20200501.tt', '20200501.tum', '20200501.tw', '20200501.ty', '20200501.tyv', '20200501.udm', '20200501.ug', '20200501.uk', '20200501.ur', '20200501.uz', '20200501.ve', '20200501.vec', '20200501.vep', '20200501.vi', '20200501.vls', '20200501.vo', '20200501.wa', '20200501.war', '20200501.wo', '20200501.wuu', '20200501.xal', '20200501.xh', '20200501.xmf', '20200501.yi', '20200501.yo', '20200501.za', '20200501.zea', '20200501.zh', '20200501.zh-classical', '20200501.zh-min-nan', '20200501.zh-yue', '20200501.zu'] ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.0.0 - Platform: Ubuntu - Python version: 3.6 - PyArrow version: 6.0.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4143/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4143/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4142
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4142/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4142/comments
https://api.github.com/repos/huggingface/datasets/issues/4142/events
https://github.com/huggingface/datasets/issues/4142
1,199,794,750
I_kwDODunzps5Hg2o-
4,142
Add ObjectFolder 2.0 dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/7246357?v=4", "events_url": "https://api.github.com/users/osanseviero/events{/privacy}", "followers_url": "https://api.github.com/users/osanseviero/followers", "following_url": "https://api.github.com/users/osanseviero/following{/other_user}", "gists_url": "https://api.github.com/users/osanseviero/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/osanseviero", "id": 7246357, "login": "osanseviero", "node_id": "MDQ6VXNlcjcyNDYzNTc=", "organizations_url": "https://api.github.com/users/osanseviero/orgs", "received_events_url": "https://api.github.com/users/osanseviero/received_events", "repos_url": "https://api.github.com/users/osanseviero/repos", "site_admin": false, "starred_url": "https://api.github.com/users/osanseviero/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/osanseviero/subscriptions", "type": "User", "url": "https://api.github.com/users/osanseviero" }
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
open
false
null
[]
null
[]
2022-04-11T10:57:51Z
2022-10-05T10:30:49Z
null
MEMBER
null
null
null
## Adding a Dataset - **Name:** ObjectFolder 2.0 - **Description:** ObjectFolder 2.0 is a dataset of 1,000 objects in the form of implicit representations. It contains 1,000 Object Files each containing the complete multisensory profile for an object instance. - **Paper:** [*link to the dataset paper if available*](https://arxiv.org/abs/2204.02389) - **Data:** https://github.com/rhgao/ObjectFolder Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4142/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4142/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4141
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4141/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4141/comments
https://api.github.com/repos/huggingface/datasets/issues/4141/events
https://github.com/huggingface/datasets/issues/4141
1,199,610,885
I_kwDODunzps5HgJwF
4,141
Why is the dataset not visible under the dataset preview section?
{ "avatar_url": "https://avatars.githubusercontent.com/u/75028682?v=4", "events_url": "https://api.github.com/users/Nid989/events{/privacy}", "followers_url": "https://api.github.com/users/Nid989/followers", "following_url": "https://api.github.com/users/Nid989/following{/other_user}", "gists_url": "https://api.github.com/users/Nid989/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Nid989", "id": 75028682, "login": "Nid989", "node_id": "MDQ6VXNlcjc1MDI4Njgy", "organizations_url": "https://api.github.com/users/Nid989/orgs", "received_events_url": "https://api.github.com/users/Nid989/received_events", "repos_url": "https://api.github.com/users/Nid989/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Nid989/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nid989/subscriptions", "type": "User", "url": "https://api.github.com/users/Nid989" }
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
closed
false
null
[]
null
[]
2022-04-11T08:36:42Z
2022-04-11T18:55:32Z
2022-04-11T17:09:49Z
NONE
null
null
null
## Dataset viewer issue for '*name of the dataset*' **Link:** *link to the dataset viewer page* *short description of the issue* Am I the one who added this dataset ? Yes-No
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4141/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4141/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4140
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4140/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4140/comments
https://api.github.com/repos/huggingface/datasets/issues/4140/events
https://github.com/huggingface/datasets/issues/4140
1,199,492,356
I_kwDODunzps5Hfs0E
4,140
Error loading arxiv data set
{ "avatar_url": "https://avatars.githubusercontent.com/u/5383918?v=4", "events_url": "https://api.github.com/users/yjqiu/events{/privacy}", "followers_url": "https://api.github.com/users/yjqiu/followers", "following_url": "https://api.github.com/users/yjqiu/following{/other_user}", "gists_url": "https://api.github.com/users/yjqiu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yjqiu", "id": 5383918, "login": "yjqiu", "node_id": "MDQ6VXNlcjUzODM5MTg=", "organizations_url": "https://api.github.com/users/yjqiu/orgs", "received_events_url": "https://api.github.com/users/yjqiu/received_events", "repos_url": "https://api.github.com/users/yjqiu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yjqiu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yjqiu/subscriptions", "type": "User", "url": "https://api.github.com/users/yjqiu" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-04-11T07:06:34Z
2022-04-12T16:24:08Z
2022-04-12T16:24:08Z
NONE
null
null
null
## Describe the bug A clear and concise description of what the bug is. I met the error below when loading arxiv dataset via `nlp.load_dataset('scientific_papers', 'arxiv',)`. ``` Traceback (most recent call last): File "scripts/summarization.py", line 354, in <module> main(args) File "scripts/summarization.py", line 306, in main model.hf_datasets = nlp.load_dataset('scientific_papers', 'arxiv') File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/load.py", line 549, in load_dataset download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications, File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/builder.py", line 463, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/builder.py", line 522, in _download_and_prepare self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), "dataset source files" File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/utils/info_utils.py", line 38, in verify_checksums raise NonMatchingChecksumError(error_msg + str(bad_urls)) nlp.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://drive.google.com/uc?id=1b3rmCSIoh6VhD4HKWjI4HOW-cSwcwbeC&export=download', 'https://drive.google.com/uc?id=1lvsqvsFi3W-pE1SqNZI0s8NR9rC1tsja&export=download'] ``` I then tried to ignore verification steps by `ignore_verifications=True` and there is another error. ``` Traceback (most recent call last): File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/builder.py", line 537, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/builder.py", line 810, in _prepare_split for key, record in utils.tqdm(generator, unit=" examples", total=split_info.num_examples, leave=False): File "/opt/conda/envs/longformer/lib/python3.7/site-packages/tqdm/std.py", line 1195, in __iter__ for obj in iterable: File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/datasets/scientific_papers/9e4f2cfe3d8494e9f34a84ce49c3214605b4b52a3d8eb199104430d04c52cc12/scientific_papers.py", line 108, in _generate_examples with open(path, encoding="utf-8") as f: NotADirectoryError: [Errno 20] Not a directory: '/home/username/.cache/huggingface/datasets/downloads/c0deae7af7d9c87f25dfadf621f7126f708d7dcac6d353c7564883084a000076/arxiv-dataset/train.txt' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "scripts/summarization.py", line 354, in <module> main(args) File "scripts/summarization.py", line 306, in main model.hf_datasets = nlp.load_dataset('scientific_papers', 'arxiv', ignore_verifications=True) File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/load.py", line 549, in load_dataset download_config=download_config, download_mode=download_mode, ignore_verifications=ignore_verifications, File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/builder.py", line 463, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "/opt/conda/envs/longformer/lib/python3.7/site-packages/nlp/builder.py", line 539, in _download_and_prepare raise OSError("Cannot find data file. " + (self.manual_download_instructions or "")) OSError: Cannot find data file. ``` ## Steps to reproduce the bug ```python # Sample code to reproduce the bug ``` ## Expected results A clear and concise description of the expected results. ## Actual results Specify the actual results or traceback. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: - Python version: - PyArrow version:
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4140/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4140/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4139
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4139/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4139/comments
https://api.github.com/repos/huggingface/datasets/issues/4139/events
https://github.com/huggingface/datasets/issues/4139
1,199,443,822
I_kwDODunzps5Hfg9u
4,139
Dataset viewer issue for Winoground
{ "avatar_url": "https://avatars.githubusercontent.com/u/7438704?v=4", "events_url": "https://api.github.com/users/alcinos/events{/privacy}", "followers_url": "https://api.github.com/users/alcinos/followers", "following_url": "https://api.github.com/users/alcinos/following{/other_user}", "gists_url": "https://api.github.com/users/alcinos/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alcinos", "id": 7438704, "login": "alcinos", "node_id": "MDQ6VXNlcjc0Mzg3MDQ=", "organizations_url": "https://api.github.com/users/alcinos/orgs", "received_events_url": "https://api.github.com/users/alcinos/received_events", "repos_url": "https://api.github.com/users/alcinos/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alcinos/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alcinos/subscriptions", "type": "User", "url": "https://api.github.com/users/alcinos" }
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" }, { "color": "51F745", "default": false, "description": "", "id": 4030248571, "name": "dataset-viewer-gated", "node_id": "LA_kwDODunzps7wOLZ7", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer-gated" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }, { "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "events_url": "https://api.github.com/users/SBrandeis/events{/privacy}", "followers_url": "https://api.github.com/users/SBrandeis/followers", "following_url": "https://api.github.com/users/SBrandeis/following{/other_user}", "gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SBrandeis", "id": 33657802, "login": "SBrandeis", "node_id": "MDQ6VXNlcjMzNjU3ODAy", "organizations_url": "https://api.github.com/users/SBrandeis/orgs", "received_events_url": "https://api.github.com/users/SBrandeis/received_events", "repos_url": "https://api.github.com/users/SBrandeis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions", "type": "User", "url": "https://api.github.com/users/SBrandeis" }, { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
[]
2022-04-11T06:11:41Z
2022-06-21T16:43:58Z
2022-06-21T16:43:58Z
NONE
null
null
null
## Dataset viewer issue for 'Winoground' **Link:** [*link to the dataset viewer page*](https://huggingface.co/datasets/facebook/winoground/viewer/facebook--winoground/train) *short description of the issue* Getting 401, message='Unauthorized' The dataset is subject to authorization, but I can access the files from the interface, so I assume I'm granted to access it. I'd assume the permission somehow doesn't propagate to the dataset viewer tool. Am I the one who added this dataset ? No
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4139/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4139/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4138
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4138/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4138/comments
https://api.github.com/repos/huggingface/datasets/issues/4138/events
https://github.com/huggingface/datasets/issues/4138
1,199,291,730
I_kwDODunzps5He71S
4,138
Incorrect Russian filenames encoding after extraction by datasets.DownloadManager.download_and_extract()
{ "avatar_url": "https://avatars.githubusercontent.com/u/55381086?v=4", "events_url": "https://api.github.com/users/iluvvatar/events{/privacy}", "followers_url": "https://api.github.com/users/iluvvatar/followers", "following_url": "https://api.github.com/users/iluvvatar/following{/other_user}", "gists_url": "https://api.github.com/users/iluvvatar/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/iluvvatar", "id": 55381086, "login": "iluvvatar", "node_id": "MDQ6VXNlcjU1MzgxMDg2", "organizations_url": "https://api.github.com/users/iluvvatar/orgs", "received_events_url": "https://api.github.com/users/iluvvatar/received_events", "repos_url": "https://api.github.com/users/iluvvatar/repos", "site_admin": false, "starred_url": "https://api.github.com/users/iluvvatar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iluvvatar/subscriptions", "type": "User", "url": "https://api.github.com/users/iluvvatar" }
[]
closed
false
null
[]
null
[]
2022-04-11T02:07:13Z
2022-04-19T03:15:46Z
2022-04-16T15:46:29Z
NONE
null
null
null
## Dataset viewer issue for 'MalakhovIlya/RuREBus' **Link:** https://huggingface.co/datasets/MalakhovIlya/RuREBus **Description** Using os.walk(topdown=False) in DatasetBuilder causes following error: Status code: 400 Exception: TypeError Message: xwalk() got an unexpected keyword argument 'topdown' Couldn't find where "xwalk" come from. How can I fix this? Am I the one who added this dataset ? Yes
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4138/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4138/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4137
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4137/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4137/comments
https://api.github.com/repos/huggingface/datasets/issues/4137/events
https://github.com/huggingface/datasets/pull/4137
1,199,000,453
PR_kwDODunzps419D6A
4,137
Add single dataset citations for TweetEval
{ "avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4", "events_url": "https://api.github.com/users/gchhablani/events{/privacy}", "followers_url": "https://api.github.com/users/gchhablani/followers", "following_url": "https://api.github.com/users/gchhablani/following{/other_user}", "gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/gchhablani", "id": 29076344, "login": "gchhablani", "node_id": "MDQ6VXNlcjI5MDc2MzQ0", "organizations_url": "https://api.github.com/users/gchhablani/orgs", "received_events_url": "https://api.github.com/users/gchhablani/received_events", "repos_url": "https://api.github.com/users/gchhablani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions", "type": "User", "url": "https://api.github.com/users/gchhablani" }
[]
closed
false
null
[]
null
[]
2022-04-10T11:51:54Z
2022-04-12T07:57:22Z
2022-04-12T07:51:15Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4137.diff", "html_url": "https://github.com/huggingface/datasets/pull/4137", "merged_at": "2022-04-12T07:51:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/4137.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4137" }
This PR adds single data citations as per request of the original creators of the TweetEval dataset. This is a recent email from the creator: > Could I ask you a favor? Would you be able to add at the end of the README the citations of the single datasets as well? You can just copy our readme maybe? https://github.com/cardiffnlp/tweeteval#citing-tweeteval (just to be sure that the creator of the single datasets also get credits when tweeteval is used) Please let me know if this looks okay or if any changes are needed. Thanks, Gunjan
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4137/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4137/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4135
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4135/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4135/comments
https://api.github.com/repos/huggingface/datasets/issues/4135/events
https://github.com/huggingface/datasets/pull/4135
1,198,307,610
PR_kwDODunzps416-Rn
4,135
Support streaming xtreme dataset for PAN-X config
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-09T06:19:48Z
2022-05-06T08:39:40Z
2022-04-11T06:54:14Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4135.diff", "html_url": "https://github.com/huggingface/datasets/pull/4135", "merged_at": "2022-04-11T06:54:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/4135.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4135" }
Support streaming xtreme dataset for PAN-X config.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4135/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4135/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4134
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4134/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4134/comments
https://api.github.com/repos/huggingface/datasets/issues/4134/events
https://github.com/huggingface/datasets/issues/4134
1,197,937,146
I_kwDODunzps5HZxH6
4,134
ELI5 supporting documents
{ "avatar_url": "https://avatars.githubusercontent.com/u/69015896?v=4", "events_url": "https://api.github.com/users/saurabh-0077/events{/privacy}", "followers_url": "https://api.github.com/users/saurabh-0077/followers", "following_url": "https://api.github.com/users/saurabh-0077/following{/other_user}", "gists_url": "https://api.github.com/users/saurabh-0077/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/saurabh-0077", "id": 69015896, "login": "saurabh-0077", "node_id": "MDQ6VXNlcjY5MDE1ODk2", "organizations_url": "https://api.github.com/users/saurabh-0077/orgs", "received_events_url": "https://api.github.com/users/saurabh-0077/received_events", "repos_url": "https://api.github.com/users/saurabh-0077/repos", "site_admin": false, "starred_url": "https://api.github.com/users/saurabh-0077/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/saurabh-0077/subscriptions", "type": "User", "url": "https://api.github.com/users/saurabh-0077" }
[ { "color": "d876e3", "default": true, "description": "Further information is requested", "id": 1935892912, "name": "question", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question" } ]
open
false
null
[]
null
[]
2022-04-08T23:36:27Z
2022-04-13T13:52:46Z
null
NONE
null
null
null
if i am using dense search to create supporting documents for eli5 how much time it will take bcz i read somewhere that it takes about 18 hrs??
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4134/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4134/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4133
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4133/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4133/comments
https://api.github.com/repos/huggingface/datasets/issues/4133/events
https://github.com/huggingface/datasets/issues/4133
1,197,830,623
I_kwDODunzps5HZXHf
4,133
HANS dataset preview broken
{ "avatar_url": "https://avatars.githubusercontent.com/u/61748653?v=4", "events_url": "https://api.github.com/users/pietrolesci/events{/privacy}", "followers_url": "https://api.github.com/users/pietrolesci/followers", "following_url": "https://api.github.com/users/pietrolesci/following{/other_user}", "gists_url": "https://api.github.com/users/pietrolesci/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/pietrolesci", "id": 61748653, "login": "pietrolesci", "node_id": "MDQ6VXNlcjYxNzQ4NjUz", "organizations_url": "https://api.github.com/users/pietrolesci/orgs", "received_events_url": "https://api.github.com/users/pietrolesci/received_events", "repos_url": "https://api.github.com/users/pietrolesci/repos", "site_admin": false, "starred_url": "https://api.github.com/users/pietrolesci/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pietrolesci/subscriptions", "type": "User", "url": "https://api.github.com/users/pietrolesci" }
[ { "color": "fef2c0", "default": false, "description": "", "id": 3287858981, "name": "streaming", "node_id": "MDU6TGFiZWwzMjg3ODU4OTgx", "url": "https://api.github.com/repos/huggingface/datasets/labels/streaming" } ]
closed
false
null
[]
null
[]
2022-04-08T21:06:15Z
2022-04-13T11:57:34Z
2022-04-13T11:57:34Z
NONE
null
null
null
## Dataset viewer issue for '*hans*' **Link:** [https://huggingface.co/datasets/hans](https://huggingface.co/datasets/hans) HANS dataset preview is broken with error 400 Am I the one who added this dataset ? No
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4133/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4133/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4132
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4132/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4132/comments
https://api.github.com/repos/huggingface/datasets/issues/4132/events
https://github.com/huggingface/datasets/pull/4132
1,197,661,720
PR_kwDODunzps41460R
4,132
Support streaming xtreme dataset for PAWS-X config
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-08T18:25:32Z
2022-05-06T08:39:42Z
2022-04-08T21:02:44Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4132.diff", "html_url": "https://github.com/huggingface/datasets/pull/4132", "merged_at": "2022-04-08T21:02:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/4132.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4132" }
Support streaming xtreme dataset for PAWS-X config.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4132/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4132/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4131
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4131/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4131/comments
https://api.github.com/repos/huggingface/datasets/issues/4131/events
https://github.com/huggingface/datasets/pull/4131
1,197,472,249
PR_kwDODunzps414Zt1
4,131
Support streaming xtreme dataset for udpos config
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-08T15:30:49Z
2022-05-06T08:39:46Z
2022-04-08T16:28:07Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4131.diff", "html_url": "https://github.com/huggingface/datasets/pull/4131", "merged_at": "2022-04-08T16:28:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/4131.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4131" }
Support streaming xtreme dataset for udpos config.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4131/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4131/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4130
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4130/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4130/comments
https://api.github.com/repos/huggingface/datasets/issues/4130/events
https://github.com/huggingface/datasets/pull/4130
1,197,456,857
PR_kwDODunzps414Wqx
4,130
Add SBU Captions Photo Dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
[]
closed
false
null
[]
null
[]
2022-04-08T15:17:39Z
2022-04-12T10:47:31Z
2022-04-12T10:41:29Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4130.diff", "html_url": "https://github.com/huggingface/datasets/pull/4130", "merged_at": "2022-04-12T10:41:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/4130.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4130" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4130/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4130/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4129
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4129/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4129/comments
https://api.github.com/repos/huggingface/datasets/issues/4129/events
https://github.com/huggingface/datasets/issues/4129
1,197,376,796
I_kwDODunzps5HXoUc
4,129
dataset metadata for reproducibility
{ "avatar_url": "https://avatars.githubusercontent.com/u/24982805?v=4", "events_url": "https://api.github.com/users/nbroad1881/events{/privacy}", "followers_url": "https://api.github.com/users/nbroad1881/followers", "following_url": "https://api.github.com/users/nbroad1881/following{/other_user}", "gists_url": "https://api.github.com/users/nbroad1881/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nbroad1881", "id": 24982805, "login": "nbroad1881", "node_id": "MDQ6VXNlcjI0OTgyODA1", "organizations_url": "https://api.github.com/users/nbroad1881/orgs", "received_events_url": "https://api.github.com/users/nbroad1881/received_events", "repos_url": "https://api.github.com/users/nbroad1881/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nbroad1881/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nbroad1881/subscriptions", "type": "User", "url": "https://api.github.com/users/nbroad1881" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
2022-04-08T14:17:28Z
2022-04-08T14:17:28Z
null
NONE
null
null
null
When pulling a dataset from the hub, it would be useful to have some metadata about the specific dataset and version that is used. The metadata could then be passed to the `Trainer` which could then be saved to a model card. This is useful for people who run many experiments on different versions (commits/branches) of the same dataset. The dataset could have a list of “source datasets” metadata and ignore what happens to them before arriving in the Trainer (i.e. ignore mapping, filtering, etc.). Here is a basic representation (made by @lhoestq ) ```python >>> from datasets import load_dataset >>> >>> my_dataset = load_dataset(...)["train"] >>> my_dataset = my_dataset.map(...) >>> >>> my_dataset.sources [HFHubDataset(repo_id=..., revision=..., arguments={...})] ```
{ "+1": 4, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 4, "url": "https://api.github.com/repos/huggingface/datasets/issues/4129/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4129/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4128
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4128/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4128/comments
https://api.github.com/repos/huggingface/datasets/issues/4128/events
https://github.com/huggingface/datasets/pull/4128
1,197,326,311
PR_kwDODunzps4138I6
4,128
More robust `cast_to_python_objects` in `TypedSequence`
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[]
2022-04-08T13:33:35Z
2022-04-13T14:07:41Z
2022-04-13T14:01:16Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4128.diff", "html_url": "https://github.com/huggingface/datasets/pull/4128", "merged_at": "2022-04-13T14:01:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/4128.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4128" }
Adds a fallback to run an expensive version of `cast_to_python_objects` which exhaustively checks entire lists to avoid the `ArrowInvalid: Could not convert` error in `TypedSequence`. Currently, this error can happen in situations where only some images are decoded in `map`, in which case `cast_to_python_objects` fails to recognize that it needs to cast `PIL.Image` objects if they are not at the beginning of the sequence and stops after the first image dictionary (e.g., if `data` is `[{"bytes": None, "path": "some path"}, PIL.Image(), ...]`) Fix #4124
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4128/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4128/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4127
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4127/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4127/comments
https://api.github.com/repos/huggingface/datasets/issues/4127/events
https://github.com/huggingface/datasets/pull/4127
1,197,297,756
PR_kwDODunzps4132EN
4,127
Add configs with processed data in medical_dialog dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-08T13:08:16Z
2022-05-06T08:39:50Z
2022-04-08T16:20:51Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4127.diff", "html_url": "https://github.com/huggingface/datasets/pull/4127", "merged_at": "2022-04-08T16:20:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/4127.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4127" }
There exist processed data files that do not require parsing the raw data files (which can take long time). Fix #4122.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4127/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4127/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4126
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4126/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4126/comments
https://api.github.com/repos/huggingface/datasets/issues/4126/events
https://github.com/huggingface/datasets/issues/4126
1,196,665,194
I_kwDODunzps5HU6lq
4,126
dataset viewer issue for common_voice
{ "avatar_url": "https://avatars.githubusercontent.com/u/24724502?v=4", "events_url": "https://api.github.com/users/laphang/events{/privacy}", "followers_url": "https://api.github.com/users/laphang/followers", "following_url": "https://api.github.com/users/laphang/following{/other_user}", "gists_url": "https://api.github.com/users/laphang/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/laphang", "id": 24724502, "login": "laphang", "node_id": "MDQ6VXNlcjI0NzI0NTAy", "organizations_url": "https://api.github.com/users/laphang/orgs", "received_events_url": "https://api.github.com/users/laphang/received_events", "repos_url": "https://api.github.com/users/laphang/repos", "site_admin": false, "starred_url": "https://api.github.com/users/laphang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/laphang/subscriptions", "type": "User", "url": "https://api.github.com/users/laphang" }
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" }, { "color": "F83ACF", "default": false, "description": "", "id": 4027368468, "name": "audio_column", "node_id": "LA_kwDODunzps7wDMQU", "url": "https://api.github.com/repos/huggingface/datasets/labels/audio_column" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" } ]
null
[]
2022-04-07T23:34:28Z
2022-04-25T13:42:17Z
2022-04-25T13:42:16Z
NONE
null
null
null
## Dataset viewer issue for 'common_voice' **Link:** https://huggingface.co/datasets/common_voice Server Error Status code: 400 Exception: TypeError Message: __init__() got an unexpected keyword argument 'audio_column' Am I the one who added this dataset ? No
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4126/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4126/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4125
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4125/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4125/comments
https://api.github.com/repos/huggingface/datasets/issues/4125/events
https://github.com/huggingface/datasets/pull/4125
1,196,633,936
PR_kwDODunzps411qeR
4,125
BIG-bench
{ "avatar_url": "https://avatars.githubusercontent.com/u/43357549?v=4", "events_url": "https://api.github.com/users/andersjohanandreassen/events{/privacy}", "followers_url": "https://api.github.com/users/andersjohanandreassen/followers", "following_url": "https://api.github.com/users/andersjohanandreassen/following{/other_user}", "gists_url": "https://api.github.com/users/andersjohanandreassen/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/andersjohanandreassen", "id": 43357549, "login": "andersjohanandreassen", "node_id": "MDQ6VXNlcjQzMzU3NTQ5", "organizations_url": "https://api.github.com/users/andersjohanandreassen/orgs", "received_events_url": "https://api.github.com/users/andersjohanandreassen/received_events", "repos_url": "https://api.github.com/users/andersjohanandreassen/repos", "site_admin": false, "starred_url": "https://api.github.com/users/andersjohanandreassen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andersjohanandreassen/subscriptions", "type": "User", "url": "https://api.github.com/users/andersjohanandreassen" }
[]
closed
false
null
[]
null
[]
2022-04-07T22:33:30Z
2022-06-08T17:57:48Z
2022-06-08T17:32:32Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4125.diff", "html_url": "https://github.com/huggingface/datasets/pull/4125", "merged_at": "2022-06-08T17:32:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/4125.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4125" }
This PR adds all BIG-bench json tasks to huggingface/datasets.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4125/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4125/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4124
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4124/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4124/comments
https://api.github.com/repos/huggingface/datasets/issues/4124/events
https://github.com/huggingface/datasets/issues/4124
1,196,469,842
I_kwDODunzps5HUK5S
4,124
Image decoding often fails when transforming Image datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/17025191?v=4", "events_url": "https://api.github.com/users/RafayAK/events{/privacy}", "followers_url": "https://api.github.com/users/RafayAK/followers", "following_url": "https://api.github.com/users/RafayAK/following{/other_user}", "gists_url": "https://api.github.com/users/RafayAK/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/RafayAK", "id": 17025191, "login": "RafayAK", "node_id": "MDQ6VXNlcjE3MDI1MTkx", "organizations_url": "https://api.github.com/users/RafayAK/orgs", "received_events_url": "https://api.github.com/users/RafayAK/received_events", "repos_url": "https://api.github.com/users/RafayAK/repos", "site_admin": false, "starred_url": "https://api.github.com/users/RafayAK/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/RafayAK/subscriptions", "type": "User", "url": "https://api.github.com/users/RafayAK" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-07T19:17:25Z
2022-04-13T14:01:16Z
2022-04-13T14:01:16Z
NONE
null
null
null
## Describe the bug When transforming/modifying images in an image dataset using the `map` function the PIL images often fail to decode in time for the image transforms, causing errors. Using a debugger it is easy to see what the problem is, the Image decode invocation does not take place and the resulting image passed around is still raw bytes: ``` [{'bytes': b'\x89PNG\r\n\x1a\n\x00\x00\x00\rIHDR\x00\x00\x00 \x00\x00\x00 \x08\x02\x00\x00\x00\xfc\x18\xed\xa3\x00\x00\x08\x02IDATx\x9cEVIs[\xc7\x11\xeemf\xde\x82\x8d\x80\x08\x89"\xb5V\\\xb6\x94(\xe5\x9f\x90\xca5\x7f$\xa7T\xe5\x9f&9\xd9\x8a\\.\xdb\xa4$J\xa4\x00\x02x\xc0{\xb3t\xe7\x00\xca\x99\xd3\\f\xba\xba\xbf\xa5?|\xfa\xf4\xa2\xeb\xba\xedv\xa3f^\xf8\xd5\x0bY\xb6\x10\xb3\xaaDq\xcd\x83\x87\xdf5\xf3gZ\x1a\x04\x0f\xa0fp\xfa\xe0\xd4\x07?\x9dN\xc4\xb1\x99\xfd\xf2\xcb/\x97\x97\x97H\xa2\xaaf\x16\x82\xaf\xeb\xca{\xbf\xd9l.\xdf\x7f\xfa\xcb_\xff&\x88\x08\x00\x80H\xc0\x80@.;\x0f\x8c@#v\xe3\xe5\xfc\xd1\x9f\xee6q\xbf\xdf\xa6\x14\'\x93\xf1\xc3\xe5\xe3\xd1x\x14c\x8c1\xa5\x1c\x9dsM\xd3\xb4\xed\x08\x89SJ)\xa5\xedv\xbb^\xafNO\x97D\x84Hf .... ``` ## Steps to reproduce the bug ```python from datasets import load_dataset, Dataset import numpy as np # seeded NumPy random number generator for reprodducinble results. rng = np.random.default_rng(seed=0) test_dataset = load_dataset('cifar100', split="test") def preprocess_data(dataset): """ Helper function to pre-process HuggingFace Cifar-100 Dataset to remove fine_label and coarse_label columns and add is_flipped column Args: dataset: HuggingFace CIFAR-100 Dataset Object Returns: new_dataset: A Dataset object with "img" and "is_flipped" columns only """ # remove fine_label and coarse_label columns new_dataset = dataset.remove_columns(['fine_label', 'coarse_label']) # add the column for is_flipped new_dataset = new_dataset.add_column(name="is_flipped", column=np.zeros((len(new_dataset)), dtype=np.uint8)) return new_dataset def generate_flipped_data(example, p=0.5): """ A Dataset mapping function that transforms some of the images up-side-down. If the probability value (p) is 0.5 approximately half the images will be flipped upside-down Args: example: An example from the dataset containing a Python dictionary with "img" and "is_flipped" key-value pair p: the probability of flipping the image up-side-down, Default 0.5 Returns: example: A Dataset object """ # example['img'] = example['img'] if rng.random() > p: # the flip the image and set is_flipped column to 1 example['img'] = example['img'].transpose( 1) # ImageOps.flip(example['img']) #example['img'].transpose(Image.FLIP_TOP_BOTTOM) example['is_flipped'] = 1 return example my_test = preprocess_data(test_dataset) my_test = my_test.map(generate_flipped_data) ``` ## Expected results The dataset should be transformed without problems. ## Actual results ``` /home/rafay/anaconda3/envs/pytorch_new/bin/python /home/rafay/Documents/you_only_live_once/upside_down_detector/create_dataset.py Reusing dataset cifar100 (/home/rafay/.cache/huggingface/datasets/cifar100/cifar100/1.0.0/f365c8b725c23e8f0f8d725c3641234d9331cd2f62919d1381d1baa5b3ba3142) Reusing dataset cifar100 (/home/rafay/.cache/huggingface/datasets/cifar100/cifar100/1.0.0/f365c8b725c23e8f0f8d725c3641234d9331cd2f62919d1381d1baa5b3ba3142) 20%|█▉ | 1999/10000 [00:00<00:01, 5560.44ex/s] Traceback (most recent call last): File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2326, in _map_single writer.write(example) File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_writer.py", line 441, in write self.write_examples_on_file() File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_writer.py", line 399, in write_examples_on_file self.write_batch(batch_examples=batch_examples) File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_writer.py", line 492, in write_batch arrays.append(pa.array(typed_sequence)) File "pyarrow/array.pxi", line 230, in pyarrow.lib.array File "pyarrow/array.pxi", line 110, in pyarrow.lib._handle_arrow_array_protocol File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_writer.py", line 185, in __arrow_array__ out = pa.array(cast_to_python_objects(data, only_1d_for_numpy=True)) File "pyarrow/array.pxi", line 316, in pyarrow.lib.array File "pyarrow/array.pxi", line 39, in pyarrow.lib._sequence_to_array File "pyarrow/error.pxi", line 143, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 99, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: Could not convert <PIL.Image.Image image mode=RGB size=32x32 at 0x7F56AEE61DE0> with type Image: did not recognize Python value type when inferring an Arrow data type During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/rafay/Documents/you_only_live_once/upside_down_detector/create_dataset.py", line 55, in <module> my_test = my_test.map(generate_flipped_data) File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 1953, in map return self._map_single( File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 519, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 486, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/fingerprint.py", line 458, in wrapper out = func(self, *args, **kwargs) File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2360, in _map_single writer.finalize() File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_writer.py", line 522, in finalize self.write_examples_on_file() File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_writer.py", line 399, in write_examples_on_file self.write_batch(batch_examples=batch_examples) File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_writer.py", line 492, in write_batch arrays.append(pa.array(typed_sequence)) File "pyarrow/array.pxi", line 230, in pyarrow.lib.array File "pyarrow/array.pxi", line 110, in pyarrow.lib._handle_arrow_array_protocol File "/home/rafay/anaconda3/envs/pytorch_new/lib/python3.10/site-packages/datasets/arrow_writer.py", line 185, in __arrow_array__ out = pa.array(cast_to_python_objects(data, only_1d_for_numpy=True)) File "pyarrow/array.pxi", line 316, in pyarrow.lib.array File "pyarrow/array.pxi", line 39, in pyarrow.lib._sequence_to_array File "pyarrow/error.pxi", line 143, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 99, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: Could not convert <PIL.Image.Image image mode=RGB size=32x32 at 0x7F56AEE61DE0> with type Image: did not recognize Python value type when inferring an Arrow data type Process finished with exit code 1 ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.0.0 - Platform: Linux(Fedora 35) - Python version: 3.10 - PyArrow version: 7.0.0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4124/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4124/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4123
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4123/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4123/comments
https://api.github.com/repos/huggingface/datasets/issues/4123/events
https://github.com/huggingface/datasets/issues/4123
1,196,367,512
I_kwDODunzps5HTx6Y
4,123
Building C4 takes forever
{ "avatar_url": "https://avatars.githubusercontent.com/u/15899312?v=4", "events_url": "https://api.github.com/users/StellaAthena/events{/privacy}", "followers_url": "https://api.github.com/users/StellaAthena/followers", "following_url": "https://api.github.com/users/StellaAthena/following{/other_user}", "gists_url": "https://api.github.com/users/StellaAthena/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/StellaAthena", "id": 15899312, "login": "StellaAthena", "node_id": "MDQ6VXNlcjE1ODk5MzEy", "organizations_url": "https://api.github.com/users/StellaAthena/orgs", "received_events_url": "https://api.github.com/users/StellaAthena/received_events", "repos_url": "https://api.github.com/users/StellaAthena/repos", "site_admin": false, "starred_url": "https://api.github.com/users/StellaAthena/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/StellaAthena/subscriptions", "type": "User", "url": "https://api.github.com/users/StellaAthena" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
open
false
null
[]
null
[]
2022-04-07T17:41:30Z
2022-04-08T13:22:19Z
null
NONE
null
null
null
## Describe the bug C4-en is a 300 GB dataset. However, when I try to download it through the hub it takes over _six hours_ to generate the train/test split from the downloaded files. This is an absurd amount of time and an unnecessary waste of resources. ## Steps to reproduce the bug ```python c4 = datasets.load("c4", "en") ``` ## Expected results I would like to be able to download pre-split data. ## Environment info - `datasets` version: 2.0.0 - Platform: Linux-5.13.0-35-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 7.0.0 - Pandas version: 1.4.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4123/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4123/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4122
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4122/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4122/comments
https://api.github.com/repos/huggingface/datasets/issues/4122/events
https://github.com/huggingface/datasets/issues/4122
1,196,095,072
I_kwDODunzps5HSvZg
4,122
medical_dialog zh has very slow _generate_examples
{ "avatar_url": "https://avatars.githubusercontent.com/u/24982805?v=4", "events_url": "https://api.github.com/users/nbroad1881/events{/privacy}", "followers_url": "https://api.github.com/users/nbroad1881/followers", "following_url": "https://api.github.com/users/nbroad1881/following{/other_user}", "gists_url": "https://api.github.com/users/nbroad1881/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nbroad1881", "id": 24982805, "login": "nbroad1881", "node_id": "MDQ6VXNlcjI0OTgyODA1", "organizations_url": "https://api.github.com/users/nbroad1881/orgs", "received_events_url": "https://api.github.com/users/nbroad1881/received_events", "repos_url": "https://api.github.com/users/nbroad1881/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nbroad1881/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nbroad1881/subscriptions", "type": "User", "url": "https://api.github.com/users/nbroad1881" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-07T14:00:51Z
2022-04-08T16:20:51Z
2022-04-08T16:20:51Z
NONE
null
null
null
## Describe the bug After downloading the files from Google Drive, `load_dataset("medical_dialog", "zh", data_dir="./")` takes an unreasonable amount of time. Generating the train/test split for 33% of the dataset takes over 4.5 hours. ## Steps to reproduce the bug The easiest way I've found to download files from Google Drive is to use `gdown` and use Google Colab because the download speeds will be very high due to the fact that they are both in Google Cloud. ```python file_ids = [ "1AnKxGEuzjeQsDHHqL3NqI_aplq2hVL_E", "1tt7weAT1SZknzRFyLXOT2fizceUUVRXX", "1A64VBbsQ_z8wZ2LDox586JIyyO6mIwWc", "1AKntx-ECnrxjB07B6BlVZcFRS4YPTB-J", "1xUk8AAua_x27bHUr-vNoAuhEAjTxOvsu", "1ezKTfe7BgqVN5o-8Vdtr9iAF0IueCSjP", "1tA7bSOxR1RRNqZst8cShzhuNHnayUf7c", "1pA3bCFA5nZDhsQutqsJcH3d712giFb0S", "1pTLFMdN1A3ro-KYghk4w4sMz6aGaMOdU", "1dUSnG0nUPq9TEQyHd6ZWvaxO0OpxVjXD", "1UfCH05nuWiIPbDZxQzHHGAHyMh8dmPQH", ] for i in file_ids: url = f"https://drive.google.com/uc?id={i}" !gdown $url from datasets import load_dataset ds = load_dataset("medical_dialog", "zh", data_dir="./") ``` ## Expected results Faster load time ## Actual results `Generating train split: 33%: 625519/1921127 [4:31:03<31:39:20, 11.37 examples/s]` ## Environment info - `datasets` version: 2.0.0 - Platform: Linux-5.4.144+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - PyArrow version: 6.0.1 - Pandas version: 1.3.5 @vrindaprabhu , could you take a look at this since you implemented it? I think the `_generate_examples` function might need to be rewritten
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4122/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4122/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4121
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4121/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4121/comments
https://api.github.com/repos/huggingface/datasets/issues/4121/events
https://github.com/huggingface/datasets/issues/4121
1,196,000,018
I_kwDODunzps5HSYMS
4,121
datasets.load_metric can not load a local metirc
{ "avatar_url": "https://avatars.githubusercontent.com/u/51749469?v=4", "events_url": "https://api.github.com/users/Gare-Ng/events{/privacy}", "followers_url": "https://api.github.com/users/Gare-Ng/followers", "following_url": "https://api.github.com/users/Gare-Ng/following{/other_user}", "gists_url": "https://api.github.com/users/Gare-Ng/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Gare-Ng", "id": 51749469, "login": "Gare-Ng", "node_id": "MDQ6VXNlcjUxNzQ5NDY5", "organizations_url": "https://api.github.com/users/Gare-Ng/orgs", "received_events_url": "https://api.github.com/users/Gare-Ng/received_events", "repos_url": "https://api.github.com/users/Gare-Ng/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Gare-Ng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Gare-Ng/subscriptions", "type": "User", "url": "https://api.github.com/users/Gare-Ng" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-04-07T12:48:56Z
2022-04-07T13:53:27Z
2022-04-07T13:53:27Z
NONE
null
null
null
## Describe the bug No matter how I hard try to tell load_metric that I want to load a local metric file, it still continues to fetch things on the Internet. And unfortunately it says 'ConnectionError: Couldn't reach'. However I can download this file without connectionerror and tell load_metric its local directory. And it comes back where it begins... ## Steps to reproduce the bug ```python metric = load_metric(path=r'C:\Users\Gare\PycharmProjects\Gare\blue\bleu.py') ConnectionError: Couldn't reach https://github.com/tensorflow/nmt/raw/master/nmt/scripts/bleu.py metric = load_metric(path='bleu') ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/1.12.1/metrics/bleu/bleu.py metric = load_metric(path='./blue/bleu.py') ConnectionError: Couldn't reach https://github.com/tensorflow/nmt/raw/master/nmt/scripts/bleu.py ``` ## Expected results I do read the docs [here](https://huggingface.co/docs/datasets/package_reference/loading_methods#datasets.load_metric). There are no other parameters that help function to distinguish from local and online file but path. As what I code above, it should load from local. ## Actual results > metric = load_metric(path=r'C:\Users\Gare\PycharmProjects\Gare\blue\bleu.py') > ~\AppData\Local\Temp\ipykernel_19636\1855752034.py in <module> ----> 1 metric = load_metric(path=r'C:\Users\Gare\PycharmProjects\Gare\blue\bleu.py') D:\Program Files\Anaconda\envs\Gare\lib\site-packages\datasets\load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs) 817 if data_files is None and data_dir is not None: 818 data_files = os.path.join(data_dir, "**") --> 819 820 self.name = name 821 self.revision = revision D:\Program Files\Anaconda\envs\Gare\lib\site-packages\datasets\load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, dynamic_modules_path, return_resolved_file_path, return_associated_base_path, data_files, **download_kwargs) 639 self, 640 path: str, --> 641 download_config: Optional[DownloadConfig] = None, 642 download_mode: Optional[DownloadMode] = None, 643 dynamic_modules_path: Optional[str] = None, D:\Program Files\Anaconda\envs\Gare\lib\site-packages\datasets\utils\file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs) 297 token = hf_api.HfFolder.get_token() 298 if token: --> 299 headers["authorization"] = f"Bearer {token}" 300 return headers 301 D:\Program Files\Anaconda\envs\Gare\lib\site-packages\datasets\utils\file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token) 604 def _resumable_file_manager(): 605 with open(incomplete_path, "a+b") as f: --> 606 yield f 607 608 temp_file_manager = _resumable_file_manager ConnectionError: Couldn't reach https://github.com/tensorflow/nmt/raw/master/nmt/scripts/bleu.py ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.0.0 - Platform: Windows-10-10.0.22000-SP0 - Python version: 3.7.13 - PyArrow version: 7.0.0 - Pandas version: 1.3.4 Any advice would be appreciated.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4121/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4121/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4120
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4120/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4120/comments
https://api.github.com/repos/huggingface/datasets/issues/4120/events
https://github.com/huggingface/datasets/issues/4120
1,195,887,430
I_kwDODunzps5HR8tG
4,120
Representing dictionaries (json) objects as features
{ "avatar_url": "https://avatars.githubusercontent.com/u/8031035?v=4", "events_url": "https://api.github.com/users/yanaiela/events{/privacy}", "followers_url": "https://api.github.com/users/yanaiela/followers", "following_url": "https://api.github.com/users/yanaiela/following{/other_user}", "gists_url": "https://api.github.com/users/yanaiela/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yanaiela", "id": 8031035, "login": "yanaiela", "node_id": "MDQ6VXNlcjgwMzEwMzU=", "organizations_url": "https://api.github.com/users/yanaiela/orgs", "received_events_url": "https://api.github.com/users/yanaiela/received_events", "repos_url": "https://api.github.com/users/yanaiela/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yanaiela/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yanaiela/subscriptions", "type": "User", "url": "https://api.github.com/users/yanaiela" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
2022-04-07T11:07:41Z
2022-04-07T11:07:41Z
null
CONTRIBUTOR
null
null
null
In the process of adding a new dataset to the hub, I stumbled upon the inability to represent dictionaries that contain different key names, unknown in advance (and may differ between samples), original asked in the [forum](https://discuss.huggingface.co/t/representing-nested-dictionary-with-different-keys/16442). For instance: ``` sample1 = {"nps": { "a": {"id": 0, "text": "text1"}, "b": {"id": 1, "text": "text2"}, }} sample2 = {"nps": { "a": {"id": 0, "text": "text1"}, "b": {"id": 1, "text": "text2"}, "c": {"id": 2, "text": "text3"}, }} sample3 = {"nps": { "a": {"id": 0, "text": "text1"}, "b": {"id": 1, "text": "text2"}, "c": {"id": 2, "text": "text3"}, "d": {"id": 3, "text": "text4"}, }} ``` the `nps` field cannot be represented as a Feature while maintaining its original structure. @lhoestq suggested to add JSON as a new feature type, which will solve this problem. It seems like an alternative solution would be to change the original data format, which isn't an optimal solution in my case. Moreover, JSON is a common structure, that will likely to be useful in future datasets as well.
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4120/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4120/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4119
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4119/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4119/comments
https://api.github.com/repos/huggingface/datasets/issues/4119/events
https://github.com/huggingface/datasets/pull/4119
1,195,641,298
PR_kwDODunzps41yXHF
4,119
Hotfix failing CI tests on Windows
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-07T07:38:46Z
2022-04-07T09:47:24Z
2022-04-07T07:57:13Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4119.diff", "html_url": "https://github.com/huggingface/datasets/pull/4119", "merged_at": "2022-04-07T07:57:13Z", "patch_url": "https://github.com/huggingface/datasets/pull/4119.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4119" }
This PR makes a hotfix for our CI Windows tests: https://app.circleci.com/pipelines/github/huggingface/datasets/11092/workflows/9cfdb1dd-0fec-4fe0-8122-5f533192ebdc/jobs/67414 Fix #4118 I guess this issue is related to this PR: - huggingface/huggingface_hub#815
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4119/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4119/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4118
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4118/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4118/comments
https://api.github.com/repos/huggingface/datasets/issues/4118/events
https://github.com/huggingface/datasets/issues/4118
1,195,638,944
I_kwDODunzps5HRACg
4,118
Failing CI tests on Windows
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-07T07:36:25Z
2022-04-07T07:57:13Z
2022-04-07T07:57:13Z
MEMBER
null
null
null
## Describe the bug Our CI Windows tests are failing from yesterday: https://app.circleci.com/pipelines/github/huggingface/datasets/11092/workflows/9cfdb1dd-0fec-4fe0-8122-5f533192ebdc/jobs/67414
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4118/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4118/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4117
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4117/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4117/comments
https://api.github.com/repos/huggingface/datasets/issues/4117/events
https://github.com/huggingface/datasets/issues/4117
1,195,552,406
I_kwDODunzps5HQq6W
4,117
AttributeError: module 'huggingface_hub' has no attribute 'hf_api'
{ "avatar_url": "https://avatars.githubusercontent.com/u/4567991?v=4", "events_url": "https://api.github.com/users/arymbe/events{/privacy}", "followers_url": "https://api.github.com/users/arymbe/followers", "following_url": "https://api.github.com/users/arymbe/following{/other_user}", "gists_url": "https://api.github.com/users/arymbe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/arymbe", "id": 4567991, "login": "arymbe", "node_id": "MDQ6VXNlcjQ1Njc5OTE=", "organizations_url": "https://api.github.com/users/arymbe/orgs", "received_events_url": "https://api.github.com/users/arymbe/received_events", "repos_url": "https://api.github.com/users/arymbe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/arymbe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/arymbe/subscriptions", "type": "User", "url": "https://api.github.com/users/arymbe" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-07T05:52:36Z
2022-07-28T16:44:04Z
2022-04-19T15:36:35Z
NONE
null
null
null
## Describe the bug Could you help me please. I got this following error. AttributeError: module 'huggingface_hub' has no attribute 'hf_api' ## Steps to reproduce the bug when I imported the datasets # Sample code to reproduce the bug from datasets import list_datasets, load_dataset, list_metrics, load_metric ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.0.0 - Platform: macOS-12.3-x86_64-i386-64bit - Python version: 3.8.9 - PyArrow version: 7.0.0 - Pandas version: 1.3.5 - Huggingface-hub: 0.5.0 - Transformers: 4.18.0 Thank you in advance.
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4117/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4117/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4116
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4116/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4116/comments
https://api.github.com/repos/huggingface/datasets/issues/4116/events
https://github.com/huggingface/datasets/pull/4116
1,194,926,459
PR_kwDODunzps41wCEO
4,116
Pretty print dataset info files
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[]
2022-04-06T17:40:48Z
2022-04-08T11:28:01Z
2022-04-08T11:21:53Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4116.diff", "html_url": "https://github.com/huggingface/datasets/pull/4116", "merged_at": "2022-04-08T11:21:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/4116.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4116" }
Adds indentation to the `dataset_infos.json` file when saving for nicer diffs. (suggested by @julien-c) This PR also updates the info files of the GH datasets. Note that this change adds more than **10 MB** to the repo size (the total file size before the change: 29.672298 MB, after: 41.666475 MB), so I'm not sure this change is a good idea. `src/datasets/info.py` is the only relevant file for reviewers.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4116/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4116/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4115
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4115/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4115/comments
https://api.github.com/repos/huggingface/datasets/issues/4115/events
https://github.com/huggingface/datasets/issues/4115
1,194,907,555
I_kwDODunzps5HONej
4,115
ImageFolder add option to ignore some folders like '.ipynb_checkpoints'
{ "avatar_url": "https://avatars.githubusercontent.com/u/15624271?v=4", "events_url": "https://api.github.com/users/cceyda/events{/privacy}", "followers_url": "https://api.github.com/users/cceyda/followers", "following_url": "https://api.github.com/users/cceyda/following{/other_user}", "gists_url": "https://api.github.com/users/cceyda/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cceyda", "id": 15624271, "login": "cceyda", "node_id": "MDQ6VXNlcjE1NjI0Mjcx", "organizations_url": "https://api.github.com/users/cceyda/orgs", "received_events_url": "https://api.github.com/users/cceyda/received_events", "repos_url": "https://api.github.com/users/cceyda/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cceyda/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cceyda/subscriptions", "type": "User", "url": "https://api.github.com/users/cceyda" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[]
2022-04-06T17:29:43Z
2022-06-01T13:04:16Z
2022-06-01T13:04:16Z
CONTRIBUTOR
null
null
null
**Is your feature request related to a problem? Please describe.** I sometimes like to peek at the dataset images from jupyterlab. thus '.ipynb_checkpoints' folder appears where my dataset is and (just realized) leads to accidental duplicate image additions. I think this is an easy enough thing to miss especially if the dataset is very large. **Describe the solution you'd like** maybe have an option `ignore` or something .gitignore style `dataset = load_dataset("imagefolder", data_dir="./data/original", ignore="regex?")` **Describe alternatives you've considered** Could filter out manually
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4115/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4115/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4114
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4114/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4114/comments
https://api.github.com/repos/huggingface/datasets/issues/4114/events
https://github.com/huggingface/datasets/issues/4114
1,194,855,345
I_kwDODunzps5HOAux
4,114
Allow downloading just some columns of a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/7246357?v=4", "events_url": "https://api.github.com/users/osanseviero/events{/privacy}", "followers_url": "https://api.github.com/users/osanseviero/followers", "following_url": "https://api.github.com/users/osanseviero/following{/other_user}", "gists_url": "https://api.github.com/users/osanseviero/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/osanseviero", "id": 7246357, "login": "osanseviero", "node_id": "MDQ6VXNlcjcyNDYzNTc=", "organizations_url": "https://api.github.com/users/osanseviero/orgs", "received_events_url": "https://api.github.com/users/osanseviero/received_events", "repos_url": "https://api.github.com/users/osanseviero/repos", "site_admin": false, "starred_url": "https://api.github.com/users/osanseviero/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/osanseviero/subscriptions", "type": "User", "url": "https://api.github.com/users/osanseviero" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
2022-04-06T16:38:46Z
2022-04-07T07:56:26Z
null
MEMBER
null
null
null
**Is your feature request related to a problem? Please describe.** Some people are interested in doing label analysis of a CV dataset without downloading all the images. Downloading the whole dataset does not always makes sense for this kind of use case **Describe the solution you'd like** Be able to just download some columns of a dataset, such as doing ```python load_dataset("huggan/wikiart",columns=["artist", "genre"]) ``` Although this might make things a bit complicated in terms of local caching of datasets.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4114/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4114/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4113
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4113/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4113/comments
https://api.github.com/repos/huggingface/datasets/issues/4113/events
https://github.com/huggingface/datasets/issues/4113
1,194,843,532
I_kwDODunzps5HN92M
4,113
Multiprocessing with FileLock fails in python 3.9
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-04-06T16:27:09Z
2022-11-28T11:49:14Z
2022-11-28T11:49:14Z
MEMBER
null
null
null
On python 3.9, this code hangs: ```python from multiprocessing import Pool from filelock import FileLock def run(i): print(f"got the lock in multi process [{i}]") with FileLock("tmp.lock"): with Pool(2) as pool: pool.map(run, range(2)) ``` This is because the subprocesses try to acquire the lock from the main process for some reason. This is not the case in older versions of python. This can cause many issues in python 3.9. In particular, we use multiprocessing to fetch data files when you load a dataset (as long as there are >16 data files). Therefore `imagefolder` hangs, and I expect any dataset that needs to download >16 files to hang as well. Let's see if we can fix this and have a CI that runs on 3.9. cc @mariosasko @julien-c
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4113/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4113/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4112
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4112/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4112/comments
https://api.github.com/repos/huggingface/datasets/issues/4112/events
https://github.com/huggingface/datasets/issues/4112
1,194,752,765
I_kwDODunzps5HNnr9
4,112
ImageFolder with Grayscale images dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/50595514?v=4", "events_url": "https://api.github.com/users/ChainYo/events{/privacy}", "followers_url": "https://api.github.com/users/ChainYo/followers", "following_url": "https://api.github.com/users/ChainYo/following{/other_user}", "gists_url": "https://api.github.com/users/ChainYo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ChainYo", "id": 50595514, "login": "ChainYo", "node_id": "MDQ6VXNlcjUwNTk1NTE0", "organizations_url": "https://api.github.com/users/ChainYo/orgs", "received_events_url": "https://api.github.com/users/ChainYo/received_events", "repos_url": "https://api.github.com/users/ChainYo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ChainYo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ChainYo/subscriptions", "type": "User", "url": "https://api.github.com/users/ChainYo" }
[]
closed
false
null
[]
null
[]
2022-04-06T15:10:00Z
2022-04-22T10:21:53Z
2022-04-22T10:21:52Z
NONE
null
null
null
Hi, I'm facing a problem with a grayscale images dataset I have uploaded [here](https://huggingface.co/datasets/ChainYo/rvl-cdip) (RVL-CDIP) I'm getting an error while I want to use images for training a model with PyTorch DataLoader. Here is the full traceback: ```bash AttributeError: Caught AttributeError in DataLoader worker process 0. Original Traceback (most recent call last): File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/torch/utils/data/_utils/worker.py", line 287, in _worker_loop data = fetcher.fetch(index) File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 49, in fetch data = [self.dataset[idx] for idx in possibly_batched_index] File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 49, in <listcomp> data = [self.dataset[idx] for idx in possibly_batched_index] File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 1765, in __getitem__ return self._getitem( File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 1750, in _getitem formatted_output = format_table( File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/formatting/formatting.py", line 532, in format_table return formatter(pa_table, query_type=query_type) File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/formatting/formatting.py", line 281, in __call__ return self.format_row(pa_table) File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/formatting/torch_formatter.py", line 58, in format_row return self.recursive_tensorize(row) File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/formatting/torch_formatter.py", line 54, in recursive_tensorize return map_nested(self._recursive_tensorize, data_struct, map_list=False) File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 314, in map_nested mapped = [ File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 315, in <listcomp> _single_map_nested((function, obj, types, None, True, None)) File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 267, in _single_map_nested return {k: _single_map_nested((function, v, types, None, True, None)) for k, v in pbar} File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 267, in <dictcomp> return {k: _single_map_nested((function, v, types, None, True, None)) for k, v in pbar} File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 251, in _single_map_nested return function(data_struct) File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/formatting/torch_formatter.py", line 51, in _recursive_tensorize return self._tensorize(data_struct) File "/home/chainyo/miniconda3/envs/gan-bird/lib/python3.8/site-packages/datasets/formatting/torch_formatter.py", line 38, in _tensorize if np.issubdtype(value.dtype, np.integer): AttributeError: 'bytes' object has no attribute 'dtype' ``` I don't really understand why the image is still a bytes object while I used transformations on it. Here the code I used to upload the dataset (and it worked well): ```python train_dataset = load_dataset("imagefolder", data_dir="data/train") train_dataset = train_dataset["train"] test_dataset = load_dataset("imagefolder", data_dir="data/test") test_dataset = test_dataset["train"] val_dataset = load_dataset("imagefolder", data_dir="data/val") val_dataset = val_dataset["train"] dataset = DatasetDict({ "train": train_dataset, "val": val_dataset, "test": test_dataset }) dataset.push_to_hub("ChainYo/rvl-cdip") ``` Now here is the code I am using to get the dataset and prepare it for training: ```python img_size = 512 batch_size = 128 normalize = [(0.5), (0.5)] data_dir = "ChainYo/rvl-cdip" dataset = load_dataset(data_dir, split="train") transforms = transforms.Compose([ transforms.Resize(img_size), transforms.CenterCrop(img_size), transforms.ToTensor(), transforms.Normalize(*normalize) ]) transformed_dataset = dataset.with_transform(transforms) transformed_dataset.set_format(type="torch", device="cuda") train_dataloader = torch.utils.data.DataLoader( transformed_dataset, batch_size=batch_size, shuffle=True, num_workers=4, pin_memory=True ) ``` But this get me the error above. I don't understand why it's doing this kind of weird thing? Do I need to map something on the dataset? Something like this: ```python labels = dataset.features["label"].names num_labels = dataset.features["label"].num_classes def preprocess_data(examples): images = [ex.convert("RGB") for ex in examples["image"]] labels = [ex for ex in examples["label"]] return {"images": images, "labels": labels} features = Features({ "images": Image(decode=True, id=None), "labels": ClassLabel(num_classes=num_labels, names=labels) }) decoded_dataset = dataset.map(preprocess_data, remove_columns=dataset.column_names, features=features, batched=True, batch_size=100) ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4112/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4112/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4111
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4111/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4111/comments
https://api.github.com/repos/huggingface/datasets/issues/4111/events
https://github.com/huggingface/datasets/pull/4111
1,194,660,699
PR_kwDODunzps41vJCt
4,111
Update security policy
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-06T13:59:51Z
2022-04-07T09:46:30Z
2022-04-07T09:40:27Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4111.diff", "html_url": "https://github.com/huggingface/datasets/pull/4111", "merged_at": "2022-04-07T09:40:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/4111.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4111" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4111/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4111/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4110
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4110/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4110/comments
https://api.github.com/repos/huggingface/datasets/issues/4110/events
https://github.com/huggingface/datasets/pull/4110
1,194,581,375
PR_kwDODunzps41u4Je
4,110
Matthews Correlation Metric Card
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
closed
false
null
[]
null
[]
2022-04-06T12:59:35Z
2022-05-03T13:43:17Z
2022-05-03T13:36:13Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4110.diff", "html_url": "https://github.com/huggingface/datasets/pull/4110", "merged_at": "2022-05-03T13:36:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/4110.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4110" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4110/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4110/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4109
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4109/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4109/comments
https://api.github.com/repos/huggingface/datasets/issues/4109/events
https://github.com/huggingface/datasets/pull/4109
1,194,579,257
PR_kwDODunzps41u3sm
4,109
Add Spearmanr Metric Card
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
closed
false
null
[]
null
[]
2022-04-06T12:57:53Z
2022-05-03T16:50:26Z
2022-05-03T16:43:37Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4109.diff", "html_url": "https://github.com/huggingface/datasets/pull/4109", "merged_at": "2022-05-03T16:43:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/4109.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4109" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4109/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4109/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4108
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4108/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4108/comments
https://api.github.com/repos/huggingface/datasets/issues/4108/events
https://github.com/huggingface/datasets/pull/4108
1,194,578,584
PR_kwDODunzps41u3j2
4,108
Perplexity Speedup
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
closed
false
null
[]
null
[]
2022-04-06T12:57:21Z
2022-04-20T13:00:54Z
2022-04-20T12:54:42Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4108.diff", "html_url": "https://github.com/huggingface/datasets/pull/4108", "merged_at": "2022-04-20T12:54:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/4108.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4108" }
This PR makes necessary changes to perplexity such that: - it runs much faster (via batching) - it throws an error when input is empty, or when input is one word without <BOS> token - it adds the option to add a <BOS> token Issues: - The values returned are extremely high, and I'm worried they aren't correct. Even if they are correct, they are sometimes returned as `inf`, which is not very useful (see [comment below](https://github.com/huggingface/datasets/pull/4108#discussion_r843931094) for some of the output values). - If the values are not correct, can you help me find the error? - If the values are correct, it might be worth it to measure something like perplexity per word, which would allow us to get actual values for the larger perplexities, instead of just `inf` Future: - `stride` is not currently implemented here. I have some thoughts on how to make it happen with batching, but I think it would be better to get another set of eyes to look at any possible errors causing such large values now rather than later.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4108/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4108/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4107
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4107/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4107/comments
https://api.github.com/repos/huggingface/datasets/issues/4107/events
https://github.com/huggingface/datasets/issues/4107
1,194,484,885
I_kwDODunzps5HMmSV
4,107
Unable to view the dataset and loading the same dataset throws the error - ArrowInvalid: Exceeded maximum rows
{ "avatar_url": "https://avatars.githubusercontent.com/u/23344465?v=4", "events_url": "https://api.github.com/users/Pavithree/events{/privacy}", "followers_url": "https://api.github.com/users/Pavithree/followers", "following_url": "https://api.github.com/users/Pavithree/following{/other_user}", "gists_url": "https://api.github.com/users/Pavithree/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Pavithree", "id": 23344465, "login": "Pavithree", "node_id": "MDQ6VXNlcjIzMzQ0NDY1", "organizations_url": "https://api.github.com/users/Pavithree/orgs", "received_events_url": "https://api.github.com/users/Pavithree/received_events", "repos_url": "https://api.github.com/users/Pavithree/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Pavithree/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Pavithree/subscriptions", "type": "User", "url": "https://api.github.com/users/Pavithree" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-04-06T11:37:15Z
2022-04-08T07:13:07Z
2022-04-06T14:39:55Z
NONE
null
null
null
## Dataset viewer issue - -ArrowInvalid: Exceeded maximum rows **Link:** *https://huggingface.co/datasets/Pavithree/explainLikeImFive* *This is the subset of original eli5 dataset https://huggingface.co/datasets/vblagoje/lfqa. I just filtered the data samples which belongs to one particular subreddit thread. However, the dataset preview for train split returns the below mentioned error: Status code: 400 Exception: ArrowInvalid Message: Exceeded maximum rows When I try to load the same dataset it returns ArrowInvalid: Exceeded maximum rows error* Am I the one who added this dataset ? Yes
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4107/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4107/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4106
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4106/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4106/comments
https://api.github.com/repos/huggingface/datasets/issues/4106/events
https://github.com/huggingface/datasets/pull/4106
1,194,393,892
PR_kwDODunzps41uPpa
4,106
Support huggingface_hub 0.5
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2022-04-06T10:15:25Z
2022-04-08T10:28:43Z
2022-04-08T10:22:23Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4106.diff", "html_url": "https://github.com/huggingface/datasets/pull/4106", "merged_at": "2022-04-08T10:22:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/4106.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4106" }
Following https://github.com/huggingface/datasets/issues/4105 `huggingface_hub` deprecated some parameters in `HfApi` in 0.5. This PR updates all the calls to HfApi to remove all the deprecations, <s>and I set the `hugginface_hub` requirement to `>=0.5.0`</s> cc @adrinjalali @LysandreJik
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4106/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4106/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4105
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4105/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4105/comments
https://api.github.com/repos/huggingface/datasets/issues/4105/events
https://github.com/huggingface/datasets/issues/4105
1,194,297,119
I_kwDODunzps5HL4cf
4,105
push to hub fails with huggingface-hub 0.5.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/2518789?v=4", "events_url": "https://api.github.com/users/frascuchon/events{/privacy}", "followers_url": "https://api.github.com/users/frascuchon/followers", "following_url": "https://api.github.com/users/frascuchon/following{/other_user}", "gists_url": "https://api.github.com/users/frascuchon/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/frascuchon", "id": 2518789, "login": "frascuchon", "node_id": "MDQ6VXNlcjI1MTg3ODk=", "organizations_url": "https://api.github.com/users/frascuchon/orgs", "received_events_url": "https://api.github.com/users/frascuchon/received_events", "repos_url": "https://api.github.com/users/frascuchon/repos", "site_admin": false, "starred_url": "https://api.github.com/users/frascuchon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/frascuchon/subscriptions", "type": "User", "url": "https://api.github.com/users/frascuchon" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-04-06T08:59:57Z
2022-04-13T14:30:47Z
2022-04-13T14:30:47Z
NONE
null
null
null
## Describe the bug `ds.push_to_hub` is failing when updating a dataset in the form "org_id/repo_id" ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_dataset("rubrix/news_test") ds.push_to_hub("<your-user>/news_test", token="<your-token>") ``` ## Expected results The dataset is successfully uploaded ## Actual results An error validation is raised: ```bash if repo_id and (name or organization): > raise ValueError( "Only pass `repo_id` and leave deprecated `name` and " "`organization` to be None." E ValueError: Only pass `repo_id` and leave deprecated `name` and `organization` to be None. ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.18.1 - `huggingface-hub`: 0.5 - Platform: macOS - Python version: 3.8.12 - PyArrow version: 6.0.0 cc @adrinjalali
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4105/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4105/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4104
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4104/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4104/comments
https://api.github.com/repos/huggingface/datasets/issues/4104/events
https://github.com/huggingface/datasets/issues/4104
1,194,072,966
I_kwDODunzps5HLBuG
4,104
Add time series data - stock market
{ "avatar_url": "https://avatars.githubusercontent.com/u/45640029?v=4", "events_url": "https://api.github.com/users/INF800/events{/privacy}", "followers_url": "https://api.github.com/users/INF800/followers", "following_url": "https://api.github.com/users/INF800/following{/other_user}", "gists_url": "https://api.github.com/users/INF800/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/INF800", "id": 45640029, "login": "INF800", "node_id": "MDQ6VXNlcjQ1NjQwMDI5", "organizations_url": "https://api.github.com/users/INF800/orgs", "received_events_url": "https://api.github.com/users/INF800/received_events", "repos_url": "https://api.github.com/users/INF800/repos", "site_admin": false, "starred_url": "https://api.github.com/users/INF800/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/INF800/subscriptions", "type": "User", "url": "https://api.github.com/users/INF800" }
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
open
false
null
[]
null
[]
2022-04-06T05:46:58Z
2022-04-11T09:07:10Z
null
NONE
null
null
null
## Adding a Time Series Dataset - **Name:** 2min ticker data for stock market - **Description:** 8 stocks' data collected for 1month post ukraine-russia war. 4 NSE stocks and 4 NASDAQ stocks. Along with technical indicators (additional features) as shown in below image - **Data:** Collected by myself from investing.com - **Motivation:** Test applicability of transformer based model on stock market / time series problem ![image](https://user-images.githubusercontent.com/45640029/161904077-52fe97cb-3720-4e3f-98ee-7f6720a056e2.png)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4104/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4104/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4103
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4103/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4103/comments
https://api.github.com/repos/huggingface/datasets/issues/4103/events
https://github.com/huggingface/datasets/pull/4103
1,193,987,104
PR_kwDODunzps41s3T4
4,103
Add the `GSM8K` dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/41410219?v=4", "events_url": "https://api.github.com/users/jon-tow/events{/privacy}", "followers_url": "https://api.github.com/users/jon-tow/followers", "following_url": "https://api.github.com/users/jon-tow/following{/other_user}", "gists_url": "https://api.github.com/users/jon-tow/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jon-tow", "id": 41410219, "login": "jon-tow", "node_id": "MDQ6VXNlcjQxNDEwMjE5", "organizations_url": "https://api.github.com/users/jon-tow/orgs", "received_events_url": "https://api.github.com/users/jon-tow/received_events", "repos_url": "https://api.github.com/users/jon-tow/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jon-tow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jon-tow/subscriptions", "type": "User", "url": "https://api.github.com/users/jon-tow" }
[]
closed
false
null
[]
null
[]
2022-04-06T04:07:52Z
2022-04-12T15:38:28Z
2022-04-12T10:21:16Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4103.diff", "html_url": "https://github.com/huggingface/datasets/pull/4103", "merged_at": "2022-04-12T10:21:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/4103.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4103" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4103/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4103/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4102
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4102/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4102/comments
https://api.github.com/repos/huggingface/datasets/issues/4102/events
https://github.com/huggingface/datasets/pull/4102
1,193,616,722
PR_kwDODunzps41roGx
4,102
[hub] Fix `api.create_repo` call?
{ "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "events_url": "https://api.github.com/users/julien-c/events{/privacy}", "followers_url": "https://api.github.com/users/julien-c/followers", "following_url": "https://api.github.com/users/julien-c/following{/other_user}", "gists_url": "https://api.github.com/users/julien-c/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/julien-c", "id": 326577, "login": "julien-c", "node_id": "MDQ6VXNlcjMyNjU3Nw==", "organizations_url": "https://api.github.com/users/julien-c/orgs", "received_events_url": "https://api.github.com/users/julien-c/received_events", "repos_url": "https://api.github.com/users/julien-c/repos", "site_admin": false, "starred_url": "https://api.github.com/users/julien-c/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/julien-c/subscriptions", "type": "User", "url": "https://api.github.com/users/julien-c" }
[]
closed
false
null
[]
null
[]
2022-04-05T19:21:52Z
2022-04-12T08:41:46Z
2022-04-12T08:41:46Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4102.diff", "html_url": "https://github.com/huggingface/datasets/pull/4102", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4102.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4102" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4102/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4102/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4101
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4101/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4101/comments
https://api.github.com/repos/huggingface/datasets/issues/4101/events
https://github.com/huggingface/datasets/issues/4101
1,193,399,204
I_kwDODunzps5HIdOk
4,101
How can I download only the train and test split for full numbers using load_dataset()?
{ "avatar_url": "https://avatars.githubusercontent.com/u/64383902?v=4", "events_url": "https://api.github.com/users/Nakkhatra/events{/privacy}", "followers_url": "https://api.github.com/users/Nakkhatra/followers", "following_url": "https://api.github.com/users/Nakkhatra/following{/other_user}", "gists_url": "https://api.github.com/users/Nakkhatra/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Nakkhatra", "id": 64383902, "login": "Nakkhatra", "node_id": "MDQ6VXNlcjY0MzgzOTAy", "organizations_url": "https://api.github.com/users/Nakkhatra/orgs", "received_events_url": "https://api.github.com/users/Nakkhatra/received_events", "repos_url": "https://api.github.com/users/Nakkhatra/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Nakkhatra/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nakkhatra/subscriptions", "type": "User", "url": "https://api.github.com/users/Nakkhatra" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
2022-04-05T16:00:15Z
2022-04-06T13:09:01Z
null
NONE
null
null
null
How can I download only the train and test split for full numbers using load_dataset()? I do not need the extra split and it will take 40 mins just to download in Colab. I have very short time in hand. Please help.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4101/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4101/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4100
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4100/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4100/comments
https://api.github.com/repos/huggingface/datasets/issues/4100/events
https://github.com/huggingface/datasets/pull/4100
1,193,393,959
PR_kwDODunzps41q4ce
4,100
Improve RedCaps dataset card
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[]
2022-04-05T15:57:14Z
2022-04-13T14:08:54Z
2022-04-13T14:02:26Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4100.diff", "html_url": "https://github.com/huggingface/datasets/pull/4100", "merged_at": "2022-04-13T14:02:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/4100.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4100" }
This PR modifies the RedCaps card to: * fix the formatting of the Point of Contact fields on the Hub * speed up the image fetching logic (aligns it with the [img2dataset](https://github.com/rom1504/img2dataset) tool) and make it more robust (return None if **any** exception is thrown)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4100/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4100/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4099
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4099/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4099/comments
https://api.github.com/repos/huggingface/datasets/issues/4099/events
https://github.com/huggingface/datasets/issues/4099
1,193,253,768
I_kwDODunzps5HH5uI
4,099
UnicodeDecodeError: 'ascii' codec can't decode byte 0xe5 in position 213: ordinal not in range(128)
{ "avatar_url": "https://avatars.githubusercontent.com/u/20210017?v=4", "events_url": "https://api.github.com/users/andreybond/events{/privacy}", "followers_url": "https://api.github.com/users/andreybond/followers", "following_url": "https://api.github.com/users/andreybond/following{/other_user}", "gists_url": "https://api.github.com/users/andreybond/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/andreybond", "id": 20210017, "login": "andreybond", "node_id": "MDQ6VXNlcjIwMjEwMDE3", "organizations_url": "https://api.github.com/users/andreybond/orgs", "received_events_url": "https://api.github.com/users/andreybond/received_events", "repos_url": "https://api.github.com/users/andreybond/repos", "site_admin": false, "starred_url": "https://api.github.com/users/andreybond/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andreybond/subscriptions", "type": "User", "url": "https://api.github.com/users/andreybond" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-05T14:42:38Z
2022-04-06T06:37:44Z
2022-04-06T06:35:54Z
NONE
null
null
null
## Describe the bug Error "UnicodeDecodeError: 'ascii' codec can't decode byte 0xe5 in position 213: ordinal not in range(128)" is thrown when downloading dataset. ## Steps to reproduce the bug ```python from datasets import load_dataset datasets = load_dataset("nielsr/XFUN", "xfun.ja") ``` ## Expected results Dataset should be downloaded without exceptions ## Actual results Stack trace (for the second-time execution): Downloading and preparing dataset xfun/xfun.ja to /root/.cache/huggingface/datasets/nielsr___xfun/xfun.ja/0.0.0/e06e948b673d1be9a390a83c05c10e49438bf03dd85ae9a4fe06f8747a724477... Downloading data files: 100% 2/2 [00:00<00:00, 88.48it/s] Extracting data files: 100% 2/2 [00:00<00:00, 79.60it/s] UnicodeDecodeErrorTraceback (most recent call last) <ipython-input-31-79c26bd1109c> in <module> 1 from datasets import load_dataset 2 ----> 3 datasets = load_dataset("nielsr/XFUN", "xfun.ja") /usr/local/lib/python3.6/dist-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) /usr/local/lib/python3.6/dist-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs) 604 ) 605 --> 606 # By default, return all splits 607 if split is None: 608 split = {s: s for s in self.info.splits} /usr/local/lib/python3.6/dist-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos) /usr/local/lib/python3.6/dist-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 692 Args: 693 split: `datasets.Split` which subset of the data to read. --> 694 695 Returns: 696 `Dataset` /usr/local/lib/python3.6/dist-packages/datasets/builder.py in _prepare_split(self, split_generator, check_duplicate_keys) /usr/local/lib/python3.6/dist-packages/tqdm/notebook.py in __iter__(self) 252 if not self.disable: 253 self.display(check_delay=False) --> 254 255 def __iter__(self): 256 try: /usr/local/lib/python3.6/dist-packages/tqdm/std.py in __iter__(self) 1183 for obj in iterable: 1184 yield obj -> 1185 return 1186 1187 mininterval = self.mininterval ~/.cache/huggingface/modules/datasets_modules/datasets/nielsr--XFUN/e06e948b673d1be9a390a83c05c10e49438bf03dd85ae9a4fe06f8747a724477/XFUN.py in _generate_examples(self, filepaths) 140 logger.info("Generating examples from = %s", filepath) 141 with open(filepath[0], "r") as f: --> 142 data = json.load(f) 143 144 for doc in data["documents"]: /usr/lib/python3.6/json/__init__.py in load(fp, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw) 294 295 """ --> 296 return loads(fp.read(), 297 cls=cls, object_hook=object_hook, 298 parse_float=parse_float, parse_int=parse_int, /usr/lib/python3.6/encodings/ascii.py in decode(self, input, final) 24 class IncrementalDecoder(codecs.IncrementalDecoder): 25 def decode(self, input, final=False): ---> 26 return codecs.ascii_decode(input, self.errors)[0] 27 28 class StreamWriter(Codec,codecs.StreamWriter): UnicodeDecodeError: 'ascii' codec can't decode byte 0xe5 in position 213: ordinal not in range(128) ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.0.0 (but reproduced with many previous versions) - Platform: Docker: Linux da5b74136d6b 5.3.0-1031-azure #32~18.04.1-Ubuntu SMP Mon Jun 22 15:27:23 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux ; Base docker image is : huggingface/transformers-pytorch-cpu - Python version: 3.6.9 - PyArrow version: 6.0.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4099/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4099/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4098
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4098/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4098/comments
https://api.github.com/repos/huggingface/datasets/issues/4098/events
https://github.com/huggingface/datasets/pull/4098
1,193,245,522
PR_kwDODunzps41qXjo
4,098
Proposing WikiSplit metric card
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
closed
false
null
[]
null
[]
2022-04-05T14:36:34Z
2022-10-11T09:10:21Z
2022-04-05T15:42:28Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4098.diff", "html_url": "https://github.com/huggingface/datasets/pull/4098", "merged_at": "2022-04-05T15:42:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/4098.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4098" }
Pinging @lhoestq to ensure that my distinction between the dataset and the metric are clear :sweat_smile:
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4098/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4098/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4097
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4097/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4097/comments
https://api.github.com/repos/huggingface/datasets/issues/4097/events
https://github.com/huggingface/datasets/pull/4097
1,193,205,751
PR_kwDODunzps41qPEu
4,097
Updating FrugalScore metric card
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
closed
false
null
[]
null
[]
2022-04-05T14:09:24Z
2022-04-05T15:07:35Z
2022-04-05T15:01:46Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4097.diff", "html_url": "https://github.com/huggingface/datasets/pull/4097", "merged_at": "2022-04-05T15:01:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/4097.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4097" }
removing duplicate paragraph
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4097/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4097/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4096
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4096/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4096/comments
https://api.github.com/repos/huggingface/datasets/issues/4096/events
https://github.com/huggingface/datasets/issues/4096
1,193,165,229
I_kwDODunzps5HHkGt
4,096
Add support for streaming Zarr stores for hosted datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/7170359?v=4", "events_url": "https://api.github.com/users/jacobbieker/events{/privacy}", "followers_url": "https://api.github.com/users/jacobbieker/followers", "following_url": "https://api.github.com/users/jacobbieker/following{/other_user}", "gists_url": "https://api.github.com/users/jacobbieker/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jacobbieker", "id": 7170359, "login": "jacobbieker", "node_id": "MDQ6VXNlcjcxNzAzNTk=", "organizations_url": "https://api.github.com/users/jacobbieker/orgs", "received_events_url": "https://api.github.com/users/jacobbieker/received_events", "repos_url": "https://api.github.com/users/jacobbieker/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jacobbieker/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jacobbieker/subscriptions", "type": "User", "url": "https://api.github.com/users/jacobbieker" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-05T13:38:32Z
2022-04-25T08:04:12Z
2022-04-21T08:12:58Z
NONE
null
null
null
**Is your feature request related to a problem? Please describe.** Lots of geospatial data is stored in the Zarr format. This format works well for n-dimensional data and coordinates, and can have good compression. Unfortunately, HF datasets doesn't support streaming in data in Zarr format as far as I can tell. Zarr stores are designed to be easily streamed in from cloud storage, especially with xarray and fsspec. Since geospatial data tends to be very large, and on the order of TBs of data or 10's of TBs of data for a single dataset, it can be difficult to store the dataset locally for users. Just adding Zarr stores with HF git doesn't work well (see https://github.com/huggingface/datasets/issues/3823) as Zarr splits the data into lots of small chunks for fast loading, and that doesn't work well with git. I've somewhat gotten around that issue by tarring each Zarr store and uploading them as a single file, which seems to be working (see https://huggingface.co/datasets/openclimatefix/gfs-reforecast for example data files, although the script isn't written yet). This does mean that streaming doesn't quite work though. On the other hand, in https://huggingface.co/datasets/openclimatefix/eumetsat_uk_hrv we stream in a Zarr store from a public GCP bucket quite easily. **Describe the solution you'd like** A way to upload Zarr stores for hosted datasets so that we can stream it with xarray and fsspec. **Describe alternatives you've considered** Tarring each Zarr store individually and just extracting them in the dataset script -> Downside this is a lot of data that probably doesn't fit locally for a lot of potential users. Pre-prepare examples in a format like Parquet -> Would use a lot more storage, and a lot less flexibility, in the eumetsat_uk_hrv, we use the one Zarr store for multiple different configurations.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4096/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4096/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4095
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4095/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4095/comments
https://api.github.com/repos/huggingface/datasets/issues/4095/events
https://github.com/huggingface/datasets/pull/4095
1,192,573,353
PR_kwDODunzps41oIFI
4,095
fix typo in rename_column error message
{ "avatar_url": "https://avatars.githubusercontent.com/u/680821?v=4", "events_url": "https://api.github.com/users/hunterlang/events{/privacy}", "followers_url": "https://api.github.com/users/hunterlang/followers", "following_url": "https://api.github.com/users/hunterlang/following{/other_user}", "gists_url": "https://api.github.com/users/hunterlang/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hunterlang", "id": 680821, "login": "hunterlang", "node_id": "MDQ6VXNlcjY4MDgyMQ==", "organizations_url": "https://api.github.com/users/hunterlang/orgs", "received_events_url": "https://api.github.com/users/hunterlang/received_events", "repos_url": "https://api.github.com/users/hunterlang/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hunterlang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hunterlang/subscriptions", "type": "User", "url": "https://api.github.com/users/hunterlang" }
[]
closed
false
null
[]
null
[]
2022-04-05T03:55:56Z
2022-04-05T08:54:46Z
2022-04-05T08:45:53Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4095.diff", "html_url": "https://github.com/huggingface/datasets/pull/4095", "merged_at": "2022-04-05T08:45:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/4095.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4095" }
I feel bad submitting such a tiny change as a PR but it confused me today 😄
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4095/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4095/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4094
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4094/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4094/comments
https://api.github.com/repos/huggingface/datasets/issues/4094/events
https://github.com/huggingface/datasets/issues/4094
1,192,534,414
I_kwDODunzps5HFKGO
4,094
Helo Mayfrends
{ "avatar_url": "https://avatars.githubusercontent.com/u/102933353?v=4", "events_url": "https://api.github.com/users/Budigming/events{/privacy}", "followers_url": "https://api.github.com/users/Budigming/followers", "following_url": "https://api.github.com/users/Budigming/following{/other_user}", "gists_url": "https://api.github.com/users/Budigming/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Budigming", "id": 102933353, "login": "Budigming", "node_id": "U_kgDOBiKjaQ", "organizations_url": "https://api.github.com/users/Budigming/orgs", "received_events_url": "https://api.github.com/users/Budigming/received_events", "repos_url": "https://api.github.com/users/Budigming/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Budigming/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Budigming/subscriptions", "type": "User", "url": "https://api.github.com/users/Budigming" }
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
[]
null
[]
2022-04-05T02:42:57Z
2022-04-05T07:16:42Z
2022-04-05T07:16:42Z
NONE
null
null
null
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons to have this dataset* Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4094/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4094/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4093
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4093/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4093/comments
https://api.github.com/repos/huggingface/datasets/issues/4093/events
https://github.com/huggingface/datasets/issues/4093
1,192,523,161
I_kwDODunzps5HFHWZ
4,093
elena-soare/crawled-ecommerce: missing dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/17519354?v=4", "events_url": "https://api.github.com/users/seevaratnam/events{/privacy}", "followers_url": "https://api.github.com/users/seevaratnam/followers", "following_url": "https://api.github.com/users/seevaratnam/following{/other_user}", "gists_url": "https://api.github.com/users/seevaratnam/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/seevaratnam", "id": 17519354, "login": "seevaratnam", "node_id": "MDQ6VXNlcjE3NTE5MzU0", "organizations_url": "https://api.github.com/users/seevaratnam/orgs", "received_events_url": "https://api.github.com/users/seevaratnam/received_events", "repos_url": "https://api.github.com/users/seevaratnam/repos", "site_admin": false, "starred_url": "https://api.github.com/users/seevaratnam/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/seevaratnam/subscriptions", "type": "User", "url": "https://api.github.com/users/seevaratnam" }
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" } ]
null
[]
2022-04-05T02:25:19Z
2022-04-12T09:34:53Z
2022-04-12T09:34:53Z
NONE
null
null
null
elena-soare/crawled-ecommerce **Link:** *link to the dataset viewer page* *short description of the issue* Am I the one who added this dataset ? Yes-No
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4093/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4093/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4092
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4092/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4092/comments
https://api.github.com/repos/huggingface/datasets/issues/4092/events
https://github.com/huggingface/datasets/pull/4092
1,192,499,903
PR_kwDODunzps41n40R
4,092
Fix dataset `amazon_us_reviews` metadata - 4/4/2022
{ "avatar_url": "https://avatars.githubusercontent.com/u/191985?v=4", "events_url": "https://api.github.com/users/trentonstrong/events{/privacy}", "followers_url": "https://api.github.com/users/trentonstrong/followers", "following_url": "https://api.github.com/users/trentonstrong/following{/other_user}", "gists_url": "https://api.github.com/users/trentonstrong/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/trentonstrong", "id": 191985, "login": "trentonstrong", "node_id": "MDQ6VXNlcjE5MTk4NQ==", "organizations_url": "https://api.github.com/users/trentonstrong/orgs", "received_events_url": "https://api.github.com/users/trentonstrong/received_events", "repos_url": "https://api.github.com/users/trentonstrong/repos", "site_admin": false, "starred_url": "https://api.github.com/users/trentonstrong/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/trentonstrong/subscriptions", "type": "User", "url": "https://api.github.com/users/trentonstrong" }
[]
closed
false
null
[]
null
[]
2022-04-05T01:39:45Z
2022-04-08T12:35:41Z
2022-04-08T12:29:31Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4092.diff", "html_url": "https://github.com/huggingface/datasets/pull/4092", "merged_at": "2022-04-08T12:29:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/4092.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4092" }
Fixes #4048 by running `dataset-cli test` to reprocess data and regenerate metadata. Additionally I've updated the README to include up-to-date counts for the subsets.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4092/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4092/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4091
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4091/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4091/comments
https://api.github.com/repos/huggingface/datasets/issues/4091/events
https://github.com/huggingface/datasets/issues/4091
1,192,023,855
I_kwDODunzps5HDNcv
4,091
Build a Dataset One Example at a Time Without Loading All Data Into Memory
{ "avatar_url": "https://avatars.githubusercontent.com/u/99340348?v=4", "events_url": "https://api.github.com/users/aravind-tonita/events{/privacy}", "followers_url": "https://api.github.com/users/aravind-tonita/followers", "following_url": "https://api.github.com/users/aravind-tonita/following{/other_user}", "gists_url": "https://api.github.com/users/aravind-tonita/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/aravind-tonita", "id": 99340348, "login": "aravind-tonita", "node_id": "U_kgDOBevQPA", "organizations_url": "https://api.github.com/users/aravind-tonita/orgs", "received_events_url": "https://api.github.com/users/aravind-tonita/received_events", "repos_url": "https://api.github.com/users/aravind-tonita/repos", "site_admin": false, "starred_url": "https://api.github.com/users/aravind-tonita/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aravind-tonita/subscriptions", "type": "User", "url": "https://api.github.com/users/aravind-tonita" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[]
2022-04-04T16:19:24Z
2022-04-20T14:31:00Z
2022-04-20T14:31:00Z
NONE
null
null
null
**Is your feature request related to a problem? Please describe.** I have a very large dataset stored on disk in a custom format. I have some custom code that reads one data example at a time and yields it in the form of a dictionary. I want to construct a `Dataset` with all examples, and then save it to disk. I later want to load the saved `Dataset` and use it like any other HuggingFace dataset, get splits, wrap it in a PyTorch `DataLoader`, etc. **Crucially, I do not ever want to materialize all the data in memory while building the dataset.** **Describe the solution you'd like** I would like to be able to do something like the following. Notice how each example is read and then immediately added to the dataset. We do not store all the data in memory when constructing the `Dataset`. If it helps, I will know the schema of my dataset before hand. ``` # Initialize an empty Dataset, possibly from a known schema. dataset = Dataset() # Read in examples one by one using a custom data streamer. for example_dict in custom_example_dict_streamer("/path/to/raw/data"): # Add this example to the dict but do not store it in memory. dataset.add_item(example_dict) # Save the final dataset to disk as an Arrow-backed dataset. dataset.save_to_disk("/path/to/dataset") ... # I'd like to be able to later `load_from_disk` and use the loaded Dataset # just like any other memory-mapped pyarrow-backed HuggingFace dataset... loaded_dataset = Dataset.load_from_disk("/path/to/dataset") loaded_dataset.set_format(type="torch", columnns=["foo", "bar", "baz"]) dataloader = torch.utils.data.DataLoader(loaded_dataset, batch_size=16) ... ``` **Describe alternatives you've considered** I initially tried to read all the data into memory, construct a Pandas DataFrame and then call `Dataset.from_pandas`. This would not work as it requires storing all the data in memory. It seems that there is an `add_item` method already -- I tried to implement something like the desired API written above, but I've not been able to initialize an empty `Dataset` (this seems to require several layers of constructing `datasets.table.Table` which requires constructing a `pyarrow.lib.Table`, etc). I also considered writing my data to multiple sharded CSV files or JSON files and then using `from_csv` or `from_json`. I'd prefer not to do this because (1) I'd prefer to avoid the intermediate step of creating these temp CSV/JSON files and (2) I'm not sure if `from_csv` and `from_json` use memory-mapping. Do you have any suggestions on how I'd be able to achieve this use case? Does something already exist to support this? Thank you very much in advance!
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4091/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4091/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4090
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4090/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4090/comments
https://api.github.com/repos/huggingface/datasets/issues/4090/events
https://github.com/huggingface/datasets/pull/4090
1,191,956,734
PR_kwDODunzps41mEs5
4,090
Avoid writing empty license files
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-04T15:23:37Z
2022-04-07T12:46:45Z
2022-04-07T12:40:43Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4090.diff", "html_url": "https://github.com/huggingface/datasets/pull/4090", "merged_at": "2022-04-07T12:40:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/4090.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4090" }
This PR avoids the creation of empty `LICENSE` files.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4090/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4090/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4089
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4089/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4089/comments
https://api.github.com/repos/huggingface/datasets/issues/4089/events
https://github.com/huggingface/datasets/pull/4089
1,191,915,196
PR_kwDODunzps41l7yd
4,089
Create metric card for Frugal Score
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
closed
false
null
[]
null
[]
2022-04-04T14:53:49Z
2022-04-05T14:14:46Z
2022-04-05T14:06:50Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4089.diff", "html_url": "https://github.com/huggingface/datasets/pull/4089", "merged_at": "2022-04-05T14:06:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/4089.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4089" }
Proposing metric card for Frugal Score. @albertvillanova or @lhoestq -- there are certain aspects that I'm not 100% sure on (such as how exactly the distillation between BertScore and FrugalScore is done) -- so if you find that something isn't clear, please let me know!
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4089/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4089/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4088
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4088/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4088/comments
https://api.github.com/repos/huggingface/datasets/issues/4088/events
https://github.com/huggingface/datasets/pull/4088
1,191,901,172
PR_kwDODunzps41l4yE
4,088
Remove unused legacy Beam utils
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-04T14:43:51Z
2022-04-05T15:23:27Z
2022-04-05T15:17:41Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4088.diff", "html_url": "https://github.com/huggingface/datasets/pull/4088", "merged_at": "2022-04-05T15:17:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/4088.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4088" }
This PR removes unused legacy custom `WriteToParquet`, once official Apache Beam includes the patch since version 2.22.0: - Patch PR: https://github.com/apache/beam/pull/11699 - Issue: https://issues.apache.org/jira/browse/BEAM-10022 In relation with: - #204
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4088/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4088/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4087
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4087/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4087/comments
https://api.github.com/repos/huggingface/datasets/issues/4087/events
https://github.com/huggingface/datasets/pull/4087
1,191,819,805
PR_kwDODunzps41lnfO
4,087
Fix BeamWriter output Parquet file
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-04T13:46:50Z
2022-04-05T15:00:40Z
2022-04-05T14:54:48Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4087.diff", "html_url": "https://github.com/huggingface/datasets/pull/4087", "merged_at": "2022-04-05T14:54:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/4087.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4087" }
Since now, the `BeamWriter` saved a Parquet file with a simplified schema, where each field value was serialized to JSON. That resulted in Parquet files larger than Arrow files. This PR: - writes Parquet file preserving original schema and without serialization, thus avoiding serialization overhead and resulting in a smaller output file size. - fixes `parquet_to_arrow` function
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4087/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4087/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4086
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4086/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4086/comments
https://api.github.com/repos/huggingface/datasets/issues/4086/events
https://github.com/huggingface/datasets/issues/4086
1,191,373,374
I_kwDODunzps5HAuo-
4,086
Dataset viewer issue for McGill-NLP/feedbackQA
{ "avatar_url": "https://avatars.githubusercontent.com/u/54827718?v=4", "events_url": "https://api.github.com/users/cslizc/events{/privacy}", "followers_url": "https://api.github.com/users/cslizc/followers", "following_url": "https://api.github.com/users/cslizc/following{/other_user}", "gists_url": "https://api.github.com/users/cslizc/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cslizc", "id": 54827718, "login": "cslizc", "node_id": "MDQ6VXNlcjU0ODI3NzE4", "organizations_url": "https://api.github.com/users/cslizc/orgs", "received_events_url": "https://api.github.com/users/cslizc/received_events", "repos_url": "https://api.github.com/users/cslizc/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cslizc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cslizc/subscriptions", "type": "User", "url": "https://api.github.com/users/cslizc" }
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-04T07:27:20Z
2022-04-04T22:29:53Z
2022-04-04T08:01:45Z
NONE
null
null
null
## Dataset viewer issue for '*McGill-NLP/feedbackQA*' **Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/McGill-NLP/feedbackQA)* *short description of the issue* The dataset can be loaded correctly with `load_dataset` but the preview doesn't work. Error message: ``` Status code: 400 Exception: Status400Error Message: Not found. Maybe the cache is missing, or maybe the dataset does not exist. ``` Am I the one who added this dataset ? Yes
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4086/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4086/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4085
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4085/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4085/comments
https://api.github.com/repos/huggingface/datasets/issues/4085/events
https://github.com/huggingface/datasets/issues/4085
1,190,621,345
I_kwDODunzps5G93Ch
4,085
datasets.set_progress_bar_enabled(False) not working in datasets v2
{ "avatar_url": "https://avatars.githubusercontent.com/u/3381112?v=4", "events_url": "https://api.github.com/users/virilo/events{/privacy}", "followers_url": "https://api.github.com/users/virilo/followers", "following_url": "https://api.github.com/users/virilo/following{/other_user}", "gists_url": "https://api.github.com/users/virilo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/virilo", "id": 3381112, "login": "virilo", "node_id": "MDQ6VXNlcjMzODExMTI=", "organizations_url": "https://api.github.com/users/virilo/orgs", "received_events_url": "https://api.github.com/users/virilo/received_events", "repos_url": "https://api.github.com/users/virilo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/virilo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/virilo/subscriptions", "type": "User", "url": "https://api.github.com/users/virilo" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-02T12:40:10Z
2022-09-17T02:18:03Z
2022-04-04T06:44:34Z
NONE
null
null
null
## Describe the bug datasets.set_progress_bar_enabled(False) not working in datasets v2 ## Steps to reproduce the bug ```python datasets.set_progress_bar_enabled(False) ``` ## Expected results datasets not using any progress bar ## Actual results AttributeError: module 'datasets' has no attribute 'set_progress_bar_enabled ## Environment info datasets version 2
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4085/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4085/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4084
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4084/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4084/comments
https://api.github.com/repos/huggingface/datasets/issues/4084/events
https://github.com/huggingface/datasets/issues/4084
1,190,060,415
I_kwDODunzps5G7uF_
4,084
Errors in `Train with Datasets` Tensorflow code section on Huggingface.co
{ "avatar_url": "https://avatars.githubusercontent.com/u/57095771?v=4", "events_url": "https://api.github.com/users/blackhat-coder/events{/privacy}", "followers_url": "https://api.github.com/users/blackhat-coder/followers", "following_url": "https://api.github.com/users/blackhat-coder/following{/other_user}", "gists_url": "https://api.github.com/users/blackhat-coder/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/blackhat-coder", "id": 57095771, "login": "blackhat-coder", "node_id": "MDQ6VXNlcjU3MDk1Nzcx", "organizations_url": "https://api.github.com/users/blackhat-coder/orgs", "received_events_url": "https://api.github.com/users/blackhat-coder/received_events", "repos_url": "https://api.github.com/users/blackhat-coder/repos", "site_admin": false, "starred_url": "https://api.github.com/users/blackhat-coder/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/blackhat-coder/subscriptions", "type": "User", "url": "https://api.github.com/users/blackhat-coder" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-01T17:02:47Z
2022-04-04T07:24:37Z
2022-04-04T07:21:31Z
NONE
null
null
null
## Describe the bug Hi ### Error 1 Running the Tensforlow code on [Huggingface](https://huggingface.co/docs/datasets/use_dataset) gives a TypeError: __init__() got an unexpected keyword argument 'return_tensors' ### Error 2 `DataCollatorWithPadding` isn't imported ## Steps to reproduce the bug ```python import tensorflow as tf from datasets import load_dataset from transformers import AutoTokenizer dataset = load_dataset('glue', 'mrpc', split='train') tokenizer = AutoTokenizer.from_pretrained('bert-base-cased') dataset = dataset.map(lambda e: tokenizer(e['sentence1'], truncation=True, padding='max_length'), batched=True) data_collator = DataCollatorWithPadding(tokenizer=tokenizer, return_tensors="tf") train_dataset = dataset["train"].to_tf_dataset( columns=['input_ids', 'token_type_ids', 'attention_mask', 'label'], shuffle=True, batch_size=16, collate_fn=data_collator, ) ``` This is the same code on Huggingface.co ## Actual results TypeError: __init__() got an unexpected keyword argument 'return_tensors' ## Environment info - `datasets` version: 2.0.0 - Platform: Windows-10-10.0.19044-SP0 - Python version: 3.9.7 - PyArrow version: 6.0.0 - Pandas version: 1.4.1 >
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4084/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4084/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4083
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4083/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4083/comments
https://api.github.com/repos/huggingface/datasets/issues/4083/events
https://github.com/huggingface/datasets/pull/4083
1,190,025,878
PR_kwDODunzps41gEbu
4,083
Add SacreBLEU Metric Card
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
closed
false
null
[]
null
[]
2022-04-01T16:24:56Z
2022-04-12T20:45:00Z
2022-04-12T20:38:40Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4083.diff", "html_url": "https://github.com/huggingface/datasets/pull/4083", "merged_at": "2022-04-12T20:38:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/4083.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4083" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4083/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4083/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4082
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4082/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4082/comments
https://api.github.com/repos/huggingface/datasets/issues/4082/events
https://github.com/huggingface/datasets/pull/4082
1,189,965,845
PR_kwDODunzps41f3fb
4,082
Add chrF(++) Metric Card
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
closed
false
null
[]
null
[]
2022-04-01T15:32:12Z
2022-04-12T20:43:55Z
2022-04-12T20:38:06Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4082.diff", "html_url": "https://github.com/huggingface/datasets/pull/4082", "merged_at": "2022-04-12T20:38:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/4082.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4082" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4082/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4082/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4081
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4081/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4081/comments
https://api.github.com/repos/huggingface/datasets/issues/4081/events
https://github.com/huggingface/datasets/pull/4081
1,189,916,472
PR_kwDODunzps41fsxW
4,081
Close parquet writer properly in `push_to_hub`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2022-04-01T14:58:50Z
2022-07-14T19:22:06Z
2022-04-01T16:16:19Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4081.diff", "html_url": "https://github.com/huggingface/datasets/pull/4081", "merged_at": "2022-04-01T16:16:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/4081.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4081" }
We don’t call writer.close(), which causes https://github.com/huggingface/datasets/issues/4077. It can happen that we upload the file before the writer is garbage collected and writes the footer. I fixed this by explicitly closing the parquet writer. Close https://github.com/huggingface/datasets/issues/4077.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4081/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4081/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4080
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4080/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4080/comments
https://api.github.com/repos/huggingface/datasets/issues/4080/events
https://github.com/huggingface/datasets/issues/4080
1,189,667,296
I_kwDODunzps5G6OHg
4,080
NonMatchingChecksumError for downloading conll2012_ontonotesv5 dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gists_url": "https://api.github.com/users/richarddwang/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/richarddwang", "id": 17963619, "login": "richarddwang", "node_id": "MDQ6VXNlcjE3OTYzNjE5", "organizations_url": "https://api.github.com/users/richarddwang/orgs", "received_events_url": "https://api.github.com/users/richarddwang/received_events", "repos_url": "https://api.github.com/users/richarddwang/repos", "site_admin": false, "starred_url": "https://api.github.com/users/richarddwang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/richarddwang/subscriptions", "type": "User", "url": "https://api.github.com/users/richarddwang" }
[ { "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists", "id": 1935892865, "name": "duplicate", "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate" }, { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-04-01T11:34:28Z
2022-04-01T13:59:10Z
2022-04-01T13:59:10Z
CONTRIBUTOR
null
null
null
## Steps to reproduce the bug ```python datasets.load_dataset("conll2012_ontonotesv5", "english_v12") ``` ## Actual results ``` Downloading builder script: 32.2kB [00:00, 9.72MB/s] Downloading metadata: 20.0kB [00:00, 10.4MB/s] Downloading and preparing dataset conll2012_ontonotesv5/english_v12 (download: 174.83 MiB, generated: 204.29 MiB, post-processed: Unknown size , total: 379.12 MiB) to ... Traceback (most recent call last): [315/390] File "/home/yisiang/lgtn/conll2012/run.py", line 86, in <module> train() File "/home/yisiang/lgtn/conll2012/run.py", line 65, in train trainer.fit(model, datamodule=dm) File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/pytorch_lightning/trainer/trainer.py", line 740, in fit self._call_and_handle_interrupt( File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/pytorch_lightning/trainer/trainer.py", line 685, in _call_and_handle_inte rrupt return trainer_fn(*args, **kwargs) File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/pytorch_lightning/trainer/trainer.py", line 777, in _fit_impl self._run(model, ckpt_path=ckpt_path) File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/pytorch_lightning/trainer/trainer.py", line 1131, in _run self._data_connector.prepare_data() File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/pytorch_lightning/trainer/connectors/data_connector.py", line 154, in pre pare_data self.trainer.datamodule.prepare_data() File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/pytorch_lightning/core/datamodule.py", line 474, in wrapped_fn fn(*args, **kwargs) File "/home/yisiang/lgtn/_abstract_task/data.py", line 43, in prepare_data raw_dsets = datasets.load_dataset(**load_dataset_kwargs) File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/datasets/load.py", line 1687, in load_dataset builder_instance.download_and_prepare( File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/datasets/builder.py", line 605, in download_and_prepare self._download_and_prepare( File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/datasets/builder.py", line 1104, in _download_and_prepare super()._download_and_prepare(dl_manager, verify_infos, check_duplicate_keys=verify_infos) File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/datasets/builder.py", line 676, in _download_and_prepare verify_checksums( File "/home/yisiang/miniconda3/envs/ai/lib/python3.9/site-packages/datasets/utils/info_utils.py", line 40, in verify_checksums raise NonMatchingChecksumError(error_msg + str(bad_urls)) datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://md-datasets-cache-zipfiles-prod.s3.eu-west-1.amazonaws.com/zmycy7t9h9-1.zip'] ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.0.0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4080/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4080/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4079
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4079/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4079/comments
https://api.github.com/repos/huggingface/datasets/issues/4079/events
https://github.com/huggingface/datasets/pull/4079
1,189,521,576
PR_kwDODunzps41eYRC
4,079
Increase max retries for GitHub datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-01T09:34:03Z
2022-04-01T15:32:40Z
2022-04-01T15:27:11Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4079.diff", "html_url": "https://github.com/huggingface/datasets/pull/4079", "merged_at": "2022-04-01T15:27:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/4079.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4079" }
As GitHub recurrently raises connectivity issues, this PR increases the number of max retries to request GitHub datasets, as previously done for GitHub metrics: - #4063 Note that this is a temporary solution, while we decide when and how to load GitHub datasets from the Hub: - #4059 Fix #2048 Related to: - #4051 - #3210 - #2787 - #2075 - #2036 CC: @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4079/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4079/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4078
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4078/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4078/comments
https://api.github.com/repos/huggingface/datasets/issues/4078/events
https://github.com/huggingface/datasets/pull/4078
1,189,513,572
PR_kwDODunzps41eWnl
4,078
Fix GithubMetricModuleFactory instantiation with None download_config
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-04-01T09:26:58Z
2022-04-01T14:44:51Z
2022-04-01T14:39:27Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4078.diff", "html_url": "https://github.com/huggingface/datasets/pull/4078", "merged_at": "2022-04-01T14:39:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/4078.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4078" }
Recent PR: - #4063 introduced a potential bug if `GithubMetricModuleFactory` is instantiated with None `download_config`. This PR add instantiation tests and fix that potential issue. CC: @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4078/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4078/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4077
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4077/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4077/comments
https://api.github.com/repos/huggingface/datasets/issues/4077/events
https://github.com/huggingface/datasets/issues/4077
1,189,467,585
I_kwDODunzps5G5dXB
4,077
ArrowInvalid: Parquet magic bytes not found in footer. Either the file is corrupted or this is not a parquet file.
{ "avatar_url": "https://avatars.githubusercontent.com/u/48327001?v=4", "events_url": "https://api.github.com/users/NielsRogge/events{/privacy}", "followers_url": "https://api.github.com/users/NielsRogge/followers", "following_url": "https://api.github.com/users/NielsRogge/following{/other_user}", "gists_url": "https://api.github.com/users/NielsRogge/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NielsRogge", "id": 48327001, "login": "NielsRogge", "node_id": "MDQ6VXNlcjQ4MzI3MDAx", "organizations_url": "https://api.github.com/users/NielsRogge/orgs", "received_events_url": "https://api.github.com/users/NielsRogge/received_events", "repos_url": "https://api.github.com/users/NielsRogge/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NielsRogge/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NielsRogge/subscriptions", "type": "User", "url": "https://api.github.com/users/NielsRogge" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
[]
2022-04-01T08:49:13Z
2022-04-01T16:16:19Z
2022-04-01T16:16:19Z
CONTRIBUTOR
null
null
null
## Describe the bug When uploading a relatively large image dataset of > 1GB, reloading doesn't work for me, even though pushing to the hub went just fine. Basically, I do: ``` from datasets import load_dataset dataset = load_dataset("imagefolder", data_files="path_to_my_files") dataset.push_to_hub("dataset_name") # works fine, no errors reloaded_dataset = load_dataset("dataset_name") ``` and it returns: ``` /usr/local/lib/python3.7/dist-packages/pyarrow/error.pxi in pyarrow.lib.check_status() ArrowInvalid: Parquet magic bytes not found in footer. Either the file is corrupted or this is not a parquet file. ``` I created a Colab notebook to reproduce my error: https://colab.research.google.com/drive/141LJCcM2XyqprPY83nIQ-Zk3BbxWeahq?usp=sharing
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4077/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4077/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4076
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4076/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4076/comments
https://api.github.com/repos/huggingface/datasets/issues/4076/events
https://github.com/huggingface/datasets/pull/4076
1,188,478,867
PR_kwDODunzps41a1n2
4,076
Add ROUGE Metric Card
{ "avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4", "events_url": "https://api.github.com/users/emibaylor/events{/privacy}", "followers_url": "https://api.github.com/users/emibaylor/followers", "following_url": "https://api.github.com/users/emibaylor/following{/other_user}", "gists_url": "https://api.github.com/users/emibaylor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/emibaylor", "id": 27527747, "login": "emibaylor", "node_id": "MDQ6VXNlcjI3NTI3NzQ3", "organizations_url": "https://api.github.com/users/emibaylor/orgs", "received_events_url": "https://api.github.com/users/emibaylor/received_events", "repos_url": "https://api.github.com/users/emibaylor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/emibaylor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/emibaylor/subscriptions", "type": "User", "url": "https://api.github.com/users/emibaylor" }
[]
closed
false
null
[]
null
[]
2022-03-31T18:34:34Z
2022-04-12T20:43:45Z
2022-04-12T20:37:38Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4076.diff", "html_url": "https://github.com/huggingface/datasets/pull/4076", "merged_at": "2022-04-12T20:37:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/4076.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4076" }
Add ROUGE metric card. I've left the 'Values from popular papers' section empty for the time being because I don't know the summarization literature very well and am therefore not sure which paper(s) to pull from (note that the original rouge paper does not seem to present specific values, just correlations with human judgements). Any suggestions on which paper(s) to pull from would be helpful! :)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4076/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4076/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4075
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4075/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4075/comments
https://api.github.com/repos/huggingface/datasets/issues/4075/events
https://github.com/huggingface/datasets/issues/4075
1,188,462,162
I_kwDODunzps5G1n5S
4,075
Add CCAgT dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/20444345?v=4", "events_url": "https://api.github.com/users/johnnv1/events{/privacy}", "followers_url": "https://api.github.com/users/johnnv1/followers", "following_url": "https://api.github.com/users/johnnv1/following{/other_user}", "gists_url": "https://api.github.com/users/johnnv1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/johnnv1", "id": 20444345, "login": "johnnv1", "node_id": "MDQ6VXNlcjIwNDQ0MzQ1", "organizations_url": "https://api.github.com/users/johnnv1/orgs", "received_events_url": "https://api.github.com/users/johnnv1/received_events", "repos_url": "https://api.github.com/users/johnnv1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/johnnv1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/johnnv1/subscriptions", "type": "User", "url": "https://api.github.com/users/johnnv1" }
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" }, { "color": "bfdadc", "default": false, "description": "Vision datasets", "id": 3608941089, "name": "vision", "node_id": "LA_kwDODunzps7XHBIh", "url": "https://api.github.com/repos/huggingface/datasets/labels/vision" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/20444345?v=4", "events_url": "https://api.github.com/users/johnnv1/events{/privacy}", "followers_url": "https://api.github.com/users/johnnv1/followers", "following_url": "https://api.github.com/users/johnnv1/following{/other_user}", "gists_url": "https://api.github.com/users/johnnv1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/johnnv1", "id": 20444345, "login": "johnnv1", "node_id": "MDQ6VXNlcjIwNDQ0MzQ1", "organizations_url": "https://api.github.com/users/johnnv1/orgs", "received_events_url": "https://api.github.com/users/johnnv1/received_events", "repos_url": "https://api.github.com/users/johnnv1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/johnnv1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/johnnv1/subscriptions", "type": "User", "url": "https://api.github.com/users/johnnv1" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/20444345?v=4", "events_url": "https://api.github.com/users/johnnv1/events{/privacy}", "followers_url": "https://api.github.com/users/johnnv1/followers", "following_url": "https://api.github.com/users/johnnv1/following{/other_user}", "gists_url": "https://api.github.com/users/johnnv1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/johnnv1", "id": 20444345, "login": "johnnv1", "node_id": "MDQ6VXNlcjIwNDQ0MzQ1", "organizations_url": "https://api.github.com/users/johnnv1/orgs", "received_events_url": "https://api.github.com/users/johnnv1/received_events", "repos_url": "https://api.github.com/users/johnnv1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/johnnv1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/johnnv1/subscriptions", "type": "User", "url": "https://api.github.com/users/johnnv1" } ]
null
[]
2022-03-31T18:20:28Z
2022-07-06T19:03:42Z
2022-07-06T19:03:42Z
NONE
null
null
null
## Adding a Dataset - **Name:** CCAgT dataset: Images of Cervical Cells with AgNOR Stain Technique - **Description:** The dataset contains 2540 images (1600x1200 where each pixel is 0.111μm×0.111μm) from three different slides, having at least one nucleus per image. These images are from fields belonging to a sample cervical slide, colored with silver-stained, a method known as Argyrophilic Nucleolar Organizer Regions (AgNOR). - **Paper:** https://doi.org/10.1109/cbms49503.2020.00110 - **Data:** https://arquivos.ufsc.br/d/373be2177a33426a9e6c/ or https://drive.google.com/drive/u/4/folders/1TBpYCv6S1ydASLauSzcsvO7Wc5O-WUw0 - **Motivation:** This is a unique dataset (because of the stain), for a major health problem, cervical cancer, with real data. Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md). Hi, this is a public version of the dataset that I have been working on, soon we will have another version of this dataset. But until this new version goes out, I thought I would add this dataset here, if it makes sense for the repository. You can assign the task to me if possible
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4075/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4075/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4074
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4074/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4074/comments
https://api.github.com/repos/huggingface/datasets/issues/4074/events
https://github.com/huggingface/datasets/issues/4074
1,188,449,142
I_kwDODunzps5G1kt2
4,074
Error in google/xtreme_s dataset card
{ "avatar_url": "https://avatars.githubusercontent.com/u/1048544?v=4", "events_url": "https://api.github.com/users/wranai/events{/privacy}", "followers_url": "https://api.github.com/users/wranai/followers", "following_url": "https://api.github.com/users/wranai/following{/other_user}", "gists_url": "https://api.github.com/users/wranai/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wranai", "id": 1048544, "login": "wranai", "node_id": "MDQ6VXNlcjEwNDg1NDQ=", "organizations_url": "https://api.github.com/users/wranai/orgs", "received_events_url": "https://api.github.com/users/wranai/received_events", "repos_url": "https://api.github.com/users/wranai/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wranai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wranai/subscriptions", "type": "User", "url": "https://api.github.com/users/wranai" }
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" }, { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
[]
2022-03-31T18:07:45Z
2022-04-01T08:12:56Z
2022-04-01T08:12:56Z
NONE
null
null
null
**Link:** https://huggingface.co/datasets/google/xtreme_s Not a big deal but Hungarian is considered an Eastern European language, together with Serbian, Slovak, Slovenian (all correctly categorized; Slovenia is mostly to the West of Hungary, by the way).
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4074/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4074/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4073
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4073/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4073/comments
https://api.github.com/repos/huggingface/datasets/issues/4073/events
https://github.com/huggingface/datasets/pull/4073
1,188,364,711
PR_kwDODunzps41adPA
4,073
Create a metric card for Competition MATH
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
closed
false
null
[]
null
[]
2022-03-31T16:48:59Z
2022-04-01T19:02:39Z
2022-04-01T18:57:13Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4073.diff", "html_url": "https://github.com/huggingface/datasets/pull/4073", "merged_at": "2022-04-01T18:57:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/4073.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4073" }
Proposing metric card for Competition MATH
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4073/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4073/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4072
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4072/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4072/comments
https://api.github.com/repos/huggingface/datasets/issues/4072/events
https://github.com/huggingface/datasets/pull/4072
1,188,266,410
PR_kwDODunzps41aIUG
4,072
Add installation instructions to image_process doc
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[]
2022-03-31T15:29:37Z
2022-03-31T17:05:46Z
2022-03-31T17:00:19Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4072.diff", "html_url": "https://github.com/huggingface/datasets/pull/4072", "merged_at": "2022-03-31T17:00:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/4072.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4072" }
This PR adds the installation instructions for the Image feature to the image process doc.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4072/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4072/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4071
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4071/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4071/comments
https://api.github.com/repos/huggingface/datasets/issues/4071/events
https://github.com/huggingface/datasets/issues/4071
1,187,587,683
I_kwDODunzps5GySZj
4,071
Loading issue for xuyeliu/notebookCDG dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/46160972?v=4", "events_url": "https://api.github.com/users/Jun-jie-Huang/events{/privacy}", "followers_url": "https://api.github.com/users/Jun-jie-Huang/followers", "following_url": "https://api.github.com/users/Jun-jie-Huang/following{/other_user}", "gists_url": "https://api.github.com/users/Jun-jie-Huang/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Jun-jie-Huang", "id": 46160972, "login": "Jun-jie-Huang", "node_id": "MDQ6VXNlcjQ2MTYwOTcy", "organizations_url": "https://api.github.com/users/Jun-jie-Huang/orgs", "received_events_url": "https://api.github.com/users/Jun-jie-Huang/received_events", "repos_url": "https://api.github.com/users/Jun-jie-Huang/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Jun-jie-Huang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Jun-jie-Huang/subscriptions", "type": "User", "url": "https://api.github.com/users/Jun-jie-Huang" }
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
[]
2022-03-31T06:36:29Z
2022-03-31T08:17:01Z
2022-03-31T08:16:16Z
NONE
null
null
null
## Dataset viewer issue for '*xuyeliu/notebookCDG*' **Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/xuyeliu/notebookCDG)* *Couldn't load the xuyeliu/notebookCDG with provided scripts: * ``` from datasets import load_dataset dataset = load_dataset("xuyeliu/notebookCDG/dataset_notebook.pkl") ``` I get an error message as follows: FileNotFoundError: Couldn't find a dataset script at /home/code_documentation/code/xuyeliu/notebookCDG/notebookCDG.py or any data file in the same directory. Couldn't find 'xuyeliu/notebookCDG' on the Hugging Face Hub either: FileNotFoundError: Unable to resolve any data file that matches ['**train*'] in dataset repository xuyeliu/notebookCDG with any supported extension ['csv', 'tsv', 'json', 'jsonl', 'parquet', 'txt', 'blp', 'bmp', 'dib', 'bufr', 'cur', 'pcx', 'dcx', 'dds', 'ps', 'eps', 'fit', 'fits', 'fli', 'flc', 'ftc', 'ftu', 'gbr', 'gif', 'grib', 'h5', 'hdf', 'png', 'apng', 'jp2', 'j2k', 'jpc', 'jpf', 'jpx', 'j2c', 'icns', 'ico', 'im', 'iim', 'tif', 'tiff', 'jfif', 'jpe', 'jpg', 'jpeg', 'mpg', 'mpeg', 'msp', 'pcd', 'pxr', 'pbm', 'pgm', 'ppm', 'pnm', 'psd', 'bw', 'rgb', 'rgba', 'sgi', 'ras', 'tga', 'icb', 'vda', 'vst', 'webp', 'wmf', 'emf', 'xbm', 'xpm', 'zip'] Am I the one who added this dataset ? No
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4071/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4071/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4070
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4070/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4070/comments
https://api.github.com/repos/huggingface/datasets/issues/4070/events
https://github.com/huggingface/datasets/pull/4070
1,186,810,205
PR_kwDODunzps41VMYq
4,070
Create metric card for seqeval
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
closed
false
null
[]
null
[]
2022-03-30T18:08:01Z
2022-04-01T19:02:58Z
2022-04-01T18:57:25Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4070.diff", "html_url": "https://github.com/huggingface/datasets/pull/4070", "merged_at": "2022-04-01T18:57:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/4070.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4070" }
Proposing metric card for seqeval. Not sure which values to report for Popular papers though.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4070/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4070/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4069
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4069/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4069/comments
https://api.github.com/repos/huggingface/datasets/issues/4069/events
https://github.com/huggingface/datasets/pull/4069
1,186,790,578
PR_kwDODunzps41VIMJ
4,069
Add support for metadata files to `imagefolder`
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[]
2022-03-30T17:47:51Z
2022-05-03T12:49:00Z
2022-05-03T12:42:16Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4069.diff", "html_url": "https://github.com/huggingface/datasets/pull/4069", "merged_at": "2022-05-03T12:42:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/4069.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4069" }
This PR adds support for metadata files to `imagefolder` to add an ability to specify image fields other than `image` and `label`, which are inferred from the directory structure in the loaded dataset. To be parsed as an image metadata file, a file should be named `"info.csv"` and should have the following structure: ``` image_id,some_col1_name,some_col2_name rel/path/to/image1.jpg,image1_col1_value,image1_col2_value rel/path/to/image2.jpg,image2_col1_value,image2_col2_value ... ``` This is how the resolution works: ``` - path/to/imagefolder/directory - info.csv - 10.jpg # referenced as 10.jpg in "info.csv" - Cat - 0.jpg # referenced as Cat/0.jpg in "info.csv" - 1.jpg # referenced as Cat/1.jpg in "info.csv" - Dog - 0.jpg # referenced as Dog/0.jpg in "info.csv" - 1.jpg # referenced as Dog/1.jpg in "info.csv" ``` Open questions: 1. IMO it makes more sense to store image metadata as JSON Lines than CSV. CSV is sufficient for textual metadata but not the best for representing bounding boxes, for instance. Also, JSON Lines is more strict, which is good in this case (CSV supports various delimiters, the header line is optional, etc., so it's easier to enforce rules on JSON Lines that it's on CSV) 2. A better name for the `image_id` column, which contains image identifiers? Maybe `image_file` or `image_filename`? 3. WDYT about making `with_metadata=True` the default behavior if the loaded repo/directory contains an `info.csv` file? An example repository: https://huggingface.co/datasets/mariosasko/PetImages. Can be loaded by installing `datasets` from the PR branch and running `load_dataset("mariosasko/PetImages", with_metadata=True)`. cc: @abhishekkrthakur (this PR should address https://huggingface.slack.com/archives/C02JB9L6JKF/p1645450017434029?thread_ts=1645157416.389499&cid=C02JB9L6JKF) TODOs: - [x] Test - [x] Metadata file nesting ``` - path/to/imagefolder/directory - info.csv - 10.jpg - Cat - info.csv # should have higher precedence in this directory than the top-level info.csv, but we choose the first "eligible" metadata file currently - 0.jpg - 1.jpg ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4069/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4069/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4068
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4068/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4068/comments
https://api.github.com/repos/huggingface/datasets/issues/4068/events
https://github.com/huggingface/datasets/pull/4068
1,186,765,422
PR_kwDODunzps41VC0I
4,068
Improve out of bounds error message
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2022-03-30T17:22:10Z
2022-03-31T08:39:08Z
2022-03-31T08:33:57Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4068.diff", "html_url": "https://github.com/huggingface/datasets/pull/4068", "merged_at": "2022-03-31T08:33:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/4068.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4068" }
In 1.18.4 with https://github.com/huggingface/datasets/pull/3719 we introduced an error message for users using `select` with out of bounds indices. The message ended up being confusing for some users because it mentioned negative indices, which is not the main use case. I replaced it with a message that is very similar to the one you get with you try to access a list with an out-of-range index.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4068/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4068/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4067
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4067/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4067/comments
https://api.github.com/repos/huggingface/datasets/issues/4067/events
https://github.com/huggingface/datasets/pull/4067
1,186,731,905
PR_kwDODunzps41U7qc
4,067
Update datasets task tags to align tags with models
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2022-03-30T16:49:32Z
2022-04-13T17:37:27Z
2022-04-13T17:31:11Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4067.diff", "html_url": "https://github.com/huggingface/datasets/pull/4067", "merged_at": "2022-04-13T17:31:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/4067.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4067" }
**Requires https://github.com/huggingface/datasets/pull/4066 to be merged first** Following https://github.com/huggingface/datasets/pull/4066 we need to update many dataset tags to use the new ones. This PR takes case of this and is quite big - feel free to review only certain tags if you don't want to spend too much time on it. Note that the CI will never be green for this PR, because many dataset cards have missing tags or sections, and fixing them is out of scope of this PR (the CI on master will be green anyway)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4067/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4067/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4066
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4066/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4066/comments
https://api.github.com/repos/huggingface/datasets/issues/4066/events
https://github.com/huggingface/datasets/pull/4066
1,186,728,104
PR_kwDODunzps41U63x
4,066
Tasks alignment with models
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2022-03-30T16:45:56Z
2022-04-13T13:12:52Z
2022-04-08T12:20:00Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4066.diff", "html_url": "https://github.com/huggingface/datasets/pull/4066", "merged_at": "2022-04-08T12:20:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/4066.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4066" }
I updated our `tasks.json` file with the new task taxonomy that is aligned with models. The rule that defines a task is the following: **Two tasks are different if and only if the steps of their pipelines** are different, i.e. if they can’t reasonably be implemented using the same coherent code (level of granularity/complexity of the code to be defined - ideally I’d like to say “HF user’s level”) - this is the same definition in `transformers` I will update the tags of all the datasets in this repository [in another PR](https://github.com/huggingface/datasets/pull/4067) for readability. Main changes: - conditional-text-generation is split between summarization, translation, text-generation and text2text-generation - speech-processing is split into automatic-speech-recognition, audio-classification, etc. - structure-prediction is renamed token-classification - abstractive-qa now belongs to text2text-generation Here is just a simplified YAML dump of `tasks.json`: ```yaml audio-classification: - keyword-spotting - speaker-identification - speaker-intent-classification - emotion-recognition - speaker-language-identification audio-to-audio: [] automatic-speech-recognition: [] conversational: - dialogue-generation feature-extraction: [] fill-mask: - slot-filling - masked-language-modeling image-classification: - multi-label-image-classification - multi-class-image-classification image-segmentation: - instance-segmentation - semantic-segmentation - panoptic-segmentation image-to-text: - image-captioning multiple-choice: - multiple-choice-qa - multiple-choice-coreference-resolution object-detection: - face-detection - vehicle-detection question-answering: - extractive-qa - open-domain-qa - closed-domain-qa sentence-similarity: [] tabular-classification: [] tabular-to-text: - rdf-to-text summarization: - news-articles-summarization - news-articles-headline-generation table-to-text: [] table-question-answering: [] text-classification: - acceptability-classification - entity-linking-classification - fact-checking - intent-classification - multi-class-classification - multi-label-classification - natural-language-inference - semantic-similarity-classification - sentiment-classification - topic-classification - semantic-similarity-scoring - sentiment-scoring - sentiment-analysis - hate-speech-detection - text-scoring text-generation: - dialogue-modeling - language-modeling text-retrieval: - document-retrieval - utterance-retrieval - entity-linking-retrieval - fact-checking-retrieval text-to-image: [] text-to-tabular: - relation-extraction - semantic-role-labeling text-to-speech: [] text2text-generation: - text-simplification - explanation-generation - abstractive-qa - open-domain-abstractive-qa - closed-domain-qa - open-book-qa - closed-book-qa time-series-forecasting: - univariate-time-series-forecasting - multivariate-time-series-forecasting token-classification: - named-entity-recognition - part-of-speech-tagging - parsing - lemmatization - word-sense-disambiguation - coreference-resolution translation: [] visual-question-answering: [] voice-activity-detection: [] zero-shot-classification: [] zero-shot-image-classification: [] reinforcement-learning: [] other: [] ``` Feel free to comment and give suggestions, especially if you think we can also align this list with other projects cc @julien-c @osanseviero @severo @lewtun @yjernite @albertvillanova @mariosasko @polinaeterna
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 5, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 7, "url": "https://api.github.com/repos/huggingface/datasets/issues/4066/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4066/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4065
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4065/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4065/comments
https://api.github.com/repos/huggingface/datasets/issues/4065/events
https://github.com/huggingface/datasets/pull/4065
1,186,722,478
PR_kwDODunzps41U5rq
4,065
Create metric card for METEOR
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
closed
false
null
[]
null
[]
2022-03-30T16:40:30Z
2022-03-31T17:12:10Z
2022-03-31T17:07:50Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4065.diff", "html_url": "https://github.com/huggingface/datasets/pull/4065", "merged_at": "2022-03-31T17:07:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/4065.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4065" }
Proposing a metric card for METEOR
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4065/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4065/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4064
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4064/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4064/comments
https://api.github.com/repos/huggingface/datasets/issues/4064/events
https://github.com/huggingface/datasets/pull/4064
1,186,650,321
PR_kwDODunzps41UqXS
4,064
Contributing MedMCQA dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/17107749?v=4", "events_url": "https://api.github.com/users/monk1337/events{/privacy}", "followers_url": "https://api.github.com/users/monk1337/followers", "following_url": "https://api.github.com/users/monk1337/following{/other_user}", "gists_url": "https://api.github.com/users/monk1337/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/monk1337", "id": 17107749, "login": "monk1337", "node_id": "MDQ6VXNlcjE3MTA3NzQ5", "organizations_url": "https://api.github.com/users/monk1337/orgs", "received_events_url": "https://api.github.com/users/monk1337/received_events", "repos_url": "https://api.github.com/users/monk1337/repos", "site_admin": false, "starred_url": "https://api.github.com/users/monk1337/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/monk1337/subscriptions", "type": "User", "url": "https://api.github.com/users/monk1337" }
[]
closed
false
null
[]
null
[]
2022-03-30T15:42:47Z
2022-05-06T09:40:40Z
2022-05-06T08:42:56Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4064.diff", "html_url": "https://github.com/huggingface/datasets/pull/4064", "merged_at": "2022-05-06T08:42:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/4064.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4064" }
Adding MedMCQA dataset ( https://paperswithcode.com/dataset/medmcqa ) **Name**: MedMCQA **Description**: MedMCQA is a large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions. MedMCQA has more than 194k high-quality AIIMS & NEET PG entrance exam MCQs covering 2.4k healthcare topics and 21 medical subjects are collected with an average token length of 12.77 and high topical diversity. The dataset contains questions about the following topics: Anesthesia, Anatomy, Biochemistry, Dental, ENT, Forensic Medicine (FM), Obstetrics and Gynecology (O&G), Medicine, Microbiology, Ophthalmology, Orthopedics Pathology, Pediatrics, Pharmacology, Physiology, Psychiatry, Radiology Skin, Preventive & Social Medicine (PSM), and Surgery **Code**: https://github.com/medmcqa/medmcqa All files are at place : **a dataset script** : medmcqa.py **a dataset card with tags and information** : README.md. **a metadata file** : dataset_infos.json **a dummy-data file** : Please help to generate this file, I was facing ` raise JSONDecodeError("Extra data", s, end)` error
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4064/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4064/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4063
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4063/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4063/comments
https://api.github.com/repos/huggingface/datasets/issues/4063/events
https://github.com/huggingface/datasets/pull/4063
1,186,611,368
PR_kwDODunzps41UiDm
4,063
Increase max retries for GitHub metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-03-30T15:12:48Z
2022-03-31T14:42:52Z
2022-03-31T14:37:47Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4063.diff", "html_url": "https://github.com/huggingface/datasets/pull/4063", "merged_at": "2022-03-31T14:37:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/4063.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4063" }
As GitHub recurrently raises connectivity issues, this PR increases the number of max retries to request GitHub metrics. Related to: - #3134 Also related to: - #4059
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4063/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4063/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4062
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4062/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4062/comments
https://api.github.com/repos/huggingface/datasets/issues/4062/events
https://github.com/huggingface/datasets/issues/4062
1,186,330,732
I_kwDODunzps5Gtfhs
4,062
Loading mozilla-foundation/common_voice_7_0 dataset failed
{ "avatar_url": "https://avatars.githubusercontent.com/u/19529125?v=4", "events_url": "https://api.github.com/users/aapot/events{/privacy}", "followers_url": "https://api.github.com/users/aapot/followers", "following_url": "https://api.github.com/users/aapot/following{/other_user}", "gists_url": "https://api.github.com/users/aapot/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/aapot", "id": 19529125, "login": "aapot", "node_id": "MDQ6VXNlcjE5NTI5MTI1", "organizations_url": "https://api.github.com/users/aapot/orgs", "received_events_url": "https://api.github.com/users/aapot/received_events", "repos_url": "https://api.github.com/users/aapot/repos", "site_admin": false, "starred_url": "https://api.github.com/users/aapot/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aapot/subscriptions", "type": "User", "url": "https://api.github.com/users/aapot" }
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-03-30T11:39:41Z
2022-06-21T07:36:23Z
2022-03-31T08:18:04Z
NONE
null
null
null
## Describe the bug I wanted to load `mozilla-foundation/common_voice_7_0` dataset with `fi` language and `test` split from datasets on Colab/Kaggle notebook, but I am getting an error `JSONDecodeError: [Errno Expecting value] Not Found: 0` while loading it. The bug seems to affect other languages and splits too than just the `fi` and `test` split. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("mozilla-foundation/common_voice_7_0", "fi", split="test", use_auth_token="YOUR TOKEN") ``` ## Expected results load `mozilla-foundation/common_voice_7_0` dataset succesfully ## Actual results ``` JSONDecodeError Traceback (most recent call last) /opt/conda/lib/python3.7/site-packages/requests/models.py in json(self, **kwargs) 909 try: --> 910 return complexjson.loads(self.text, **kwargs) 911 except JSONDecodeError as e: /opt/conda/lib/python3.7/site-packages/simplejson/__init__.py in loads(s, encoding, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, use_decimal, **kw) 524 and not use_decimal and not kw): --> 525 return _default_decoder.decode(s) 526 if cls is None: /opt/conda/lib/python3.7/site-packages/simplejson/decoder.py in decode(self, s, _w, _PY3) 369 s = str(s, self.encoding) --> 370 obj, end = self.raw_decode(s) 371 end = _w(s, end).end() /opt/conda/lib/python3.7/site-packages/simplejson/decoder.py in raw_decode(self, s, idx, _w, _PY3) 399 idx += 3 --> 400 return self.scan_once(s, idx=_w(s, idx).end()) JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: JSONDecodeError Traceback (most recent call last) /tmp/ipykernel_358/370980805.py in <module> 1 # load Common Voice 7.0 dataset from Huggingface with Finnish "test" split ----> 2 test_dataset = load_dataset("mozilla-foundation/common_voice_7_0", "fi", split="test", use_auth_token=True) /opt/conda/lib/python3.7/site-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) 1690 ignore_verifications=ignore_verifications, 1691 try_from_hf_gcs=try_from_hf_gcs, -> 1692 use_auth_token=use_auth_token, 1693 ) 1694 /opt/conda/lib/python3.7/site-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs) 604 if not downloaded_from_gcs: 605 self._download_and_prepare( --> 606 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 607 ) 608 # Sync info /opt/conda/lib/python3.7/site-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos) 1102 1103 def _download_and_prepare(self, dl_manager, verify_infos): -> 1104 super()._download_and_prepare(dl_manager, verify_infos, check_duplicate_keys=verify_infos) 1105 1106 def _get_examples_iterable_for_split(self, split_generator: SplitGenerator) -> ExamplesIterable: /opt/conda/lib/python3.7/site-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 670 split_dict = SplitDict(dataset_name=self.name) 671 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) --> 672 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) 673 674 # Checksums verification ~/.cache/huggingface/modules/datasets_modules/datasets/mozilla-foundation--common_voice_7_0/fe20cac47c166e25b1f096ab661832e3da7cf298ed4a91dcaa1343ad972d175b/common_voice_7_0.py in _split_generators(self, dl_manager) 151 152 self._log_download(self.config.name, bundle_version, hf_auth_token) --> 153 archive = dl_manager.download(self._get_bundle_url(self.config.name, bundle_url_template)) 154 155 if self.config.version < datasets.Version("5.0.0"): ~/.cache/huggingface/modules/datasets_modules/datasets/mozilla-foundation--common_voice_7_0/fe20cac47c166e25b1f096ab661832e3da7cf298ed4a91dcaa1343ad972d175b/common_voice_7_0.py in _get_bundle_url(self, locale, url_template) 130 path = urllib.parse.quote(path.encode("utf-8"), safe="~()*!.'") 131 use_cdn = self.config.size_bytes < 20 * 1024 * 1024 * 1024 --> 132 response = requests.get(f"{_API_URL}/bucket/dataset/{path}/{use_cdn}", timeout=10.0).json() 133 return response["url"] 134 /opt/conda/lib/python3.7/site-packages/requests/models.py in json(self, **kwargs) 915 raise RequestsJSONDecodeError(e.message) 916 else: --> 917 raise RequestsJSONDecodeError(e.msg, e.doc, e.pos) 918 919 @property JSONDecodeError: [Errno Expecting value] Not Found: 0 ``` ## Environment info - `datasets` version: 2.0.0 - Platform: Linux-5.10.90+-x86_64-with-debian-bullseye-sid - Python version: 3.7.12 - PyArrow version: 5.0.0 - Pandas version: 1.3.5
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4062/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4062/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4061
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4061/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4061/comments
https://api.github.com/repos/huggingface/datasets/issues/4061/events
https://github.com/huggingface/datasets/issues/4061
1,186,317,071
I_kwDODunzps5GtcMP
4,061
Loading cnn_dailymail dataset failed
{ "avatar_url": "https://avatars.githubusercontent.com/u/68355048?v=4", "events_url": "https://api.github.com/users/Arij-Aladel/events{/privacy}", "followers_url": "https://api.github.com/users/Arij-Aladel/followers", "following_url": "https://api.github.com/users/Arij-Aladel/following{/other_user}", "gists_url": "https://api.github.com/users/Arij-Aladel/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Arij-Aladel", "id": 68355048, "login": "Arij-Aladel", "node_id": "MDQ6VXNlcjY4MzU1MDQ4", "organizations_url": "https://api.github.com/users/Arij-Aladel/orgs", "received_events_url": "https://api.github.com/users/Arij-Aladel/received_events", "repos_url": "https://api.github.com/users/Arij-Aladel/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Arij-Aladel/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Arij-Aladel/subscriptions", "type": "User", "url": "https://api.github.com/users/Arij-Aladel" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists", "id": 1935892865, "name": "duplicate", "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-03-30T11:29:02Z
2022-03-30T13:36:14Z
2022-03-30T13:36:14Z
NONE
null
null
null
## Describe the bug I wanted to load cnn_dailymail dataset from huggingface datasets on jupyter lab, but I am getting an error ` NotADirectoryError:[Errno20] Not a directory ` while loading it. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset('cnn_dailymail', '3.0.0') ``` ## Expected results load `cnn_dailymail` dataset succesfully ## Actual results failed to load and get error > NotADirectoryError: [Errno 20] Not a directory ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` 1.8.0: - Platform: Ubuntu-20.04 - Python version: 3.9.10 - PyArrow version: 3.0.0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4061/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4061/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4060
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4060/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4060/comments
https://api.github.com/repos/huggingface/datasets/issues/4060/events
https://github.com/huggingface/datasets/pull/4060
1,186,281,033
PR_kwDODunzps41Tbmg
4,060
Deprecate canonical Multilingual Librispeech
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
closed
false
null
[]
null
[]
2022-03-30T10:56:56Z
2022-04-01T12:54:05Z
2022-04-01T12:48:51Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4060.diff", "html_url": "https://github.com/huggingface/datasets/pull/4060", "merged_at": "2022-04-01T12:48:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/4060.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4060" }
Deprecate canonical Multilingual Librispeech in favor of [the community one](https://huggingface.co/datasets/facebook/multilingual_librispeech) which supports streaming. However, there is a problem regarding new ASR template schema: since it's changed, I guess all community datasets that use this template do not work with new version of the library, including MLS. Should we somehow notify users about that or is it possible to change this line ourselves? For MLS specifically, I cannot change the code directly as I'm not the member of the Facebook org. Hm, and the code should be change after the release, no?
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4060/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4060/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4059
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4059/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4059/comments
https://api.github.com/repos/huggingface/datasets/issues/4059/events
https://github.com/huggingface/datasets/pull/4059
1,186,149,949
PR_kwDODunzps41TC-o
4,059
Load GitHub datasets from Hub
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-03-30T09:21:56Z
2022-09-16T12:43:26Z
2022-09-16T12:40:43Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4059.diff", "html_url": "https://github.com/huggingface/datasets/pull/4059", "merged_at": "2022-09-16T12:40:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/4059.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4059" }
We have recurrently had connection errors when requesting GitHub because sometimes the site is not available. This PR requests the Hub instead, once all GitHub datasets are mirrored on the Hub. Fix #2048 Related to: - #4051 - #3210 - #2787 - #2075 - #2036
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4059/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4059/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4058
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4058/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4058/comments
https://api.github.com/repos/huggingface/datasets/issues/4058/events
https://github.com/huggingface/datasets/pull/4058
1,185,611,600
PR_kwDODunzps41RPhl
4,058
Updated annotations for nli_tr dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/2246791?v=4", "events_url": "https://api.github.com/users/e-budur/events{/privacy}", "followers_url": "https://api.github.com/users/e-budur/followers", "following_url": "https://api.github.com/users/e-budur/following{/other_user}", "gists_url": "https://api.github.com/users/e-budur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/e-budur", "id": 2246791, "login": "e-budur", "node_id": "MDQ6VXNlcjIyNDY3OTE=", "organizations_url": "https://api.github.com/users/e-budur/orgs", "received_events_url": "https://api.github.com/users/e-budur/received_events", "repos_url": "https://api.github.com/users/e-budur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/e-budur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/e-budur/subscriptions", "type": "User", "url": "https://api.github.com/users/e-budur" }
[]
closed
false
null
[]
null
[]
2022-03-29T23:46:59Z
2022-04-12T20:55:12Z
2022-04-12T10:37:22Z
CONTRIBUTOR
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4058.diff", "html_url": "https://github.com/huggingface/datasets/pull/4058", "merged_at": "2022-04-12T10:37:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/4058.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4058" }
This PR adds annotation tags for `nli_tr` dataset so that the dataset can be searchable wrt. relevant query parameters. The annotations in this PR are based on the existing annotations of `snli` and `multi_nli` datasets as `nli_tr` is a machine-generated extension of those datasets. This PR is intended only for updating the annotation labels but a followup PR will focus on updating the missing sections in the `README.md` as well. Thanks for all your time to review it.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4058/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4058/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4057
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4057/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4057/comments
https://api.github.com/repos/huggingface/datasets/issues/4057/events
https://github.com/huggingface/datasets/issues/4057
1,185,442,001
I_kwDODunzps5GqGjR
4,057
`load_dataset` consumes too much memory for audio + tar archives
{ "avatar_url": "https://avatars.githubusercontent.com/u/50839826?v=4", "events_url": "https://api.github.com/users/JFCeron/events{/privacy}", "followers_url": "https://api.github.com/users/JFCeron/followers", "following_url": "https://api.github.com/users/JFCeron/following{/other_user}", "gists_url": "https://api.github.com/users/JFCeron/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JFCeron", "id": 50839826, "login": "JFCeron", "node_id": "MDQ6VXNlcjUwODM5ODI2", "organizations_url": "https://api.github.com/users/JFCeron/orgs", "received_events_url": "https://api.github.com/users/JFCeron/received_events", "repos_url": "https://api.github.com/users/JFCeron/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JFCeron/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JFCeron/subscriptions", "type": "User", "url": "https://api.github.com/users/JFCeron" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
2022-03-29T21:38:55Z
2022-08-16T10:22:55Z
2022-08-16T10:22:55Z
NONE
null
null
null
## Description `load_dataset` consumes more and more memory until it's killed, even though it's made with a generator. I'm adding a loading script for a new dataset, made up of ~15s audio coming from a tar file. Tried setting `DEFAULT_WRITER_BATCH_SIZE = 1` as per the discussion in #741 but the problem persists. ## Steps to reproduce the bug Here's my implementation of `_generate_examples`: ```python class MyDatasetBuilder(datasets.GeneratorBasedBuilder): DEFAULT_WRITER_BATCH_SIZE = 1 ... def _split_generators(self, dl_manager): archive_path = dl_manager.download(_DL_URLS[self.config.name]) return [ datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={ "audio_tarfile_path": archive_path["audio_tarfile"] }, ), ] def _generate_examples(self, audio_tarfile_path): key = 0 with tarfile.open(audio_tarfile_path, mode="r|") as audio_tarfile: for audio_tarinfo in audio_tarfile: audio_name = audio_tarinfo.name audio_file_obj = audio_tarfile.extractfile(audio_tarinfo) yield key, {"audio": {"path": audio_name, "bytes": audio_file_obj.read()}} key += 1 ``` I then try to load via `ds = load_dataset('./datasets/my_new_dataset', writer_batch_size=1)`, and memory usage grows until all 8GB of my machine are taken and process is killed (`Killed`). Also tried an untarred version of this using `os.walk` but the same happened. I created a script to confirm that one can safely go through such a generator, which runs just fine with memory <500MB at all times. ```python import tarfile def generate_examples(): audio_tarfile = tarfile.open("audios.tar", mode="r|") key = 0 for audio_tarinfo in audio_tarfile: audio_name = audio_tarinfo.name audio_file_obj = audio_tarfile.extractfile(audio_tarinfo) yield key, {"audio": {"path": audio_name, "bytes": audio_file_obj.read()}} key += 1 if __name__ == "__main__": examples = generate_examples() for example in examples: pass ``` ## Expected results Memory consumption should be similar to the non-huggingface script. ## Actual results Process is killed after consuming too much memory. ## Environment info - `datasets` version: 2.0.1.dev0 - Platform: Linux-4.19.0-20-cloud-amd64-x86_64-with-debian-10.12 - Python version: 3.7.12 - PyArrow version: 6.0.1 - Pandas version: 1.3.5
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4057/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4057/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4056
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4056/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4056/comments
https://api.github.com/repos/huggingface/datasets/issues/4056/events
https://github.com/huggingface/datasets/issues/4056
1,185,155,775
I_kwDODunzps5GpAq_
4,056
Unexpected behavior of _TempDirWithCustomCleanup
{ "avatar_url": "https://avatars.githubusercontent.com/u/22680696?v=4", "events_url": "https://api.github.com/users/JonasGeiping/events{/privacy}", "followers_url": "https://api.github.com/users/JonasGeiping/followers", "following_url": "https://api.github.com/users/JonasGeiping/following{/other_user}", "gists_url": "https://api.github.com/users/JonasGeiping/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JonasGeiping", "id": 22680696, "login": "JonasGeiping", "node_id": "MDQ6VXNlcjIyNjgwNjk2", "organizations_url": "https://api.github.com/users/JonasGeiping/orgs", "received_events_url": "https://api.github.com/users/JonasGeiping/received_events", "repos_url": "https://api.github.com/users/JonasGeiping/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JonasGeiping/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JonasGeiping/subscriptions", "type": "User", "url": "https://api.github.com/users/JonasGeiping" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
open
false
null
[]
null
[]
2022-03-29T16:58:22Z
2022-03-30T15:08:04Z
null
NONE
null
null
null
## Describe the bug This is not 100% a bug in `datasets`, but behavior that surprised me and I think this could be made more robust on the `datasets`side. When using `datasets.disable_caching()`, cache files are written to a temporary directory. This directory should be based on the environment variable TMPDIR. I want to set TMPDIR at runtime using os.ENVIRON["TMPDIR"] = something, but depending on other imported modules this can fail to take effect. ## Steps to reproduce the bug `_TempDirWithCustomCleanup` relies on `tempfile` to generate a path to a temporary directory. However, `tempfile` generates the path only once. This can be a problem when trying to set TMPDIR at runtime whenever other code imports `tempfile` first and does something unexpected. For example (after too much trial and error) I found out that a different part of the code base I work with defines a class `PatchedDataCollatorForLanguageModeling(transformers.DataCollatorForLanguageModeling)` based on a `transformers` class. This import is enough to trigger `tempfile` to generate `tempfile` to generate a temporary path and leading to the wrong path being cached in `tempfile.tempdir`. ## Suggestion: I could file this also as bug with `transformers`, but I think fixing this on the datasets would be much more robust: Datasets could recompute the temporary path once (technically possible via `tempfile._get_default_tempdir` or resetting the global variable `tempfile.tmpdir` to None) before setting its own global `_TEMP_DIR_FOR_TEMP_CACHE_FILES`.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4056/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4056/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4055
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4055/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4055/comments
https://api.github.com/repos/huggingface/datasets/issues/4055/events
https://github.com/huggingface/datasets/pull/4055
1,184,976,292
PR_kwDODunzps41PGF1
4,055
[DO NOT MERGE] Test doc-builder
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
[]
closed
false
null
[]
null
[]
2022-03-29T14:39:02Z
2022-03-30T12:31:14Z
2022-03-30T12:25:52Z
MEMBER
null
true
{ "diff_url": "https://github.com/huggingface/datasets/pull/4055.diff", "html_url": "https://github.com/huggingface/datasets/pull/4055", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4055.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4055" }
This is a test PR to ensure the changes in https://github.com/huggingface/doc-builder/pull/164 don't break anything in `datasets`
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4055/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4055/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4054
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4054/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4054/comments
https://api.github.com/repos/huggingface/datasets/issues/4054/events
https://github.com/huggingface/datasets/pull/4054
1,184,575,368
PR_kwDODunzps41Nwjz
4,054
Support float data types in pearsonr/spearmanr metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2022-03-29T09:29:10Z
2022-03-29T14:07:59Z
2022-03-29T14:02:20Z
MEMBER
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4054.diff", "html_url": "https://github.com/huggingface/datasets/pull/4054", "merged_at": "2022-03-29T14:02:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/4054.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4054" }
Fix #4053.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4054/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4054/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4053
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4053/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4053/comments
https://api.github.com/repos/huggingface/datasets/issues/4053/events
https://github.com/huggingface/datasets/issues/4053
1,184,500,378
I_kwDODunzps5Gmgqa
4,053
Modify datatype from `int32` to `float` for pearsonr, spearmanr.
{ "avatar_url": "https://avatars.githubusercontent.com/u/86637320?v=4", "events_url": "https://api.github.com/users/Woodywarhol9/events{/privacy}", "followers_url": "https://api.github.com/users/Woodywarhol9/followers", "following_url": "https://api.github.com/users/Woodywarhol9/following{/other_user}", "gists_url": "https://api.github.com/users/Woodywarhol9/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Woodywarhol9", "id": 86637320, "login": "Woodywarhol9", "node_id": "MDQ6VXNlcjg2NjM3MzIw", "organizations_url": "https://api.github.com/users/Woodywarhol9/orgs", "received_events_url": "https://api.github.com/users/Woodywarhol9/received_events", "repos_url": "https://api.github.com/users/Woodywarhol9/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Woodywarhol9/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Woodywarhol9/subscriptions", "type": "User", "url": "https://api.github.com/users/Woodywarhol9" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-03-29T08:27:41Z
2022-03-29T14:02:20Z
2022-03-29T14:02:20Z
NONE
null
null
null
**Is your feature request related to a problem? Please describe.** - Now [Pearsonr](https://github.com/huggingface/datasets/blob/master/metrics/pearsonr/pearsonr.py) and [Spearmanr](https://github.com/huggingface/datasets/blob/master/metrics/spearmanr/spearmanr.py) both get input data as 'int32'. **Describe the solution you'd like** - Considering that those metrics are widely used for the STS task(labels are in 'float' data type), it would be better to modify datatype from 'int32' to 'float' for getting exact values of similarity.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4053/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4053/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4052
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4052/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4052/comments
https://api.github.com/repos/huggingface/datasets/issues/4052/events
https://github.com/huggingface/datasets/issues/4052
1,184,447,977
I_kwDODunzps5GmT3p
4,052
metric = metric_cls( TypeError: 'NoneType' object is not callable
{ "avatar_url": "https://avatars.githubusercontent.com/u/39409233?v=4", "events_url": "https://api.github.com/users/klyuhang9/events{/privacy}", "followers_url": "https://api.github.com/users/klyuhang9/followers", "following_url": "https://api.github.com/users/klyuhang9/following{/other_user}", "gists_url": "https://api.github.com/users/klyuhang9/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/klyuhang9", "id": 39409233, "login": "klyuhang9", "node_id": "MDQ6VXNlcjM5NDA5MjMz", "organizations_url": "https://api.github.com/users/klyuhang9/orgs", "received_events_url": "https://api.github.com/users/klyuhang9/received_events", "repos_url": "https://api.github.com/users/klyuhang9/repos", "site_admin": false, "starred_url": "https://api.github.com/users/klyuhang9/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/klyuhang9/subscriptions", "type": "User", "url": "https://api.github.com/users/klyuhang9" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
[]
2022-03-29T07:43:08Z
2022-03-29T14:06:01Z
2022-03-29T14:06:01Z
NONE
null
null
null
Hi, friend. I meet a problem. When I run the code: `metric = load_metric('glue', 'rte')` There is a problem raising: `metric = metric_cls( TypeError: 'NoneType' object is not callable ` I don't know why. Thanks for your help!
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4052/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4052/timeline
null
completed
true
https://api.github.com/repos/huggingface/datasets/issues/4051
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4051/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4051/comments
https://api.github.com/repos/huggingface/datasets/issues/4051/events
https://github.com/huggingface/datasets/issues/4051
1,184,400,179
I_kwDODunzps5GmIMz
4,051
ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/2.0.0/datasets/glue/glue.py
{ "avatar_url": "https://avatars.githubusercontent.com/u/39409233?v=4", "events_url": "https://api.github.com/users/klyuhang9/events{/privacy}", "followers_url": "https://api.github.com/users/klyuhang9/followers", "following_url": "https://api.github.com/users/klyuhang9/following{/other_user}", "gists_url": "https://api.github.com/users/klyuhang9/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/klyuhang9", "id": 39409233, "login": "klyuhang9", "node_id": "MDQ6VXNlcjM5NDA5MjMz", "organizations_url": "https://api.github.com/users/klyuhang9/orgs", "received_events_url": "https://api.github.com/users/klyuhang9/received_events", "repos_url": "https://api.github.com/users/klyuhang9/repos", "site_admin": false, "starred_url": "https://api.github.com/users/klyuhang9/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/klyuhang9/subscriptions", "type": "User", "url": "https://api.github.com/users/klyuhang9" }
[]
closed
false
null
[]
null
[]
2022-03-29T07:00:31Z
2022-05-08T07:27:32Z
2022-03-29T08:29:25Z
NONE
null
null
null
Hi, I meet a problem. When I run the code: `dataset = load_dataset('glue','sst2')` There is a issue raising: ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/2.0.0/datasets/glue/glue.py I don't know why; it is ok when I use Google Chrome to view this url. Thanks for your help!
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4051/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4051/timeline
null
completed
true