url
stringlengths 58
61
| repository_url
stringclasses 1
value | labels_url
stringlengths 72
75
| comments_url
stringlengths 67
70
| events_url
stringlengths 65
68
| html_url
stringlengths 46
51
| id
int64 599M
1.5B
| node_id
stringlengths 18
32
| number
int64 1
5.38k
| title
stringlengths 1
276
| user
dict | labels
list | state
stringclasses 2
values | locked
bool 1
class | assignee
dict | assignees
list | milestone
dict | comments
list | created_at
stringlengths 20
20
| updated_at
stringlengths 20
20
| closed_at
stringlengths 20
20
⌀ | author_association
stringclasses 3
values | active_lock_reason
null | draft
bool 2
classes | pull_request
dict | body
stringlengths 0
228k
⌀ | reactions
dict | timeline_url
stringlengths 67
70
| performed_via_github_app
null | state_reason
stringclasses 3
values | is_pull_request
bool 1
class |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/2114
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2114/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2114/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2114/events
|
https://github.com/huggingface/datasets/pull/2114
| 841,207,878 |
MDExOlB1bGxSZXF1ZXN0NjAwOTc1MTA3
| 2,114 |
Support for legal NLP datasets (EURLEX, ECtHR cases and EU-REG-IR)
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/1626984?v=4",
"events_url": "https://api.github.com/users/iliaschalkidis/events{/privacy}",
"followers_url": "https://api.github.com/users/iliaschalkidis/followers",
"following_url": "https://api.github.com/users/iliaschalkidis/following{/other_user}",
"gists_url": "https://api.github.com/users/iliaschalkidis/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/iliaschalkidis",
"id": 1626984,
"login": "iliaschalkidis",
"node_id": "MDQ6VXNlcjE2MjY5ODQ=",
"organizations_url": "https://api.github.com/users/iliaschalkidis/orgs",
"received_events_url": "https://api.github.com/users/iliaschalkidis/received_events",
"repos_url": "https://api.github.com/users/iliaschalkidis/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/iliaschalkidis/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/iliaschalkidis/subscriptions",
"type": "User",
"url": "https://api.github.com/users/iliaschalkidis"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-25T18:40:17Z
|
2021-03-31T10:38:50Z
|
2021-03-31T10:38:50Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2114.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2114",
"merged_at": "2021-03-31T10:38:50Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2114.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2114"
}
|
Add support for two legal NLP datasets:
- EURLEX (https://www.aclweb.org/anthology/P19-1636/)
- ECtHR cases (https://arxiv.org/abs/2103.13084)
- EU-REG-IR (https://arxiv.org/abs/2101.10726)
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 2,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2114/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2114/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2113
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2113/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2113/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2113/events
|
https://github.com/huggingface/datasets/pull/2113
| 841,191,303 |
MDExOlB1bGxSZXF1ZXN0NjAwOTYxMDEz
| 2,113 |
Implement Dataset as context manager
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-25T18:18:30Z
|
2021-03-31T11:30:14Z
|
2021-03-31T08:30:11Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2113.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2113",
"merged_at": "2021-03-31T08:30:11Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2113.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2113"
}
|
When used as context manager, it would be safely deleted if some exception is raised.
This will avoid
> During handling of the above exception, another exception occurred:
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2113/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2113/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2112
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2112/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2112/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2112/events
|
https://github.com/huggingface/datasets/pull/2112
| 841,098,008 |
MDExOlB1bGxSZXF1ZXN0NjAwODgyMjA0
| 2,112 |
Support for legal NLP datasets (EURLEX and ECtHR cases)
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/1626984?v=4",
"events_url": "https://api.github.com/users/iliaschalkidis/events{/privacy}",
"followers_url": "https://api.github.com/users/iliaschalkidis/followers",
"following_url": "https://api.github.com/users/iliaschalkidis/following{/other_user}",
"gists_url": "https://api.github.com/users/iliaschalkidis/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/iliaschalkidis",
"id": 1626984,
"login": "iliaschalkidis",
"node_id": "MDQ6VXNlcjE2MjY5ODQ=",
"organizations_url": "https://api.github.com/users/iliaschalkidis/orgs",
"received_events_url": "https://api.github.com/users/iliaschalkidis/received_events",
"repos_url": "https://api.github.com/users/iliaschalkidis/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/iliaschalkidis/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/iliaschalkidis/subscriptions",
"type": "User",
"url": "https://api.github.com/users/iliaschalkidis"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-25T16:24:17Z
|
2021-03-25T18:39:31Z
|
2021-03-25T18:34:31Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2112.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2112",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2112.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2112"
}
|
Add support for two legal NLP datasets:
- EURLEX (https://www.aclweb.org/anthology/P19-1636/)
- ECtHR cases (https://arxiv.org/abs/2103.13084)
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2112/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2112/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2111
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2111/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2111/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2111/events
|
https://github.com/huggingface/datasets/pull/2111
| 841,082,087 |
MDExOlB1bGxSZXF1ZXN0NjAwODY4OTg5
| 2,111 |
Compute WER metric iteratively
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-25T16:06:48Z
|
2021-04-06T07:20:43Z
|
2021-04-06T07:20:43Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2111.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2111",
"merged_at": "2021-04-06T07:20:43Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2111.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2111"
}
|
Compute WER metric iteratively to avoid MemoryError.
Fix #2078.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2111/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2111/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2110
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2110/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2110/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2110/events
|
https://github.com/huggingface/datasets/pull/2110
| 840,794,995 |
MDExOlB1bGxSZXF1ZXN0NjAwNjI1NDQ5
| 2,110 |
Fix incorrect assertion in builder.py
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/2340721?v=4",
"events_url": "https://api.github.com/users/dreamgonfly/events{/privacy}",
"followers_url": "https://api.github.com/users/dreamgonfly/followers",
"following_url": "https://api.github.com/users/dreamgonfly/following{/other_user}",
"gists_url": "https://api.github.com/users/dreamgonfly/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/dreamgonfly",
"id": 2340721,
"login": "dreamgonfly",
"node_id": "MDQ6VXNlcjIzNDA3MjE=",
"organizations_url": "https://api.github.com/users/dreamgonfly/orgs",
"received_events_url": "https://api.github.com/users/dreamgonfly/received_events",
"repos_url": "https://api.github.com/users/dreamgonfly/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/dreamgonfly/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dreamgonfly/subscriptions",
"type": "User",
"url": "https://api.github.com/users/dreamgonfly"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-25T10:39:20Z
|
2021-04-12T13:33:03Z
|
2021-04-12T13:33:03Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2110.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2110",
"merged_at": "2021-04-12T13:33:03Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2110.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2110"
}
|
Fix incorrect num_examples comparison assertion in builder.py
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2110/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2110/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2109
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2109/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2109/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2109/events
|
https://github.com/huggingface/datasets/pull/2109
| 840,746,598 |
MDExOlB1bGxSZXF1ZXN0NjAwNTg1MzM5
| 2,109 |
Add more issue templates and customize issue template chooser
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-25T09:41:53Z
|
2021-04-19T06:20:11Z
|
2021-04-19T06:20:11Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2109.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2109",
"merged_at": "2021-04-19T06:20:10Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2109.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2109"
}
|
When opening an issue, it is not evident for the users how to choose a blank issue template. There is a link at the bottom of all the other issue templates (`Don’t see your issue here? Open a blank issue.`), but this is not very visible for users. This is the reason why many users finally chose the `add-dataset` template instead (this is more visible) for issues that indeed are not requesting the addition of a new dataset.
~~With this PR, the default blank issue template would be as visible as the other templates (as the `add-dataset` template), thus making easier for the users to choose it.~~
With this PR:
- more issue templates, besides `add-dataset`, are added: `bug-report` and `feature-request`
- the issue template chooser is customized, so that it now includes a link to `Discussions` for questions
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2109/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2109/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2108
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2108/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2108/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2108/events
|
https://github.com/huggingface/datasets/issues/2108
| 840,181,055 |
MDU6SXNzdWU4NDAxODEwNTU=
| 2,108 |
Is there a way to use a GPU only when training an Index in the process of add_faisis_index?
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4",
"events_url": "https://api.github.com/users/shamanez/events{/privacy}",
"followers_url": "https://api.github.com/users/shamanez/followers",
"following_url": "https://api.github.com/users/shamanez/following{/other_user}",
"gists_url": "https://api.github.com/users/shamanez/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/shamanez",
"id": 16892570,
"login": "shamanez",
"node_id": "MDQ6VXNlcjE2ODkyNTcw",
"organizations_url": "https://api.github.com/users/shamanez/orgs",
"received_events_url": "https://api.github.com/users/shamanez/received_events",
"repos_url": "https://api.github.com/users/shamanez/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/shamanez/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shamanez/subscriptions",
"type": "User",
"url": "https://api.github.com/users/shamanez"
}
|
[
{
"color": "d876e3",
"default": true,
"description": "Further information is requested",
"id": 1935892912,
"name": "question",
"node_id": "MDU6TGFiZWwxOTM1ODkyOTEy",
"url": "https://api.github.com/repos/huggingface/datasets/labels/question"
}
] |
open
| false | null |
[] | null |
[] |
2021-03-24T21:32:16Z
|
2021-03-25T06:31:43Z
| null |
NONE
| null | null | null |
Motivation - Some FAISS indexes like IVF consist of the training step that clusters the dataset into a given number of indexes. It would be nice if we can use a GPU to do the training step and covert the index back to CPU as mention in [this faiss example](https://gist.github.com/mdouze/46d6bbbaabca0b9778fca37ed2bcccf6).
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2108/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2108/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2107
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2107/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2107/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2107/events
|
https://github.com/huggingface/datasets/pull/2107
| 839,495,825 |
MDExOlB1bGxSZXF1ZXN0NTk5NTAxODE5
| 2,107 |
Metadata validation
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
|
[] |
closed
| false |
{
"avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4",
"events_url": "https://api.github.com/users/SBrandeis/events{/privacy}",
"followers_url": "https://api.github.com/users/SBrandeis/followers",
"following_url": "https://api.github.com/users/SBrandeis/following{/other_user}",
"gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/SBrandeis",
"id": 33657802,
"login": "SBrandeis",
"node_id": "MDQ6VXNlcjMzNjU3ODAy",
"organizations_url": "https://api.github.com/users/SBrandeis/orgs",
"received_events_url": "https://api.github.com/users/SBrandeis/received_events",
"repos_url": "https://api.github.com/users/SBrandeis/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions",
"type": "User",
"url": "https://api.github.com/users/SBrandeis"
}
|
[
{
"avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4",
"events_url": "https://api.github.com/users/SBrandeis/events{/privacy}",
"followers_url": "https://api.github.com/users/SBrandeis/followers",
"following_url": "https://api.github.com/users/SBrandeis/following{/other_user}",
"gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/SBrandeis",
"id": 33657802,
"login": "SBrandeis",
"node_id": "MDQ6VXNlcjMzNjU3ODAy",
"organizations_url": "https://api.github.com/users/SBrandeis/orgs",
"received_events_url": "https://api.github.com/users/SBrandeis/received_events",
"repos_url": "https://api.github.com/users/SBrandeis/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions",
"type": "User",
"url": "https://api.github.com/users/SBrandeis"
}
] | null |
[] |
2021-03-24T08:52:41Z
|
2021-04-26T08:27:14Z
|
2021-04-26T08:27:13Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2107.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2107",
"merged_at": "2021-04-26T08:27:13Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2107.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2107"
}
|
- `pydantic` metadata schema with dedicated validators against our taxonomy
- ci script to validate new changes against this schema and start a vertuous loop
- soft validation on tasks ids since we expect the taxonomy to undergo some changes in the near future
for reference with the current validation we have ~365~ 378 datasets with invalid metadata! full error report [_here_.](https://gist.github.com/theo-m/61b3c0c47fc6121d08d3174bd4c2a26b)
|
{
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2107/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2107/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2106
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2106/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2106/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2106/events
|
https://github.com/huggingface/datasets/issues/2106
| 839,084,264 |
MDU6SXNzdWU4MzkwODQyNjQ=
| 2,106 |
WMT19 Dataset for Kazakh-English is not formatted correctly
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/22580542?v=4",
"events_url": "https://api.github.com/users/trina731/events{/privacy}",
"followers_url": "https://api.github.com/users/trina731/followers",
"following_url": "https://api.github.com/users/trina731/following{/other_user}",
"gists_url": "https://api.github.com/users/trina731/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/trina731",
"id": 22580542,
"login": "trina731",
"node_id": "MDQ6VXNlcjIyNTgwNTQy",
"organizations_url": "https://api.github.com/users/trina731/orgs",
"received_events_url": "https://api.github.com/users/trina731/received_events",
"repos_url": "https://api.github.com/users/trina731/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/trina731/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/trina731/subscriptions",
"type": "User",
"url": "https://api.github.com/users/trina731"
}
|
[
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] |
open
| false | null |
[] | null |
[] |
2021-03-23T20:14:47Z
|
2021-03-25T21:36:20Z
| null |
NONE
| null | null | null |
In addition to the bug of languages being switched from Issue @415, there are incorrect translations in the dataset because the English-Kazakh translations have a one off formatting error.
The News Commentary v14 parallel data set for kk-en from http://www.statmt.org/wmt19/translation-task.html has a bug here:
> Line 94. The Swiss National Bank, for its part, has been battling with the deflationary effects of the franc’s dramatic appreciation over the past few years. Швейцарияның Ұлттық банкі өз тарапынан, соңғы бірнеше жыл ішінде франк құнының қатты өсуінің дефляциялық әсерімен күресіп келеді.
>
> Line 95. Дефляциялық күштер 2008 жылы терең және ұзаққа созылған жаһандық дағдарысқа байланысты орын алған ірі экономикалық және қаржылық орын алмасулардың арқасында босатылды. Жеке қарыз қаражаты үлесінің қысқаруы орталық банктің рефляцияға жұмсалған күш-жігеріне тұрақты соққан қарсы желдей болды.
>
> Line 96. The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008. Private deleveraging became a steady headwind to central bank efforts to reflate. 2009 жылы, алдыңғы қатарлы экономикалардың шамамен үштен бірі бағаның төмендеуін көрсетті, бұл соғыстан кейінгі жоғары деңгей болды.
As you can see, line 95 has only the Kazakh translation which should be part of line 96. This causes all of the following English-Kazakh translation pairs to be one off rendering ALL of those translations incorrect. This issue was not fixed when the dataset was imported to Huggingface. By running this code
```
import datasets
from datasets import load_dataset
dataset = load_dataset('wmt19', 'kk-en')
for key in dataset['train']['translation']:
if 'The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008.' in key['kk']:
print(key['en'])
print(key['kk'])
break
```
we get:
> 2009 жылы, алдыңғы қатарлы экономикалардың шамамен үштен бірі бағаның төмендеуін көрсетті, бұл соғыстан кейінгі жоғары деңгей болды.
> The deflationary forces were unleashed by the major economic and financial dislocations associated with the deep and protracted global crisis that erupted in 2008. Private deleveraging became a steady headwind to central bank efforts to reflate.
which shows that the issue still persists in the Huggingface dataset. The Kazakh sentence matches up to the next English sentence in the dataset instead of the current one.
Please let me know if there's you have any ideas to fix this one-off error from the dataset or if this can be fixed by Huggingface.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2106/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2106/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2105
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2105/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2105/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2105/events
|
https://github.com/huggingface/datasets/issues/2105
| 839,059,226 |
MDU6SXNzdWU4MzkwNTkyMjY=
| 2,105 |
Request to remove S2ORC dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/13603748?v=4",
"events_url": "https://api.github.com/users/kyleclo/events{/privacy}",
"followers_url": "https://api.github.com/users/kyleclo/followers",
"following_url": "https://api.github.com/users/kyleclo/following{/other_user}",
"gists_url": "https://api.github.com/users/kyleclo/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/kyleclo",
"id": 13603748,
"login": "kyleclo",
"node_id": "MDQ6VXNlcjEzNjAzNzQ4",
"organizations_url": "https://api.github.com/users/kyleclo/orgs",
"received_events_url": "https://api.github.com/users/kyleclo/received_events",
"repos_url": "https://api.github.com/users/kyleclo/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/kyleclo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kyleclo/subscriptions",
"type": "User",
"url": "https://api.github.com/users/kyleclo"
}
|
[] |
open
| false | null |
[] | null |
[] |
2021-03-23T19:43:06Z
|
2021-08-04T19:18:02Z
| null |
NONE
| null | null | null |
Hi! I was wondering if it's possible to remove [S2ORC](https://huggingface.co/datasets/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 1,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2105/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2105/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2104
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2104/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2104/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2104/events
|
https://github.com/huggingface/datasets/issues/2104
| 839,027,834 |
MDU6SXNzdWU4MzkwMjc4MzQ=
| 2,104 |
Trouble loading wiki_movies
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/35391599?v=4",
"events_url": "https://api.github.com/users/adityaarunsinghal/events{/privacy}",
"followers_url": "https://api.github.com/users/adityaarunsinghal/followers",
"following_url": "https://api.github.com/users/adityaarunsinghal/following{/other_user}",
"gists_url": "https://api.github.com/users/adityaarunsinghal/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/adityaarunsinghal",
"id": 35391599,
"login": "adityaarunsinghal",
"node_id": "MDQ6VXNlcjM1MzkxNTk5",
"organizations_url": "https://api.github.com/users/adityaarunsinghal/orgs",
"received_events_url": "https://api.github.com/users/adityaarunsinghal/received_events",
"repos_url": "https://api.github.com/users/adityaarunsinghal/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/adityaarunsinghal/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/adityaarunsinghal/subscriptions",
"type": "User",
"url": "https://api.github.com/users/adityaarunsinghal"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-23T18:59:54Z
|
2022-03-30T08:22:58Z
|
2022-03-30T08:22:58Z
|
NONE
| null | null | null |
Hello,
I am trying to load_dataset("wiki_movies") and it gives me this error -
`FileNotFoundError: Couldn't find file locally at wiki_movies/wiki_movies.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.2/datasets/wiki_movies/wiki_movies.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/wiki_movies/wiki_movies.py`
Trying to do `python run_mlm.py \
--model_name_or_path roberta-base \
--dataset_name wiki_movies \` also gives the same error.
Is this something on my end? From what I can tell, this dataset was re-added by @lhoestq a few months ago.
Thank you!
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2104/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2104/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2103
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2103/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2103/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2103/events
|
https://github.com/huggingface/datasets/issues/2103
| 838,946,916 |
MDU6SXNzdWU4Mzg5NDY5MTY=
| 2,103 |
citation, homepage, and license fields of `dataset_info.json` are duplicated many times
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/15007950?v=4",
"events_url": "https://api.github.com/users/samsontmr/events{/privacy}",
"followers_url": "https://api.github.com/users/samsontmr/followers",
"following_url": "https://api.github.com/users/samsontmr/following{/other_user}",
"gists_url": "https://api.github.com/users/samsontmr/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/samsontmr",
"id": 15007950,
"login": "samsontmr",
"node_id": "MDQ6VXNlcjE1MDA3OTUw",
"organizations_url": "https://api.github.com/users/samsontmr/orgs",
"received_events_url": "https://api.github.com/users/samsontmr/received_events",
"repos_url": "https://api.github.com/users/samsontmr/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/samsontmr/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/samsontmr/subscriptions",
"type": "User",
"url": "https://api.github.com/users/samsontmr"
}
|
[
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "7057ff",
"default": true,
"description": "Good for newcomers",
"id": 1935892877,
"name": "good first issue",
"node_id": "MDU6TGFiZWwxOTM1ODkyODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-23T17:18:09Z
|
2021-04-06T14:39:59Z
|
2021-04-06T14:39:59Z
|
NONE
| null | null | null |
This happens after a `map` operation when `num_proc` is set to `>1`. I tested this by cleaning up the json before running the `map` op on the dataset so it's unlikely it's coming from an earlier concatenation.
Example result:
```
"citation": "@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n\n@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n
```
@lhoestq and I believe this is happening due to the fields being concatenated `num_proc` times.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2103/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2103/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2102
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2102/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2102/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2102/events
|
https://github.com/huggingface/datasets/pull/2102
| 838,794,090 |
MDExOlB1bGxSZXF1ZXN0NTk4OTEyNzUw
| 2,102 |
Move Dataset.to_csv to csv module
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[
{
"color": "B67A40",
"default": false,
"description": "Restructuring existing code without changing its external behavior",
"id": 2851292821,
"name": "refactoring",
"node_id": "MDU6TGFiZWwyODUxMjkyODIx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/refactoring"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-23T14:35:46Z
|
2021-03-24T14:07:35Z
|
2021-03-24T14:07:34Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2102.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2102",
"merged_at": "2021-03-24T14:07:34Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2102.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2102"
}
|
Move the implementation of `Dataset.to_csv` to module `datasets.io.csv`.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2102/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2102/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2101
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2101/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2101/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2101/events
|
https://github.com/huggingface/datasets/pull/2101
| 838,586,184 |
MDExOlB1bGxSZXF1ZXN0NTk4NzQzMDM4
| 2,101 |
MIAM dataset - new citation details
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/1551356?v=4",
"events_url": "https://api.github.com/users/eusip/events{/privacy}",
"followers_url": "https://api.github.com/users/eusip/followers",
"following_url": "https://api.github.com/users/eusip/following{/other_user}",
"gists_url": "https://api.github.com/users/eusip/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/eusip",
"id": 1551356,
"login": "eusip",
"node_id": "MDQ6VXNlcjE1NTEzNTY=",
"organizations_url": "https://api.github.com/users/eusip/orgs",
"received_events_url": "https://api.github.com/users/eusip/received_events",
"repos_url": "https://api.github.com/users/eusip/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/eusip/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/eusip/subscriptions",
"type": "User",
"url": "https://api.github.com/users/eusip"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-23T10:41:23Z
|
2021-03-23T18:08:10Z
|
2021-03-23T18:08:10Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2101.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2101",
"merged_at": "2021-03-23T18:08:09Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2101.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2101"
}
|
Hi @lhoestq, I have updated the citations to reference an OpenReview preprint.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2101/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2101/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2100
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2100/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2100/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2100/events
|
https://github.com/huggingface/datasets/pull/2100
| 838,574,631 |
MDExOlB1bGxSZXF1ZXN0NTk4NzMzOTM0
| 2,100 |
Fix deprecated warning message and docstring
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[
{
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation",
"id": 1935892861,
"name": "documentation",
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-23T10:27:52Z
|
2021-03-24T08:19:41Z
|
2021-03-23T18:03:49Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2100.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2100",
"merged_at": "2021-03-23T18:03:49Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2100.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2100"
}
|
Fix deprecated warnings:
- Use deprecated Sphinx directive in docstring
- Fix format of deprecated message
- Raise FutureWarning
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2100/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2100/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2099
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2099/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2099/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2099/events
|
https://github.com/huggingface/datasets/issues/2099
| 838,523,819 |
MDU6SXNzdWU4Mzg1MjM4MTk=
| 2,099 |
load_from_disk takes a long time to load local dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/15007950?v=4",
"events_url": "https://api.github.com/users/samsontmr/events{/privacy}",
"followers_url": "https://api.github.com/users/samsontmr/followers",
"following_url": "https://api.github.com/users/samsontmr/following{/other_user}",
"gists_url": "https://api.github.com/users/samsontmr/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/samsontmr",
"id": 15007950,
"login": "samsontmr",
"node_id": "MDQ6VXNlcjE1MDA3OTUw",
"organizations_url": "https://api.github.com/users/samsontmr/orgs",
"received_events_url": "https://api.github.com/users/samsontmr/received_events",
"repos_url": "https://api.github.com/users/samsontmr/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/samsontmr/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/samsontmr/subscriptions",
"type": "User",
"url": "https://api.github.com/users/samsontmr"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-23T09:28:37Z
|
2021-03-23T17:12:16Z
|
2021-03-23T17:12:16Z
|
NONE
| null | null | null |
I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helping (the total size seems to be smaller though).
Does anyone know what could be the issue? Or does the casting of that column to `int8` need to happen in the function that writes the arrow table instead of in the `map` where I create the list of integers?
Tagging @lhoestq since you seem to be working on these issues and PRs :)
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2099/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2099/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2098
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2098/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2098/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2098/events
|
https://github.com/huggingface/datasets/issues/2098
| 838,447,959 |
MDU6SXNzdWU4Mzg0NDc5NTk=
| 2,098 |
SQuAD version
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/39556019?v=4",
"events_url": "https://api.github.com/users/h-peng17/events{/privacy}",
"followers_url": "https://api.github.com/users/h-peng17/followers",
"following_url": "https://api.github.com/users/h-peng17/following{/other_user}",
"gists_url": "https://api.github.com/users/h-peng17/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/h-peng17",
"id": 39556019,
"login": "h-peng17",
"node_id": "MDQ6VXNlcjM5NTU2MDE5",
"organizations_url": "https://api.github.com/users/h-peng17/orgs",
"received_events_url": "https://api.github.com/users/h-peng17/received_events",
"repos_url": "https://api.github.com/users/h-peng17/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/h-peng17/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/h-peng17/subscriptions",
"type": "User",
"url": "https://api.github.com/users/h-peng17"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-23T07:47:54Z
|
2021-03-26T09:48:54Z
|
2021-03-26T09:48:54Z
|
NONE
| null | null | null |
Hi~
I want train on squad dataset. What's the version of the squad? Is it 1.1 or 1.0? I'm new in QA, I don't find some descriptions about it.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2098/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2098/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2097
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2097/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2097/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2097/events
|
https://github.com/huggingface/datasets/pull/2097
| 838,105,289 |
MDExOlB1bGxSZXF1ZXN0NTk4MzM4MTA3
| 2,097 |
fixes issue #1110 by descending further if `obj["_type"]` is a dict
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/15979778?v=4",
"events_url": "https://api.github.com/users/dcfidalgo/events{/privacy}",
"followers_url": "https://api.github.com/users/dcfidalgo/followers",
"following_url": "https://api.github.com/users/dcfidalgo/following{/other_user}",
"gists_url": "https://api.github.com/users/dcfidalgo/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/dcfidalgo",
"id": 15979778,
"login": "dcfidalgo",
"node_id": "MDQ6VXNlcjE1OTc5Nzc4",
"organizations_url": "https://api.github.com/users/dcfidalgo/orgs",
"received_events_url": "https://api.github.com/users/dcfidalgo/received_events",
"repos_url": "https://api.github.com/users/dcfidalgo/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/dcfidalgo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dcfidalgo/subscriptions",
"type": "User",
"url": "https://api.github.com/users/dcfidalgo"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-22T21:00:55Z
|
2021-03-22T21:01:11Z
|
2021-03-22T21:01:11Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2097.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2097",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2097.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2097"
}
|
Check metrics
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2097/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2097/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2096
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2096/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2096/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2096/events
|
https://github.com/huggingface/datasets/issues/2096
| 838,038,379 |
MDU6SXNzdWU4MzgwMzgzNzk=
| 2,096 |
CoNLL 2003 dataset not including German
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8406802?v=4",
"events_url": "https://api.github.com/users/rxian/events{/privacy}",
"followers_url": "https://api.github.com/users/rxian/followers",
"following_url": "https://api.github.com/users/rxian/following{/other_user}",
"gists_url": "https://api.github.com/users/rxian/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rxian",
"id": 8406802,
"login": "rxian",
"node_id": "MDQ6VXNlcjg0MDY4MDI=",
"organizations_url": "https://api.github.com/users/rxian/orgs",
"received_events_url": "https://api.github.com/users/rxian/received_events",
"repos_url": "https://api.github.com/users/rxian/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rxian/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rxian/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rxian"
}
|
[
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] |
open
| false | null |
[] | null |
[] |
2021-03-22T19:23:56Z
|
2022-09-07T23:34:30Z
| null |
NONE
| null | null | null |
Hello, thanks for all the work on developing and maintaining this amazing platform, which I am enjoying working with!
I was wondering if there is a reason why the German CoNLL 2003 dataset is not included in the [repository](https://github.com/huggingface/datasets/tree/master/datasets/conll2003), since a copy of it could be found in some places on the internet such as GitHub? I could help adding the German data to the hub, unless there are some copyright issues that I am unaware of...
This is considering that many work use the union of CoNLL 2002 and 2003 datasets for comparing cross-lingual NER transfer performance in `en`, `de`, `es`, and `nl`. E.g., [XLM-R](https://www.aclweb.org/anthology/2020.acl-main.747.pdf).
## Adding a Dataset
- **Name:** CoNLL 2003 German
- **Paper:** https://www.aclweb.org/anthology/W03-0419/
- **Data:** https://github.com/huggingface/datasets/tree/master/datasets/conll2003
|
{
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2096/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2096/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2093
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2093/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2093/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2093/events
|
https://github.com/huggingface/datasets/pull/2093
| 837,209,211 |
MDExOlB1bGxSZXF1ZXN0NTk3NTgyNjUx
| 2,093 |
Fix: Allows a feature to be named "_type"
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/15979778?v=4",
"events_url": "https://api.github.com/users/dcfidalgo/events{/privacy}",
"followers_url": "https://api.github.com/users/dcfidalgo/followers",
"following_url": "https://api.github.com/users/dcfidalgo/following{/other_user}",
"gists_url": "https://api.github.com/users/dcfidalgo/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/dcfidalgo",
"id": 15979778,
"login": "dcfidalgo",
"node_id": "MDQ6VXNlcjE1OTc5Nzc4",
"organizations_url": "https://api.github.com/users/dcfidalgo/orgs",
"received_events_url": "https://api.github.com/users/dcfidalgo/received_events",
"repos_url": "https://api.github.com/users/dcfidalgo/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/dcfidalgo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dcfidalgo/subscriptions",
"type": "User",
"url": "https://api.github.com/users/dcfidalgo"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-21T23:21:57Z
|
2021-03-25T14:35:54Z
|
2021-03-25T14:35:54Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2093.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2093",
"merged_at": "2021-03-25T14:35:54Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2093.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2093"
}
|
This PR tries to fix issue #1110. Sorry for taking so long to come back to this.
It's a simple fix, but i am not sure if it works for all possible types of `obj`. Let me know what you think @lhoestq
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2093/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2093/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2092
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2092/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2092/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2092/events
|
https://github.com/huggingface/datasets/issues/2092
| 836,984,043 |
MDU6SXNzdWU4MzY5ODQwNDM=
| 2,092 |
How to disable making arrow tables in load_dataset ?
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/48825663?v=4",
"events_url": "https://api.github.com/users/Jeevesh8/events{/privacy}",
"followers_url": "https://api.github.com/users/Jeevesh8/followers",
"following_url": "https://api.github.com/users/Jeevesh8/following{/other_user}",
"gists_url": "https://api.github.com/users/Jeevesh8/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Jeevesh8",
"id": 48825663,
"login": "Jeevesh8",
"node_id": "MDQ6VXNlcjQ4ODI1NjYz",
"organizations_url": "https://api.github.com/users/Jeevesh8/orgs",
"received_events_url": "https://api.github.com/users/Jeevesh8/received_events",
"repos_url": "https://api.github.com/users/Jeevesh8/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Jeevesh8/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Jeevesh8/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Jeevesh8"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-21T04:50:07Z
|
2022-06-01T16:49:52Z
|
2022-06-01T16:49:52Z
|
NONE
| null | null | null |
Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2092/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2092/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2091
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2091/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2091/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2091/events
|
https://github.com/huggingface/datasets/pull/2091
| 836,831,403 |
MDExOlB1bGxSZXF1ZXN0NTk3Mjk4ODI3
| 2,091 |
Fix copy snippet in docs
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko"
}
|
[
{
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation",
"id": 1935892861,
"name": "documentation",
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-20T15:08:22Z
|
2021-03-24T08:20:50Z
|
2021-03-23T17:18:31Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2091.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2091",
"merged_at": "2021-03-23T17:18:31Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2091.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2091"
}
|
With this change the lines starting with `...` in the code blocks can be properly copied to clipboard.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2091/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2091/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2090
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2090/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2090/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2090/events
|
https://github.com/huggingface/datasets/pull/2090
| 836,807,498 |
MDExOlB1bGxSZXF1ZXN0NTk3MjgwNTEy
| 2,090 |
Add machine translated multilingual STS benchmark dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/229382?v=4",
"events_url": "https://api.github.com/users/PhilipMay/events{/privacy}",
"followers_url": "https://api.github.com/users/PhilipMay/followers",
"following_url": "https://api.github.com/users/PhilipMay/following{/other_user}",
"gists_url": "https://api.github.com/users/PhilipMay/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/PhilipMay",
"id": 229382,
"login": "PhilipMay",
"node_id": "MDQ6VXNlcjIyOTM4Mg==",
"organizations_url": "https://api.github.com/users/PhilipMay/orgs",
"received_events_url": "https://api.github.com/users/PhilipMay/received_events",
"repos_url": "https://api.github.com/users/PhilipMay/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/PhilipMay/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/PhilipMay/subscriptions",
"type": "User",
"url": "https://api.github.com/users/PhilipMay"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-20T13:28:07Z
|
2021-03-29T13:24:42Z
|
2021-03-29T13:00:15Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2090.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2090",
"merged_at": "2021-03-29T13:00:15Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2090.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2090"
}
|
also see here https://github.com/PhilipMay/stsb-multi-mt
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2090/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2090/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2089
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2089/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2089/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2089/events
|
https://github.com/huggingface/datasets/issues/2089
| 836,788,019 |
MDU6SXNzdWU4MzY3ODgwMTk=
| 2,089 |
Add documentaton for dataset README.md files
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/229382?v=4",
"events_url": "https://api.github.com/users/PhilipMay/events{/privacy}",
"followers_url": "https://api.github.com/users/PhilipMay/followers",
"following_url": "https://api.github.com/users/PhilipMay/following{/other_user}",
"gists_url": "https://api.github.com/users/PhilipMay/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/PhilipMay",
"id": 229382,
"login": "PhilipMay",
"node_id": "MDQ6VXNlcjIyOTM4Mg==",
"organizations_url": "https://api.github.com/users/PhilipMay/orgs",
"received_events_url": "https://api.github.com/users/PhilipMay/received_events",
"repos_url": "https://api.github.com/users/PhilipMay/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/PhilipMay/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/PhilipMay/subscriptions",
"type": "User",
"url": "https://api.github.com/users/PhilipMay"
}
|
[] |
open
| false | null |
[] | null |
[] |
2021-03-20T11:44:38Z
|
2021-07-12T17:41:40Z
| null |
CONTRIBUTOR
| null | null | null |
Hi,
the dataset README files have special headers.
Somehow a documenation of the allowed values and tags is missing.
Could you add that?
Just to give some concrete questions that should be answered imo:
- which values can be passted to multilinguality?
- what should be passed to language_creators?
- which values should licenses have? What do I say when it is a custom license? Should I add a link?
- how should I choose size_categories ? What are valid ranges?
- what are valid task_categories?
Thanks
Philip
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2089/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2089/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2088
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2088/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2088/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2088/events
|
https://github.com/huggingface/datasets/pull/2088
| 836,763,733 |
MDExOlB1bGxSZXF1ZXN0NTk3MjQ4Mzk1
| 2,088 |
change bibtex template to author instead of authors
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/229382?v=4",
"events_url": "https://api.github.com/users/PhilipMay/events{/privacy}",
"followers_url": "https://api.github.com/users/PhilipMay/followers",
"following_url": "https://api.github.com/users/PhilipMay/following{/other_user}",
"gists_url": "https://api.github.com/users/PhilipMay/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/PhilipMay",
"id": 229382,
"login": "PhilipMay",
"node_id": "MDQ6VXNlcjIyOTM4Mg==",
"organizations_url": "https://api.github.com/users/PhilipMay/orgs",
"received_events_url": "https://api.github.com/users/PhilipMay/received_events",
"repos_url": "https://api.github.com/users/PhilipMay/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/PhilipMay/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/PhilipMay/subscriptions",
"type": "User",
"url": "https://api.github.com/users/PhilipMay"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-20T09:23:44Z
|
2021-03-23T15:40:12Z
|
2021-03-23T15:40:12Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2088.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2088",
"merged_at": "2021-03-23T15:40:12Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2088.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2088"
}
|
Hi,
IMO when using BibTex Author should be used instead of Authors.
See here: http://www.bibtex.org/Using/de/
Thanks
Philip
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2088/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2088/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2087
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2087/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2087/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2087/events
|
https://github.com/huggingface/datasets/pull/2087
| 836,587,392 |
MDExOlB1bGxSZXF1ZXN0NTk3MDg4NTk2
| 2,087 |
Update metadata if dataset features are modified
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-20T02:05:23Z
|
2021-04-09T09:25:33Z
|
2021-04-09T09:25:33Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2087.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2087",
"merged_at": "2021-04-09T09:25:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2087.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2087"
}
|
This PR adds a decorator that updates the dataset metadata if a previously executed transform modifies its features.
Fixes #2083
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 1,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2087/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2087/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2086
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2086/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2086/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2086/events
|
https://github.com/huggingface/datasets/pull/2086
| 836,249,587 |
MDExOlB1bGxSZXF1ZXN0NTk2Nzg0Mjcz
| 2,086 |
change user permissions to -rw-r--r--
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4",
"events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}",
"followers_url": "https://api.github.com/users/bhavitvyamalik/followers",
"following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}",
"gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/bhavitvyamalik",
"id": 19718818,
"login": "bhavitvyamalik",
"node_id": "MDQ6VXNlcjE5NzE4ODE4",
"organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs",
"received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events",
"repos_url": "https://api.github.com/users/bhavitvyamalik/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions",
"type": "User",
"url": "https://api.github.com/users/bhavitvyamalik"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-19T18:14:56Z
|
2021-03-24T13:59:04Z
|
2021-03-24T13:59:04Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2086.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2086",
"merged_at": "2021-03-24T13:59:04Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2086.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2086"
}
|
Fix for #2065
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2086/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2086/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2085
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2085/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2085/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2085/events
|
https://github.com/huggingface/datasets/pull/2085
| 835,870,994 |
MDExOlB1bGxSZXF1ZXN0NTk2NDYyOTc2
| 2,085 |
Fix max_wait_time in requests
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-19T11:22:26Z
|
2021-03-23T15:36:38Z
|
2021-03-23T15:36:37Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2085.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2085",
"merged_at": "2021-03-23T15:36:37Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2085.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2085"
}
|
it was handled as a min time, not max cc @SBrandeis
|
{
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2085/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2085/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2084
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2084/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2084/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2084/events
|
https://github.com/huggingface/datasets/issues/2084
| 835,750,671 |
MDU6SXNzdWU4MzU3NTA2NzE=
| 2,084 |
CUAD - Contract Understanding Atticus Dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
|
[
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-19T09:27:43Z
|
2021-04-16T08:50:44Z
|
2021-04-16T08:50:44Z
|
CONTRIBUTOR
| null | null | null |
## Adding a Dataset
- **Name:** CUAD - Contract Understanding Atticus Dataset
- **Description:** As one of the only large, specialized NLP benchmarks annotated by experts, CUAD can serve as a challenging research benchmark for the broader NLP community.
- **Paper:** https://arxiv.org/abs/2103.06268
- **Data:** https://github.com/TheAtticusProject/cuad/
- **Motivation:** good domain specific datasets are valuable
Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2084/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2084/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2083
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2083/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2083/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2083/events
|
https://github.com/huggingface/datasets/issues/2083
| 835,695,425 |
MDU6SXNzdWU4MzU2OTU0MjU=
| 2,083 |
`concatenate_datasets` throws error when changing the order of datasets to concatenate
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}",
"followers_url": "https://api.github.com/users/patrickvonplaten/followers",
"following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}",
"gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/patrickvonplaten",
"id": 23423619,
"login": "patrickvonplaten",
"node_id": "MDQ6VXNlcjIzNDIzNjE5",
"organizations_url": "https://api.github.com/users/patrickvonplaten/orgs",
"received_events_url": "https://api.github.com/users/patrickvonplaten/received_events",
"repos_url": "https://api.github.com/users/patrickvonplaten/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions",
"type": "User",
"url": "https://api.github.com/users/patrickvonplaten"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-19T08:29:48Z
|
2021-04-09T09:25:33Z
|
2021-04-09T09:25:33Z
|
MEMBER
| null | null | null |
Hey,
I played around with the `concatenate_datasets(...)` function: https://huggingface.co/docs/datasets/package_reference/main_classes.html?highlight=concatenate_datasets#datasets.concatenate_datasets
and noticed that when the order in which the datasets are concatenated changes an error is thrown where it should not IMO.
Here is a google colab to reproduce the error: https://colab.research.google.com/drive/17VTFU4KQ735-waWZJjeOHS6yDTfV5ekK?usp=sharing
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2083/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2083/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2082
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2082/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2082/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2082/events
|
https://github.com/huggingface/datasets/pull/2082
| 835,401,555 |
MDExOlB1bGxSZXF1ZXN0NTk2MDY1NTM0
| 2,082 |
Updated card using information from data statement and datasheet
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4",
"events_url": "https://api.github.com/users/mcmillanmajora/events{/privacy}",
"followers_url": "https://api.github.com/users/mcmillanmajora/followers",
"following_url": "https://api.github.com/users/mcmillanmajora/following{/other_user}",
"gists_url": "https://api.github.com/users/mcmillanmajora/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mcmillanmajora",
"id": 26722925,
"login": "mcmillanmajora",
"node_id": "MDQ6VXNlcjI2NzIyOTI1",
"organizations_url": "https://api.github.com/users/mcmillanmajora/orgs",
"received_events_url": "https://api.github.com/users/mcmillanmajora/received_events",
"repos_url": "https://api.github.com/users/mcmillanmajora/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mcmillanmajora/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mcmillanmajora/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mcmillanmajora"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-19T00:39:38Z
|
2021-03-19T14:29:09Z
|
2021-03-19T14:29:09Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2082.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2082",
"merged_at": "2021-03-19T14:29:08Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2082.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2082"
}
|
I updated and clarified the REFreSD [data card](https://github.com/mcmillanmajora/datasets/blob/refresd_card/datasets/refresd/README.md) with information from the Eleftheria's [website](https://elbria.github.io/post/refresd/). I added brief descriptions where the initial card referred to the paper, and I also recreated some of the tables in the paper to show relevant dataset statistics.
I'll email Eleftheria to see if she has any comments on the card.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2082/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2082/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2081
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2081/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2081/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2081/events
|
https://github.com/huggingface/datasets/pull/2081
| 835,112,968 |
MDExOlB1bGxSZXF1ZXN0NTk1ODE3OTM4
| 2,081 |
Fix docstrings issues
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[
{
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation",
"id": 1935892861,
"name": "documentation",
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-18T18:11:01Z
|
2021-04-07T14:37:43Z
|
2021-04-07T14:37:43Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2081.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2081",
"merged_at": "2021-04-07T14:37:43Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2081.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2081"
}
|
Fix docstring issues.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2081/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2081/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2080
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2080/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2080/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2080/events
|
https://github.com/huggingface/datasets/issues/2080
| 835,023,000 |
MDU6SXNzdWU4MzUwMjMwMDA=
| 2,080 |
Multidimensional arrays in a Dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/3142085?v=4",
"events_url": "https://api.github.com/users/vermouthmjl/events{/privacy}",
"followers_url": "https://api.github.com/users/vermouthmjl/followers",
"following_url": "https://api.github.com/users/vermouthmjl/following{/other_user}",
"gists_url": "https://api.github.com/users/vermouthmjl/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/vermouthmjl",
"id": 3142085,
"login": "vermouthmjl",
"node_id": "MDQ6VXNlcjMxNDIwODU=",
"organizations_url": "https://api.github.com/users/vermouthmjl/orgs",
"received_events_url": "https://api.github.com/users/vermouthmjl/received_events",
"repos_url": "https://api.github.com/users/vermouthmjl/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/vermouthmjl/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/vermouthmjl/subscriptions",
"type": "User",
"url": "https://api.github.com/users/vermouthmjl"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-18T16:29:14Z
|
2021-03-25T12:46:53Z
|
2021-03-25T12:46:53Z
|
NONE
| null | null | null |
Hi,
I'm trying to put together a `datasets.Dataset` to be used with LayoutLM which is available in `transformers`. This model requires as input the bounding boxes of each of the token of a sequence. This is when I realized that `Dataset` does not support multi-dimensional arrays as a value for a column in a row.
The following code results in conversion error in pyarrow (`pyarrow.lib.ArrowInvalid: ('Can only convert 1-dimensional array values', 'Conversion failed for column bbox with type object')`)
```
from datasets import Dataset
import pandas as pd
import numpy as np
dataset = pd.DataFrame({
'bbox': [
np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),
np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),
np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]]),
np.array([[1,2,3,4],[1,2,3,4],[1,2,3,4]])
],
'input_ids': [1, 2, 3, 4]
})
dataset = Dataset.from_pandas(dataset)
```
Since I wanted to use pytorch for the downstream training task, I also tried a few ways to directly put in a column of 2-D pytorch tensor in a formatted dataset, but I can only have a list of 1-D tensors, or a list of arrays, or a list of lists.
```
import torch
from datasets import Dataset
import pandas as pd
dataset = pd.DataFrame({
'bbox': [
[[1,2,3,4],[1,2,3,4],[1,2,3,4]],
[[1,2,3,4],[1,2,3,4],[1,2,3,4]],
[[1,2,3,4],[1,2,3,4],[1,2,3,4]],
[[1,2,3,4],[1,2,3,4],[1,2,3,4]]
],
'input_ids': [1, 2, 3, 4]
})
dataset = Dataset.from_pandas(dataset)
def test(examples):
return {'bbbox': torch.Tensor(examples['bbox'])}
dataset = dataset.map(test)
print(dataset[0]['bbox'])
print(dataset[0]['bbbox'])
dataset.set_format(type='torch', columns=['input_ids', 'bbox'], output_all_columns=True)
print(dataset[0]['bbox'])
print(dataset[0]['bbbox'])
def test2(examples):
return {'bbbox': torch.stack(examples['bbox'])}
dataset = dataset.map(test2)
print(dataset[0]['bbox'])
print(dataset[0]['bbbox'])
```
Is is possible to support n-D arrays/tensors in datasets?
It seems that it can also be useful for this [feature request](https://github.com/huggingface/datasets/issues/263).
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2080/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2080/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2079
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2079/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2079/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2079/events
|
https://github.com/huggingface/datasets/pull/2079
| 834,920,493 |
MDExOlB1bGxSZXF1ZXN0NTk1NjU2MDQ5
| 2,079 |
Refactorize Metric.compute signature to force keyword arguments only
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-18T15:05:50Z
|
2021-03-23T15:31:44Z
|
2021-03-23T15:31:44Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2079.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2079",
"merged_at": "2021-03-23T15:31:44Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2079.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2079"
}
|
Minor refactoring of Metric.compute signature to force the use of keyword arguments, by using the single star syntax.
|
{
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2079/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2079/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2078
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2078/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2078/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2078/events
|
https://github.com/huggingface/datasets/issues/2078
| 834,694,819 |
MDU6SXNzdWU4MzQ2OTQ4MTk=
| 2,078 |
MemoryError when computing WER metric
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/5707233?v=4",
"events_url": "https://api.github.com/users/diego-fustes/events{/privacy}",
"followers_url": "https://api.github.com/users/diego-fustes/followers",
"following_url": "https://api.github.com/users/diego-fustes/following{/other_user}",
"gists_url": "https://api.github.com/users/diego-fustes/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/diego-fustes",
"id": 5707233,
"login": "diego-fustes",
"node_id": "MDQ6VXNlcjU3MDcyMzM=",
"organizations_url": "https://api.github.com/users/diego-fustes/orgs",
"received_events_url": "https://api.github.com/users/diego-fustes/received_events",
"repos_url": "https://api.github.com/users/diego-fustes/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/diego-fustes/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/diego-fustes/subscriptions",
"type": "User",
"url": "https://api.github.com/users/diego-fustes"
}
|
[
{
"color": "25b21e",
"default": false,
"description": "A bug in a metric script",
"id": 2067393914,
"name": "metric bug",
"node_id": "MDU6TGFiZWwyMDY3MzkzOTE0",
"url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug"
}
] |
closed
| false |
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
] | null |
[] |
2021-03-18T11:30:05Z
|
2021-05-01T08:31:49Z
|
2021-04-06T07:20:43Z
|
NONE
| null | null | null |
Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation:
```
wer = load_metric("wer")
print(wer.compute(predictions=result["predicted"], references=result["target"]))
```
However, I receive the following exception:
`Traceback (most recent call last):
File "/home/diego/IpGlobal/wav2vec/test_wav2vec.py", line 51, in <module>
print(wer.compute(predictions=result["predicted"], references=result["target"]))
File "/home/diego/miniconda3/envs/wav2vec3.6/lib/python3.6/site-packages/datasets/metric.py", line 403, in compute
output = self._compute(predictions=predictions, references=references, **kwargs)
File "/home/diego/.cache/huggingface/modules/datasets_modules/metrics/wer/73b2d32b723b7fb8f204d785c00980ae4d937f12a65466f8fdf78706e2951281/wer.py", line 94, in _compute
return wer(references, predictions)
File "/home/diego/miniconda3/envs/wav2vec3.6/lib/python3.6/site-packages/jiwer/measures.py", line 81, in wer
truth, hypothesis, truth_transform, hypothesis_transform, **kwargs
File "/home/diego/miniconda3/envs/wav2vec3.6/lib/python3.6/site-packages/jiwer/measures.py", line 192, in compute_measures
H, S, D, I = _get_operation_counts(truth, hypothesis)
File "/home/diego/miniconda3/envs/wav2vec3.6/lib/python3.6/site-packages/jiwer/measures.py", line 273, in _get_operation_counts
editops = Levenshtein.editops(source_string, destination_string)
MemoryError`
My system has more than 10GB of available RAM. Looking at the code, I think that it could be related to the way jiwer does the calculation, as it is pasting all the sentences in a single string before calling Levenshtein editops function.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2078/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2078/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2077
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2077/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2077/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2077/events
|
https://github.com/huggingface/datasets/pull/2077
| 834,649,536 |
MDExOlB1bGxSZXF1ZXN0NTk1NDI0MTYw
| 2,077 |
Bump huggingface_hub version
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4",
"events_url": "https://api.github.com/users/SBrandeis/events{/privacy}",
"followers_url": "https://api.github.com/users/SBrandeis/followers",
"following_url": "https://api.github.com/users/SBrandeis/following{/other_user}",
"gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/SBrandeis",
"id": 33657802,
"login": "SBrandeis",
"node_id": "MDQ6VXNlcjMzNjU3ODAy",
"organizations_url": "https://api.github.com/users/SBrandeis/orgs",
"received_events_url": "https://api.github.com/users/SBrandeis/received_events",
"repos_url": "https://api.github.com/users/SBrandeis/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions",
"type": "User",
"url": "https://api.github.com/users/SBrandeis"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-18T10:54:34Z
|
2021-03-18T11:33:26Z
|
2021-03-18T11:33:26Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2077.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2077",
"merged_at": "2021-03-18T11:33:26Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2077.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2077"
}
|
`0.0.2 => 0.0.6`
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2077/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2077/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2076
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2076/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2076/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2076/events
|
https://github.com/huggingface/datasets/issues/2076
| 834,445,296 |
MDU6SXNzdWU4MzQ0NDUyOTY=
| 2,076 |
Issue: Dataset download error
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/20436061?v=4",
"events_url": "https://api.github.com/users/XuhuiZhou/events{/privacy}",
"followers_url": "https://api.github.com/users/XuhuiZhou/followers",
"following_url": "https://api.github.com/users/XuhuiZhou/following{/other_user}",
"gists_url": "https://api.github.com/users/XuhuiZhou/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/XuhuiZhou",
"id": 20436061,
"login": "XuhuiZhou",
"node_id": "MDQ6VXNlcjIwNDM2MDYx",
"organizations_url": "https://api.github.com/users/XuhuiZhou/orgs",
"received_events_url": "https://api.github.com/users/XuhuiZhou/received_events",
"repos_url": "https://api.github.com/users/XuhuiZhou/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/XuhuiZhou/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/XuhuiZhou/subscriptions",
"type": "User",
"url": "https://api.github.com/users/XuhuiZhou"
}
|
[
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] |
open
| false | null |
[] | null |
[] |
2021-03-18T06:36:06Z
|
2021-03-22T11:52:31Z
| null |
NONE
| null | null | null |
The download link in `iwslt2017.py` file does not seem to work anymore.
For example, `FileNotFoundError: Couldn't find file at https://wit3.fbk.eu/archive/2017-01-trnted/texts/zh/en/zh-en.tgz`
Would be nice if we could modify it script and use the new downloadable link?
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2076/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2076/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2075
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2075/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2075/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2075/events
|
https://github.com/huggingface/datasets/issues/2075
| 834,301,246 |
MDU6SXNzdWU4MzQzMDEyNDY=
| 2,075 |
ConnectionError: Couldn't reach common_voice.py
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/6188893?v=4",
"events_url": "https://api.github.com/users/LifaSun/events{/privacy}",
"followers_url": "https://api.github.com/users/LifaSun/followers",
"following_url": "https://api.github.com/users/LifaSun/following{/other_user}",
"gists_url": "https://api.github.com/users/LifaSun/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/LifaSun",
"id": 6188893,
"login": "LifaSun",
"node_id": "MDQ6VXNlcjYxODg4OTM=",
"organizations_url": "https://api.github.com/users/LifaSun/orgs",
"received_events_url": "https://api.github.com/users/LifaSun/received_events",
"repos_url": "https://api.github.com/users/LifaSun/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/LifaSun/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/LifaSun/subscriptions",
"type": "User",
"url": "https://api.github.com/users/LifaSun"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-18T01:19:06Z
|
2021-03-20T10:29:41Z
|
2021-03-20T10:29:41Z
|
NONE
| null | null | null |
When I run:
from datasets import load_dataset, load_metric
common_voice_train = load_dataset("common_voice", "zh-CN", split="train+validation")
common_voice_test = load_dataset("common_voice", "zh-CN", split="test")
Got:
ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/master/datasets/common_voice/common_voice.py
Version:
1.4.1
Thanks! @lhoestq @LysandreJik @thomwolf
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2075/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2075/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2074
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2074/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2074/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2074/events
|
https://github.com/huggingface/datasets/pull/2074
| 834,268,463 |
MDExOlB1bGxSZXF1ZXN0NTk1MTIzMjYw
| 2,074 |
Fix size categories in YAML Tags
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4",
"events_url": "https://api.github.com/users/gchhablani/events{/privacy}",
"followers_url": "https://api.github.com/users/gchhablani/followers",
"following_url": "https://api.github.com/users/gchhablani/following{/other_user}",
"gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gchhablani",
"id": 29076344,
"login": "gchhablani",
"node_id": "MDQ6VXNlcjI5MDc2MzQ0",
"organizations_url": "https://api.github.com/users/gchhablani/orgs",
"received_events_url": "https://api.github.com/users/gchhablani/received_events",
"repos_url": "https://api.github.com/users/gchhablani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gchhablani"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-18T00:02:36Z
|
2021-03-23T17:11:10Z
|
2021-03-23T17:11:10Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2074.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2074",
"merged_at": "2021-03-23T17:11:09Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2074.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2074"
}
|
This PR fixes several `size_categories` in YAML tags and makes them consistent. Additionally, I have added a few more categories after `1M`, up to `1T`. I would like to add that to the streamlit app also.
This PR also adds a couple of infos that I found missing.
The code for generating this:
```python
for dataset in sorted(os.listdir('./datasets/')):
if '.' not in dataset and dataset not in ['c4', 'csv', 'downloads', 'cc100', 'ccaligned_multilingual', 'celeb_a', 'chr_en', 'emea', 'glue']:
infos = {}
stats = {}
st = ''
with open(f'datasets/{dataset}/README.md') as f:
d = f.read()
start_dash = d.find('---') + 3
end_dash = d[start_dash:].find('---') + 3
rest_text = d[end_dash + 3:]
try:
full_yaml = OmegaConf.create(d[start_dash:end_dash])
readme = OmegaConf.to_container(full_yaml['size_categories'], resolve=True)
except Exception as e:
print(e)
continue
try:
with open(f'datasets/{dataset}/dataset_infos.json') as f:
data = json.load(f)
except Exception as e:
print(e)
continue # Skip those without infos.
done_set = set([])
num_keys = len(data.keys())
for keys in data:
# dataset = load_dataset('opus100', f'{dirs}')
total = 0
for split in data[keys]['splits']:
total = total + data[keys]['splits'][split]['num_examples']
if total < 1000:
st += "- n<1K" + '\n'
infos[keys] = ["n<1K"]
elif total >= 1000 and total < 10000:
infos[keys] = ["1K<n<10K"]
elif total >= 10000 and total < 100000:
infos[keys] = ["10K<n<100K"]
elif total >= 100000 and total < 1000000:
infos[keys] = ["100K<n<1M"]
elif total >= 1000000 and total < 10000000:
infos[keys] = ["1M<n<10M"]
elif total >= 10000000 and total < 100000000:
infos[keys] = ["10M<n<100M"]
elif total >= 100000000 and total < 1000000000:
infos[keys] = ["100M<n<1B"]
elif total >= 1000000000 and total < 10000000000:
infos[keys] = ["1B<n<10B"]
elif total >= 10000000000 and total < 100000000000:
infos[keys] = ["10B<n<100B"]
elif total >= 100000000000 and total < 1000000000000:
infos[keys] = ["100B<n<1T"]
else:
infos[keys] = ["n>1T"]
done_set = done_set.union(infos[keys])
if (isinstance(readme, list) and list(infos.values())[0] != readme) or (isinstance(readme, dict) and readme != infos):
print('-' * 30)
print(done_set)
print(f"Changing Full YAML for {dataset}")
print(OmegaConf.to_yaml(full_yaml))
if len(done_set) == 1:
full_yaml['size_categories'] = list(done_set)
else:
full_yaml['size_categories'] = dict([(k, v) for k, v in sorted(infos.items(), key=lambda x: x[0])])
full_yaml_string = OmegaConf.to_yaml(full_yaml)
print('-' * 30)
print(full_yaml_string)
inp = input('Do you wish to continue?(Y/N)')
if inp == 'Y':
with open(f'./datasets/{dataset}/README.md', 'w') as f:
f.write('---\n')
f.write(full_yaml_string)
f.write('---')
f.write(rest_text)
else:
break
```
Note that the lower-bound is inclusive. I'm unsure if this is how it is done in the tagging app.
EDIT:
It would be great if there was a way to make the task categories consistent too. For this, the streamlit app can look into all the datasets and check for existing categories and show them in the list. This may add some consistency.
EDIT:
I understand this will not work for cases where only the infos for some of the configs are present, for example: `ccaligned_multingual` has only 5 out of several configs present, and infos has only information about them. Hence, I have skipped a few datasets in the code, if there are more such datasets, then I'll ignore them too.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2074/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2074/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2073
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2073/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2073/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2073/events
|
https://github.com/huggingface/datasets/pull/2073
| 834,192,501 |
MDExOlB1bGxSZXF1ZXN0NTk1MDYyMzQ2
| 2,073 |
Fixes check of TF_AVAILABLE and TORCH_AVAILABLE
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/32632186?v=4",
"events_url": "https://api.github.com/users/philschmid/events{/privacy}",
"followers_url": "https://api.github.com/users/philschmid/followers",
"following_url": "https://api.github.com/users/philschmid/following{/other_user}",
"gists_url": "https://api.github.com/users/philschmid/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/philschmid",
"id": 32632186,
"login": "philschmid",
"node_id": "MDQ6VXNlcjMyNjMyMTg2",
"organizations_url": "https://api.github.com/users/philschmid/orgs",
"received_events_url": "https://api.github.com/users/philschmid/received_events",
"repos_url": "https://api.github.com/users/philschmid/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/philschmid/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/philschmid/subscriptions",
"type": "User",
"url": "https://api.github.com/users/philschmid"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-17T21:28:53Z
|
2021-03-18T09:09:25Z
|
2021-03-18T09:09:24Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2073.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2073",
"merged_at": "2021-03-18T09:09:24Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2073.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2073"
}
|
# What is this PR doing
This PR implements the checks if `Tensorflow` and `Pytorch` are available the same way as `transformers` does it. I added the additional checks for the different `Tensorflow` and `torch` versions. #2068
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2073/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2073/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2072
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2072/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2072/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2072/events
|
https://github.com/huggingface/datasets/pull/2072
| 834,054,837 |
MDExOlB1bGxSZXF1ZXN0NTk0OTQ5NjA4
| 2,072 |
Fix docstring issues
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[
{
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation",
"id": 1935892861,
"name": "documentation",
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-17T18:13:44Z
|
2021-03-24T08:20:57Z
|
2021-03-18T12:41:21Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2072.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2072",
"merged_at": "2021-03-18T12:41:21Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2072.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2072"
}
|
Fix docstring issues.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2072/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2072/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2071
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2071/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2071/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2071/events
|
https://github.com/huggingface/datasets/issues/2071
| 833,950,824 |
MDU6SXNzdWU4MzM5NTA4MjQ=
| 2,071 |
Multiprocessing is slower than single process
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
|
[
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-17T16:08:58Z
|
2021-03-18T09:10:23Z
|
2021-03-18T09:10:23Z
|
CONTRIBUTOR
| null | null | null |
```python
# benchmark_filter.py
import logging
import sys
import time
from datasets import load_dataset, set_caching_enabled
if __name__ == "__main__":
set_caching_enabled(False)
logging.basicConfig(level=logging.DEBUG)
bc = load_dataset("bookcorpus")
now = time.time()
try:
bc["train"].filter(lambda x: len(x["text"]) < 64, num_proc=int(sys.argv[1]))
except Exception as e:
print(f"cancelled: {e}")
elapsed = time.time() - now
print(elapsed)
```
Running `python benchmark_filter.py 1` (20min+) is faster than `python benchmark_filter.py 2` (2hrs+)
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2071/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2071/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2070
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2070/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2070/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2070/events
|
https://github.com/huggingface/datasets/issues/2070
| 833,799,035 |
MDU6SXNzdWU4MzM3OTkwMzU=
| 2,070 |
ArrowInvalid issue for squad v2 dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29818977?v=4",
"events_url": "https://api.github.com/users/MichaelYxWang/events{/privacy}",
"followers_url": "https://api.github.com/users/MichaelYxWang/followers",
"following_url": "https://api.github.com/users/MichaelYxWang/following{/other_user}",
"gists_url": "https://api.github.com/users/MichaelYxWang/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/MichaelYxWang",
"id": 29818977,
"login": "MichaelYxWang",
"node_id": "MDQ6VXNlcjI5ODE4OTc3",
"organizations_url": "https://api.github.com/users/MichaelYxWang/orgs",
"received_events_url": "https://api.github.com/users/MichaelYxWang/received_events",
"repos_url": "https://api.github.com/users/MichaelYxWang/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/MichaelYxWang/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/MichaelYxWang/subscriptions",
"type": "User",
"url": "https://api.github.com/users/MichaelYxWang"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-17T13:51:49Z
|
2021-08-04T17:57:16Z
|
2021-08-04T17:57:16Z
|
NONE
| null | null | null |
Hello, I am using the huggingface official question answering example notebook (https://colab.research.google.com/github/huggingface/notebooks/blob/master/examples/question_answering.ipynb).
In the prepare_validation_features function, I made some modifications to tokenize a new set of quesions with the original contexts and save them in three different list called candidate_input_dis, candidate_attetion_mask and candidate_token_type_ids. When I try to run the next cell for dataset.map, I got the following error:
`ArrowInvalid: Column 1 named candidate_attention_mask expected length 1180 but got length 1178`
My code is as follows:
```
def generate_candidate_questions(examples):
val_questions = examples["question"]
candididate_questions = random.sample(datasets["train"]["question"], len(val_questions))
candididate_questions = [x[:max_length] for x in candididate_questions]
return candididate_questions
def prepare_validation_features(examples, use_mixing=False):
pad_on_right = tokenizer.padding_side == "right"
tokenized_examples = tokenizer(
examples["question" if pad_on_right else "context"],
examples["context" if pad_on_right else "question"],
truncation="only_second" if pad_on_right else "only_first",
max_length=max_length,
stride=doc_stride,
return_overflowing_tokens=True,
return_offsets_mapping=True,
padding="max_length",
)
if use_mixing:
candidate_questions = generate_candidate_questions(examples)
tokenized_candidates = tokenizer(
candidate_questions if pad_on_right else examples["context"],
examples["context"] if pad_on_right else candidate_questions,
truncation="only_second" if pad_on_right else "only_first",
max_length=max_length,
stride=doc_stride,
return_overflowing_tokens=True,
return_offsets_mapping=True,
padding="max_length",
)
sample_mapping = tokenized_examples.pop("overflow_to_sample_mapping")
tokenized_examples["example_id"] = []
if use_mixing:
tokenized_examples["candidate_input_ids"] = tokenized_candidates["input_ids"]
tokenized_examples["candidate_attention_mask"] = tokenized_candidates["attention_mask"]
tokenized_examples["candidate_token_type_ids"] = tokenized_candidates["token_type_ids"]
for i in range(len(tokenized_examples["input_ids"])):
sequence_ids = tokenized_examples.sequence_ids(i)
context_index = 1 if pad_on_right else 0
sample_index = sample_mapping[i]
tokenized_examples["example_id"].append(examples["id"][sample_index])
tokenized_examples["offset_mapping"][i] = [
(o if sequence_ids[k] == context_index else None)
for k, o in enumerate(tokenized_examples["offset_mapping"][i])
]
return tokenized_examples
validation_features = datasets["validation"].map(
lambda xs: prepare_validation_features(xs, True),
batched=True,
remove_columns=datasets["validation"].column_names
)
```
I guess this might happen because of the batched=True. I see similar issues in this repo related to arrow table length mismatch error, but in their cases, the numbers vary a lot. In my case, this error always happens when the expected length and unexpected length are very close. Thanks for the help!
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2070/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2070/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2069
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2069/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2069/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2069/events
|
https://github.com/huggingface/datasets/pull/2069
| 833,768,926 |
MDExOlB1bGxSZXF1ZXN0NTk0NzA5ODYw
| 2,069 |
Add and fix docstring for NamedSplit
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-17T13:19:28Z
|
2021-03-18T10:27:40Z
|
2021-03-18T10:27:40Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2069.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2069",
"merged_at": "2021-03-18T10:27:40Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2069.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2069"
}
|
Add and fix docstring for `NamedSplit`, which was missing.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2069/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2069/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2068
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2068/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2068/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2068/events
|
https://github.com/huggingface/datasets/issues/2068
| 833,602,832 |
MDU6SXNzdWU4MzM2MDI4MzI=
| 2,068 |
PyTorch not available error on SageMaker GPU docker though it is installed
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/1651457?v=4",
"events_url": "https://api.github.com/users/sivakhno/events{/privacy}",
"followers_url": "https://api.github.com/users/sivakhno/followers",
"following_url": "https://api.github.com/users/sivakhno/following{/other_user}",
"gists_url": "https://api.github.com/users/sivakhno/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sivakhno",
"id": 1651457,
"login": "sivakhno",
"node_id": "MDQ6VXNlcjE2NTE0NTc=",
"organizations_url": "https://api.github.com/users/sivakhno/orgs",
"received_events_url": "https://api.github.com/users/sivakhno/received_events",
"repos_url": "https://api.github.com/users/sivakhno/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sivakhno/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sivakhno/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sivakhno"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-17T10:04:27Z
|
2021-06-14T04:47:30Z
|
2021-06-14T04:47:30Z
|
NONE
| null | null | null |
I get en error when running data loading using SageMaker SDK
```
File "main.py", line 34, in <module>
run_training()
File "main.py", line 25, in run_training
dm.setup('fit')
File "/opt/conda/lib/python3.6/site-packages/pytorch_lightning/core/datamodule.py", line 92, in wrapped_fn
return fn(*args, **kwargs)
File "/opt/ml/code/data_module.py", line 103, in setup
self.dataset[split].set_format(type="torch", columns=self.columns)
File "/opt/conda/lib/python3.6/site-packages/datasets/fingerprint.py", line 337, in wrapper
out = func(self, *args, **kwargs)
File "/opt/conda/lib/python3.6/site-packages/datasets/arrow_dataset.py", line 995, in set_format
_ = get_formatter(type, **format_kwargs)
File "/opt/conda/lib/python3.6/site-packages/datasets/formatting/__init__.py", line 114, in get_formatter
raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type]
ValueError: PyTorch needs to be installed to be able to return PyTorch tensors.
```
when trying to execute dataset loading using this notebook https://github.com/PyTorchLightning/pytorch-lightning/blob/master/notebooks/04-transformers-text-classification.ipynb, specifically lines
```
self.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns]
self.dataset[split].set_format(type="torch", columns=self.columns)
```
The SageMaker docker image used is 763104351884.dkr.ecr.eu-central-1.amazonaws.com/pytorch-training:1.4.0-gpu-py3 .
By running container interactively I have checked that torch loading completes successfully by executing `https://github.com/huggingface/datasets/blob/master/src/datasets/config.py#L39`.
Also as a first line in the data loading module I have
```
import os
os.environ["USE_TF"] = "0"
os.environ["USE_TORCH"] = "1"
````
But unfortunately the error stills persists. Any suggestions would be appreciated as I am stack.
Many Thanks!
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2068/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2068/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2067
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2067/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2067/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2067/events
|
https://github.com/huggingface/datasets/issues/2067
| 833,559,940 |
MDU6SXNzdWU4MzM1NTk5NDA=
| 2,067 |
Multiprocessing windows error
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47894090?v=4",
"events_url": "https://api.github.com/users/flozi00/events{/privacy}",
"followers_url": "https://api.github.com/users/flozi00/followers",
"following_url": "https://api.github.com/users/flozi00/following{/other_user}",
"gists_url": "https://api.github.com/users/flozi00/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/flozi00",
"id": 47894090,
"login": "flozi00",
"node_id": "MDQ6VXNlcjQ3ODk0MDkw",
"organizations_url": "https://api.github.com/users/flozi00/orgs",
"received_events_url": "https://api.github.com/users/flozi00/received_events",
"repos_url": "https://api.github.com/users/flozi00/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/flozi00/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/flozi00/subscriptions",
"type": "User",
"url": "https://api.github.com/users/flozi00"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-17T09:12:28Z
|
2021-08-04T17:59:08Z
|
2021-08-04T17:59:08Z
|
CONTRIBUTOR
| null | null | null |
As described here https://huggingface.co/blog/fine-tune-xlsr-wav2vec2
When using the num_proc argument on windows the whole Python environment crashes and hanging in loop.
For example at the map_to_array part.
An error occures because the cache file already exists and windows throws and error. After this the log crashes into an loop
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2067/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2067/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2066
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2066/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2066/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2066/events
|
https://github.com/huggingface/datasets/pull/2066
| 833,480,551 |
MDExOlB1bGxSZXF1ZXN0NTk0NDcwMjEz
| 2,066 |
Fix docstring rendering of Dataset/DatasetDict.from_csv args
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-17T07:23:10Z
|
2021-03-17T09:21:21Z
|
2021-03-17T09:21:21Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2066.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2066",
"merged_at": "2021-03-17T09:21:21Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2066.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2066"
}
|
Fix the docstring rendering of Dataset/DatasetDict.from_csv args.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2066/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2066/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2065
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2065/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2065/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2065/events
|
https://github.com/huggingface/datasets/issues/2065
| 833,291,432 |
MDU6SXNzdWU4MzMyOTE0MzI=
| 2,065 |
Only user permission of saved cache files, not group
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/57237365?v=4",
"events_url": "https://api.github.com/users/lorr1/events{/privacy}",
"followers_url": "https://api.github.com/users/lorr1/followers",
"following_url": "https://api.github.com/users/lorr1/following{/other_user}",
"gists_url": "https://api.github.com/users/lorr1/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lorr1",
"id": 57237365,
"login": "lorr1",
"node_id": "MDQ6VXNlcjU3MjM3MzY1",
"organizations_url": "https://api.github.com/users/lorr1/orgs",
"received_events_url": "https://api.github.com/users/lorr1/received_events",
"repos_url": "https://api.github.com/users/lorr1/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lorr1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lorr1/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lorr1"
}
|
[
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "7057ff",
"default": true,
"description": "Good for newcomers",
"id": 1935892877,
"name": "good first issue",
"node_id": "MDU6TGFiZWwxOTM1ODkyODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-17T00:20:22Z
|
2022-06-28T08:10:10Z
|
2021-05-10T06:45:29Z
|
NONE
| null | null | null |
Hello,
It seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you know any ways around this or a way to correctly set the permissions?
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2065/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2065/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2064
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2064/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2064/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2064/events
|
https://github.com/huggingface/datasets/pull/2064
| 833,002,360 |
MDExOlB1bGxSZXF1ZXN0NTk0MDczOTQ1
| 2,064 |
Fix ted_talks_iwslt version error
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-16T16:43:45Z
|
2021-03-16T18:00:08Z
|
2021-03-16T18:00:08Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2064.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2064",
"merged_at": "2021-03-16T18:00:07Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2064.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2064"
}
|
This PR fixes the bug where the version argument would be passed twice if the dataset configuration was created on the fly.
Fixes #2059
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2064/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2064/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2063
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2063/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2063/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2063/events
|
https://github.com/huggingface/datasets/pull/2063
| 832,993,705 |
MDExOlB1bGxSZXF1ZXN0NTk0MDY2NzI5
| 2,063 |
[Common Voice] Adapt dataset script so that no manual data download is actually needed
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}",
"followers_url": "https://api.github.com/users/patrickvonplaten/followers",
"following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}",
"gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/patrickvonplaten",
"id": 23423619,
"login": "patrickvonplaten",
"node_id": "MDQ6VXNlcjIzNDIzNjE5",
"organizations_url": "https://api.github.com/users/patrickvonplaten/orgs",
"received_events_url": "https://api.github.com/users/patrickvonplaten/received_events",
"repos_url": "https://api.github.com/users/patrickvonplaten/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions",
"type": "User",
"url": "https://api.github.com/users/patrickvonplaten"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-16T16:33:44Z
|
2021-03-17T09:42:52Z
|
2021-03-17T09:42:37Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2063.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2063",
"merged_at": "2021-03-17T09:42:37Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2063.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2063"
}
|
This PR changes the dataset script so that no manual data dir is needed anymore.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2063/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2063/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2062
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2062/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2062/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2062/events
|
https://github.com/huggingface/datasets/pull/2062
| 832,625,483 |
MDExOlB1bGxSZXF1ZXN0NTkzNzUyNTMz
| 2,062 |
docs: fix missing quotation
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/46561493?v=4",
"events_url": "https://api.github.com/users/neal2018/events{/privacy}",
"followers_url": "https://api.github.com/users/neal2018/followers",
"following_url": "https://api.github.com/users/neal2018/following{/other_user}",
"gists_url": "https://api.github.com/users/neal2018/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/neal2018",
"id": 46561493,
"login": "neal2018",
"node_id": "MDQ6VXNlcjQ2NTYxNDkz",
"organizations_url": "https://api.github.com/users/neal2018/orgs",
"received_events_url": "https://api.github.com/users/neal2018/received_events",
"repos_url": "https://api.github.com/users/neal2018/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/neal2018/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/neal2018/subscriptions",
"type": "User",
"url": "https://api.github.com/users/neal2018"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-16T10:07:54Z
|
2021-03-17T09:21:57Z
|
2021-03-17T09:21:57Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2062.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2062",
"merged_at": "2021-03-17T09:21:56Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2062.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2062"
}
|
The json code misses a quote
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2062/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2062/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2061
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2061/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2061/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2061/events
|
https://github.com/huggingface/datasets/issues/2061
| 832,596,228 |
MDU6SXNzdWU4MzI1OTYyMjg=
| 2,061 |
Cannot load udpos subsets from xtreme dataset using load_dataset()
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/55791365?v=4",
"events_url": "https://api.github.com/users/adzcodez/events{/privacy}",
"followers_url": "https://api.github.com/users/adzcodez/followers",
"following_url": "https://api.github.com/users/adzcodez/following{/other_user}",
"gists_url": "https://api.github.com/users/adzcodez/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/adzcodez",
"id": 55791365,
"login": "adzcodez",
"node_id": "MDQ6VXNlcjU1NzkxMzY1",
"organizations_url": "https://api.github.com/users/adzcodez/orgs",
"received_events_url": "https://api.github.com/users/adzcodez/received_events",
"repos_url": "https://api.github.com/users/adzcodez/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/adzcodez/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/adzcodez/subscriptions",
"type": "User",
"url": "https://api.github.com/users/adzcodez"
}
|
[
{
"color": "7057ff",
"default": true,
"description": "Good for newcomers",
"id": 1935892877,
"name": "good first issue",
"node_id": "MDU6TGFiZWwxOTM1ODkyODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-16T09:32:13Z
|
2021-06-18T11:54:11Z
|
2021-06-18T11:54:10Z
|
NONE
| null | null | null |
Hello,
I am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and faced the same error.
Reprex is:
`from datasets import load_dataset `
`dataset = load_dataset('xtreme', 'udpos.English')`
The error is:
`KeyError: '_'`
The full traceback is:
KeyError Traceback (most recent call last)
<ipython-input-5-7181359ea09d> in <module>
1 from datasets import load_dataset
----> 2 dataset = load_dataset('xtreme', 'udpos.English')
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)
738
739 # Download and prepare data
--> 740 builder_instance.download_and_prepare(
741 download_config=download_config,
742 download_mode=download_mode,
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs)
576 logger.warning("HF google storage unreachable. Downloading and preparing it from source")
577 if not downloaded_from_gcs:
--> 578 self._download_and_prepare(
579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
580 )
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)
654 try:
655 # Prepare split will record examples associated to the split
--> 656 self._prepare_split(split_generator, **prepare_split_kwargs)
657 except OSError as e:
658 raise OSError(
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\builder.py in _prepare_split(self, split_generator)
977 generator, unit=" examples", total=split_info.num_examples, leave=False, disable=not_verbose
978 ):
--> 979 example = self.info.features.encode_example(record)
980 writer.write(example)
981 finally:
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\features.py in encode_example(self, example)
946 def encode_example(self, example):
947 example = cast_to_python_objects(example)
--> 948 return encode_nested_example(self, example)
949
950 def encode_batch(self, batch):
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\features.py in encode_nested_example(schema, obj)
840 # Nested structures: we allow dict, list/tuples, sequences
841 if isinstance(schema, dict):
--> 842 return {
843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)
844 }
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\features.py in <dictcomp>(.0)
841 if isinstance(schema, dict):
842 return {
--> 843 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj)
844 }
845 elif isinstance(schema, (list, tuple)):
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\features.py in encode_nested_example(schema, obj)
868 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks
869 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)):
--> 870 return schema.encode_example(obj)
871 # Other object should be directly convertible to a native Arrow type (like Translation and Translation)
872 return obj
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\features.py in encode_example(self, example_data)
647 # If a string is given, convert to associated integer
648 if isinstance(example_data, str):
--> 649 example_data = self.str2int(example_data)
650
651 # Allowing -1 to mean no label.
~\Anaconda3\envs\mlenv\lib\site-packages\datasets\features.py in str2int(self, values)
605 if value not in self._str2int:
606 value = value.strip()
--> 607 output.append(self._str2int[str(value)])
608 else:
609 # No names provided, try to integerize
KeyError: '_'
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2061/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2061/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2060
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2060/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2060/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2060/events
|
https://github.com/huggingface/datasets/pull/2060
| 832,588,591 |
MDExOlB1bGxSZXF1ZXN0NTkzNzIxNzcx
| 2,060 |
Filtering refactor
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
|
[] |
closed
| false |
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
|
[
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
] | null |
[] |
2021-03-16T09:23:30Z
|
2021-10-13T09:09:04Z
|
2021-10-13T09:09:03Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2060.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2060",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2060.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2060"
}
|
fix https://github.com/huggingface/datasets/issues/2032
benchmarking is somewhat inconclusive, currently running on `book_corpus` with:
```python
bc = load_dataset("bookcorpus")
now = time.time()
bc.filter(lambda x: len(x["text"]) < 64)
elapsed = time.time() - now
print(elapsed)
```
this branch does it in 233 seconds, master in 1409 seconds.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2060/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2060/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2059
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2059/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2059/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2059/events
|
https://github.com/huggingface/datasets/issues/2059
| 832,579,156 |
MDU6SXNzdWU4MzI1NzkxNTY=
| 2,059 |
Error while following docs to load the `ted_talks_iwslt` dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/40426312?v=4",
"events_url": "https://api.github.com/users/ekdnam/events{/privacy}",
"followers_url": "https://api.github.com/users/ekdnam/followers",
"following_url": "https://api.github.com/users/ekdnam/following{/other_user}",
"gists_url": "https://api.github.com/users/ekdnam/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ekdnam",
"id": 40426312,
"login": "ekdnam",
"node_id": "MDQ6VXNlcjQwNDI2MzEy",
"organizations_url": "https://api.github.com/users/ekdnam/orgs",
"received_events_url": "https://api.github.com/users/ekdnam/received_events",
"repos_url": "https://api.github.com/users/ekdnam/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ekdnam/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ekdnam/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ekdnam"
}
|
[
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-16T09:12:19Z
|
2021-03-16T18:00:31Z
|
2021-03-16T18:00:07Z
|
NONE
| null | null | null |
I am currently trying to load the `ted_talks_iwslt` dataset into google colab.
The [docs](https://huggingface.co/datasets/ted_talks_iwslt) mention the following way of doing so.
```python
dataset = load_dataset("ted_talks_iwslt", language_pair=("it", "pl"), year="2014")
```
Executing it results in the error attached below.
```
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
<ipython-input-6-7dcc67154ef9> in <module>()
----> 1 dataset = load_dataset("ted_talks_iwslt", language_pair=("it", "pl"), year="2014")
4 frames
/usr/local/lib/python3.7/dist-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs)
730 hash=hash,
731 features=features,
--> 732 **config_kwargs,
733 )
734
/usr/local/lib/python3.7/dist-packages/datasets/builder.py in __init__(self, writer_batch_size, *args, **kwargs)
927
928 def __init__(self, *args, writer_batch_size=None, **kwargs):
--> 929 super(GeneratorBasedBuilder, self).__init__(*args, **kwargs)
930 # Batch size used by the ArrowWriter
931 # It defines the number of samples that are kept in memory before writing them
/usr/local/lib/python3.7/dist-packages/datasets/builder.py in __init__(self, cache_dir, name, hash, features, **config_kwargs)
241 name,
242 custom_features=features,
--> 243 **config_kwargs,
244 )
245
/usr/local/lib/python3.7/dist-packages/datasets/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs)
337 if "version" not in config_kwargs and hasattr(self, "VERSION") and self.VERSION:
338 config_kwargs["version"] = self.VERSION
--> 339 builder_config = self.BUILDER_CONFIG_CLASS(**config_kwargs)
340
341 # otherwise use the config_kwargs to overwrite the attributes
/root/.cache/huggingface/modules/datasets_modules/datasets/ted_talks_iwslt/024d06b1376b361e59245c5878ab8acf9a7576d765f2d0077f61751158e60914/ted_talks_iwslt.py in __init__(self, language_pair, year, **kwargs)
219 description=description,
220 version=datasets.Version("1.1.0", ""),
--> 221 **kwargs,
222 )
223
TypeError: __init__() got multiple values for keyword argument 'version'
```
How to resolve this?
PS: Thanks a lot @huggingface team for creating this great library!
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2059/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2059/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2058
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2058/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2058/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2058/events
|
https://github.com/huggingface/datasets/issues/2058
| 832,159,844 |
MDU6SXNzdWU4MzIxNTk4NDQ=
| 2,058 |
Is it possible to convert a `tfds` to HuggingFace `dataset`?
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/6608232?v=4",
"events_url": "https://api.github.com/users/abarbosa94/events{/privacy}",
"followers_url": "https://api.github.com/users/abarbosa94/followers",
"following_url": "https://api.github.com/users/abarbosa94/following{/other_user}",
"gists_url": "https://api.github.com/users/abarbosa94/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/abarbosa94",
"id": 6608232,
"login": "abarbosa94",
"node_id": "MDQ6VXNlcjY2MDgyMzI=",
"organizations_url": "https://api.github.com/users/abarbosa94/orgs",
"received_events_url": "https://api.github.com/users/abarbosa94/received_events",
"repos_url": "https://api.github.com/users/abarbosa94/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/abarbosa94/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/abarbosa94/subscriptions",
"type": "User",
"url": "https://api.github.com/users/abarbosa94"
}
|
[] |
open
| false | null |
[] | null |
[] |
2021-03-15T20:18:47Z
|
2022-10-05T16:55:18Z
| null |
CONTRIBUTOR
| null | null | null |
I was having some weird bugs with `C4`dataset version of HuggingFace, so I decided to try to download `C4`from `tfds`. I would like to know if it is possible to convert a tfds dataset to HuggingFace dataset format :)
I can also open a new issue reporting the bug I'm receiving with `datasets.load_dataset('c4','en')` in the future if you think that it would be useful.
Thanks!
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2058/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2058/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2057
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2057/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2057/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2057/events
|
https://github.com/huggingface/datasets/pull/2057
| 832,120,522 |
MDExOlB1bGxSZXF1ZXN0NTkzMzMzMjM0
| 2,057 |
update link to ZEST dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/619844?v=4",
"events_url": "https://api.github.com/users/matt-peters/events{/privacy}",
"followers_url": "https://api.github.com/users/matt-peters/followers",
"following_url": "https://api.github.com/users/matt-peters/following{/other_user}",
"gists_url": "https://api.github.com/users/matt-peters/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/matt-peters",
"id": 619844,
"login": "matt-peters",
"node_id": "MDQ6VXNlcjYxOTg0NA==",
"organizations_url": "https://api.github.com/users/matt-peters/orgs",
"received_events_url": "https://api.github.com/users/matt-peters/received_events",
"repos_url": "https://api.github.com/users/matt-peters/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/matt-peters/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/matt-peters/subscriptions",
"type": "User",
"url": "https://api.github.com/users/matt-peters"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-15T19:22:57Z
|
2021-03-16T17:06:28Z
|
2021-03-16T17:06:28Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2057.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2057",
"merged_at": "2021-03-16T17:06:28Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2057.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2057"
}
|
Updating the link as the original one is no longer working.
|
{
"+1": 3,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 3,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2057/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2057/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2056
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2056/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2056/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2056/events
|
https://github.com/huggingface/datasets/issues/2056
| 831,718,397 |
MDU6SXNzdWU4MzE3MTgzOTc=
| 2,056 |
issue with opus100/en-fr dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4",
"events_url": "https://api.github.com/users/dorost1234/events{/privacy}",
"followers_url": "https://api.github.com/users/dorost1234/followers",
"following_url": "https://api.github.com/users/dorost1234/following{/other_user}",
"gists_url": "https://api.github.com/users/dorost1234/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/dorost1234",
"id": 79165106,
"login": "dorost1234",
"node_id": "MDQ6VXNlcjc5MTY1MTA2",
"organizations_url": "https://api.github.com/users/dorost1234/orgs",
"received_events_url": "https://api.github.com/users/dorost1234/received_events",
"repos_url": "https://api.github.com/users/dorost1234/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/dorost1234/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dorost1234/subscriptions",
"type": "User",
"url": "https://api.github.com/users/dorost1234"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-15T11:32:42Z
|
2021-03-16T15:49:00Z
|
2021-03-16T15:48:59Z
|
NONE
| null | null | null |
Hi
I am running run_mlm.py code of huggingface repo with opus100/fr-en pair, I am getting this error, note that this error occurs for only this pairs and not the other pairs. Any idea why this is occurring? and how I can solve this?
Thanks a lot @lhoestq for your help in advance.
`
thread '<unnamed>' panicked at 'index out of bounds: the len is 617 but the index is 617', /__w/tokenizers/tokenizers/tokenizers/src/tokenizer/normalizer.rs:382:21
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
63%|██████████████████████████████████████████████████████████▊ | 626/1000 [00:27<00:16, 22.69ba/s]
Traceback (most recent call last):
File "run_mlm.py", line 550, in <module>
main()
File "run_mlm.py", line 412, in main
in zip(data_args.dataset_name, data_args.dataset_config_name)]
File "run_mlm.py", line 411, in <listcomp>
logger) for dataset_name, dataset_config_name\
File "/user/dara/dev/codes/seq2seq/data/tokenize_datasets.py", line 96, in get_tokenized_dataset
load_from_cache_file=not data_args.overwrite_cache,
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/dataset_dict.py", line 448, in map
for k, dataset in self.items()
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/dataset_dict.py", line 448, in <dictcomp>
for k, dataset in self.items()
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1309, in map
update_data=update_data,
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 204, in wrapper
out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs)
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/fingerprint.py", line 337, in wrapper
out = func(self, *args, **kwargs)
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1574, in _map_single
batch, indices, check_same_num_examples=len(self.list_indexes()) > 0, offset=offset
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1490, in apply_function_on_filtered_inputs
function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)
File "/user/dara/dev/codes/seq2seq/data/tokenize_datasets.py", line 89, in tokenize_function
return tokenizer(examples[text_column_name], return_special_tokens_mask=True)
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/transformers/tokenization_utils_base.py", line 2347, in __call__
**kwargs,
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/transformers/tokenization_utils_base.py", line 2532, in batch_encode_plus
**kwargs,
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/transformers/tokenization_utils_fast.py", line 384, in _batch_encode_plus
is_pretokenized=is_split_into_words,
pyo3_runtime.PanicException: index out of bounds: the len is 617 but the index is 617
`
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2056/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2056/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2055
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2055/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2055/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2055/events
|
https://github.com/huggingface/datasets/issues/2055
| 831,684,312 |
MDU6SXNzdWU4MzE2ODQzMTI=
| 2,055 |
is there a way to override a dataset object saved with save_to_disk?
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4",
"events_url": "https://api.github.com/users/shamanez/events{/privacy}",
"followers_url": "https://api.github.com/users/shamanez/followers",
"following_url": "https://api.github.com/users/shamanez/following{/other_user}",
"gists_url": "https://api.github.com/users/shamanez/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/shamanez",
"id": 16892570,
"login": "shamanez",
"node_id": "MDQ6VXNlcjE2ODkyNTcw",
"organizations_url": "https://api.github.com/users/shamanez/orgs",
"received_events_url": "https://api.github.com/users/shamanez/received_events",
"repos_url": "https://api.github.com/users/shamanez/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/shamanez/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shamanez/subscriptions",
"type": "User",
"url": "https://api.github.com/users/shamanez"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-15T10:50:53Z
|
2021-03-22T04:06:17Z
|
2021-03-22T04:06:17Z
|
NONE
| null | null | null |
At the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object?
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2055/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2055/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2054
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2054/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2054/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2054/events
|
https://github.com/huggingface/datasets/issues/2054
| 831,597,665 |
MDU6SXNzdWU4MzE1OTc2NjU=
| 2,054 |
Could not find file for ZEST dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/26653468?v=4",
"events_url": "https://api.github.com/users/bhadreshpsavani/events{/privacy}",
"followers_url": "https://api.github.com/users/bhadreshpsavani/followers",
"following_url": "https://api.github.com/users/bhadreshpsavani/following{/other_user}",
"gists_url": "https://api.github.com/users/bhadreshpsavani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/bhadreshpsavani",
"id": 26653468,
"login": "bhadreshpsavani",
"node_id": "MDQ6VXNlcjI2NjUzNDY4",
"organizations_url": "https://api.github.com/users/bhadreshpsavani/orgs",
"received_events_url": "https://api.github.com/users/bhadreshpsavani/received_events",
"repos_url": "https://api.github.com/users/bhadreshpsavani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/bhadreshpsavani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bhadreshpsavani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/bhadreshpsavani"
}
|
[
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-15T09:11:58Z
|
2021-05-03T09:30:24Z
|
2021-05-03T09:30:24Z
|
CONTRIBUTOR
| null | null | null |
I am trying to use zest dataset from Allen AI using below code in colab,
```
!pip install -q datasets
from datasets import load_dataset
dataset = load_dataset("zest")
```
I am getting the following error,
```
Using custom data configuration default
Downloading and preparing dataset zest/default (download: 5.53 MiB, generated: 19.96 MiB, post-processed: Unknown size, total: 25.48 MiB) to /root/.cache/huggingface/datasets/zest/default/0.0.0/1f7a230fbfc964d979bbca0f0130fbab3259fce547ee758ad8aa4f9c9bec6cca...
---------------------------------------------------------------------------
FileNotFoundError Traceback (most recent call last)
<ipython-input-6-18dbbc1a4b8a> in <module>()
1 from datasets import load_dataset
2
----> 3 dataset = load_dataset("zest")
9 frames
/usr/local/lib/python3.7/dist-packages/datasets/utils/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token)
612 )
613 elif response is not None and response.status_code == 404:
--> 614 raise FileNotFoundError("Couldn't find file at {}".format(url))
615 _raise_if_offline_mode_is_enabled(f"Tried to reach {url}")
616 raise ConnectionError("Couldn't reach {}".format(url))
FileNotFoundError: Couldn't find file at https://ai2-datasets.s3-us-west-2.amazonaws.com/zest/zest.zip
```
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2054/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2054/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2053
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2053/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2053/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2053/events
|
https://github.com/huggingface/datasets/pull/2053
| 831,151,728 |
MDExOlB1bGxSZXF1ZXN0NTkyNTM4ODY2
| 2,053 |
Add bAbI QA tasks
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4",
"events_url": "https://api.github.com/users/gchhablani/events{/privacy}",
"followers_url": "https://api.github.com/users/gchhablani/followers",
"following_url": "https://api.github.com/users/gchhablani/following{/other_user}",
"gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gchhablani",
"id": 29076344,
"login": "gchhablani",
"node_id": "MDQ6VXNlcjI5MDc2MzQ0",
"organizations_url": "https://api.github.com/users/gchhablani/orgs",
"received_events_url": "https://api.github.com/users/gchhablani/received_events",
"repos_url": "https://api.github.com/users/gchhablani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gchhablani"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-14T13:04:39Z
|
2021-03-29T12:41:48Z
|
2021-03-29T12:41:48Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2053.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2053",
"merged_at": "2021-03-29T12:41:48Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2053.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2053"
}
|
- **Name:** *The (20) QA bAbI tasks*
- **Description:** *The (20) QA bAbI tasks are a set of proxy tasks that evaluate reading comprehension via question answering. Our tasks measure understanding in several ways: whether a system is able to answer questions via chaining facts, simple induction, deduction and many more. The tasks are designed to be prerequisites for any system that aims to be capable of conversing with a human. The aim is to classify these tasks into skill sets,so that researchers can identify (and then rectify) the failings of their systems.*
- **Paper:** [arXiv](https://arxiv.org/pdf/1502.05698.pdf)
- **Data:** [Facebook Research Page](https://research.fb.com/downloads/babi/)
- **Motivation:** This is a unique dataset with story-based Question Answering. It is a part of the `bAbI` project by Facebook Research.
**Note**: I have currently added all the 160 configs. If this seems impractical, I can keep only a few. While each `dummy_data.zip` weighs a few KBs, overall it is around 1.3MB for all configurations. This is problematic. Let me know what is to be done.
Thanks :)
### Checkbox
- [x] Create the dataset script `/datasets/my_dataset/my_dataset.py` using the template
- [x] Fill the `_DESCRIPTION` and `_CITATION` variables
- [x] Implement `_infos()`, `_split_generators()` and `_generate_examples()`
- [x] Make sure that the `BUILDER_CONFIGS` class attribute is filled with the different configurations of the dataset and that the `BUILDER_CONFIG_CLASS` is specified if there is a custom config class.
- [x] Generate the metadata file `dataset_infos.json` for all configurations
- [x] Generate the dummy data `dummy_data.zip` files to have the dataset script tested and that they don't weigh too much (<50KB)
- [x] Add the dataset card `README.md` using the template : fill the tags and the various paragraphs
- [x] Both tests for the real data and the dummy data pass.
|
{
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2053/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2053/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2052
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2052/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2052/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2052/events
|
https://github.com/huggingface/datasets/issues/2052
| 831,135,704 |
MDU6SXNzdWU4MzExMzU3MDQ=
| 2,052 |
Timit_asr dataset repeats examples
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/7583522?v=4",
"events_url": "https://api.github.com/users/fermaat/events{/privacy}",
"followers_url": "https://api.github.com/users/fermaat/followers",
"following_url": "https://api.github.com/users/fermaat/following{/other_user}",
"gists_url": "https://api.github.com/users/fermaat/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/fermaat",
"id": 7583522,
"login": "fermaat",
"node_id": "MDQ6VXNlcjc1ODM1MjI=",
"organizations_url": "https://api.github.com/users/fermaat/orgs",
"received_events_url": "https://api.github.com/users/fermaat/received_events",
"repos_url": "https://api.github.com/users/fermaat/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/fermaat/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/fermaat/subscriptions",
"type": "User",
"url": "https://api.github.com/users/fermaat"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-14T11:43:43Z
|
2021-03-15T10:37:16Z
|
2021-03-15T10:37:16Z
|
NONE
| null | null | null |
Summary
When loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same
Steps to reproduce
As an example, on this code there is the text from the training part:
Code snippet:
```
from datasets import load_dataset, load_metric
timit = load_dataset("timit_asr")
timit['train']['text']
#['Would such an act of refusal be useful?',
# 'Would such an act of refusal be useful?',
# 'Would such an act of refusal be useful?',
# 'Would such an act of refusal be useful?',
# 'Would such an act of refusal be useful?',
# 'Would such an act of refusal be useful?',
```
The same behavior happens for other columns
Expected behavior:
Different info on the actual timit_asr dataset
Actual behavior:
When loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same. I've checked datasets 1.3 and the rows are different
Debug info
Streamlit version: (get it with $ streamlit version)
Python version: Python 3.6.12
Using Conda? PipEnv? PyEnv? Pex? Using pip
OS version: Centos-release-7-9.2009.1.el7.centos.x86_64
Additional information
You can check the same behavior on https://huggingface.co/datasets/viewer/?dataset=timit_asr
|
{
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2052/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2052/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2051
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2051/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2051/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2051/events
|
https://github.com/huggingface/datasets/pull/2051
| 831,027,021 |
MDExOlB1bGxSZXF1ZXN0NTkyNDQ2MDU1
| 2,051 |
Add MDD Dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4",
"events_url": "https://api.github.com/users/gchhablani/events{/privacy}",
"followers_url": "https://api.github.com/users/gchhablani/followers",
"following_url": "https://api.github.com/users/gchhablani/following{/other_user}",
"gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gchhablani",
"id": 29076344,
"login": "gchhablani",
"node_id": "MDQ6VXNlcjI5MDc2MzQ0",
"organizations_url": "https://api.github.com/users/gchhablani/orgs",
"received_events_url": "https://api.github.com/users/gchhablani/received_events",
"repos_url": "https://api.github.com/users/gchhablani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gchhablani"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-14T00:01:05Z
|
2021-03-19T11:15:44Z
|
2021-03-19T10:31:59Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2051.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2051",
"merged_at": "2021-03-19T10:31:59Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2051.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2051"
}
|
- **Name:** *MDD Dataset*
- **Description:** The Movie Dialog dataset (MDD) is designed to measure how well models can perform at goal and non-goal orientated dialog centered around the topic of movies (question answering, recommendation and discussion), from various movie reviews sources such as MovieLens and OMDb.
- **Paper:** [arXiv](https://arxiv.org/pdf/1511.06931.pdf)
- **Data:** https://research.fb.com/downloads/babi/
- **Motivation:** This is one of the popular dialog datasets, a part of Facebook Research's "bAbI project".
### Checkbox
- [x] Create the dataset script `/datasets/my_dataset/my_dataset.py` using the template
- [x] Fill the `_DESCRIPTION` and `_CITATION` variables
- [x] Implement `_infos()`, `_split_generators()` and `_generate_examples()`
- [x] Make sure that the `BUILDER_CONFIGS` class attribute is filled with the different configurations of the dataset and that the `BUILDER_CONFIG_CLASS` is specified if there is a custom config class.
- [x] Generate the metadata file `dataset_infos.json` for all configurations
- [x] Generate the dummy data `dummy_data.zip` files to have the dataset script tested and that they don't weigh too much (<50KB)
- [x] Add the dataset card `README.md` using the template : fill the tags and the various paragraphs
- [x] Both tests for the real data and the dummy data pass.
**Note**: I haven't included the following from the data files: `entities` (the file containing list of all entities in the first three subtasks), `dictionary`(the dictionary of words they use in their models), `movie_kb`(contains the knowledge base of information about the movies, actors and other entities that are mentioned in the dialogs). Please let me know if those are needed, and if yes, should I make separate configurations for them?
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2051/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2051/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2050
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2050/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2050/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2050/events
|
https://github.com/huggingface/datasets/issues/2050
| 831,006,551 |
MDU6SXNzdWU4MzEwMDY1NTE=
| 2,050 |
Build custom dataset to fine-tune Wav2Vec2
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/72882909?v=4",
"events_url": "https://api.github.com/users/Omarnabk/events{/privacy}",
"followers_url": "https://api.github.com/users/Omarnabk/followers",
"following_url": "https://api.github.com/users/Omarnabk/following{/other_user}",
"gists_url": "https://api.github.com/users/Omarnabk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Omarnabk",
"id": 72882909,
"login": "Omarnabk",
"node_id": "MDQ6VXNlcjcyODgyOTA5",
"organizations_url": "https://api.github.com/users/Omarnabk/orgs",
"received_events_url": "https://api.github.com/users/Omarnabk/received_events",
"repos_url": "https://api.github.com/users/Omarnabk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Omarnabk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Omarnabk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Omarnabk"
}
|
[
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-13T22:01:10Z
|
2021-03-15T09:27:28Z
|
2021-03-15T09:27:28Z
|
NONE
| null | null | null |
Thank you for your recent tutorial on how to finetune Wav2Vec2 on a custom dataset. The example you gave here (https://huggingface.co/blog/fine-tune-xlsr-wav2vec2) was on the CommonVoice dataset. However, what if I want to load my own dataset? I have a manifest (transcript and their audio files) in a JSON file.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2050/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2050/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2049
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2049/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2049/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2049/events
|
https://github.com/huggingface/datasets/pull/2049
| 830,978,687 |
MDExOlB1bGxSZXF1ZXN0NTkyNDE2MzQ0
| 2,049 |
Fix text-classification tags
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4",
"events_url": "https://api.github.com/users/gchhablani/events{/privacy}",
"followers_url": "https://api.github.com/users/gchhablani/followers",
"following_url": "https://api.github.com/users/gchhablani/following{/other_user}",
"gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gchhablani",
"id": 29076344,
"login": "gchhablani",
"node_id": "MDQ6VXNlcjI5MDc2MzQ0",
"organizations_url": "https://api.github.com/users/gchhablani/orgs",
"received_events_url": "https://api.github.com/users/gchhablani/received_events",
"repos_url": "https://api.github.com/users/gchhablani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gchhablani"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-13T19:51:42Z
|
2021-03-16T15:47:46Z
|
2021-03-16T15:47:46Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2049.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2049",
"merged_at": "2021-03-16T15:47:46Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2049.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2049"
}
|
There are different tags for text classification right now: `text-classification` and `text_classification`:
.
This PR fixes it.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2049/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2049/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2048
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2048/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2048/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2048/events
|
https://github.com/huggingface/datasets/issues/2048
| 830,953,431 |
MDU6SXNzdWU4MzA5NTM0MzE=
| 2,048 |
github is not always available - probably need a back up
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4",
"events_url": "https://api.github.com/users/stas00/events{/privacy}",
"followers_url": "https://api.github.com/users/stas00/followers",
"following_url": "https://api.github.com/users/stas00/following{/other_user}",
"gists_url": "https://api.github.com/users/stas00/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/stas00",
"id": 10676103,
"login": "stas00",
"node_id": "MDQ6VXNlcjEwNjc2MTAz",
"organizations_url": "https://api.github.com/users/stas00/orgs",
"received_events_url": "https://api.github.com/users/stas00/received_events",
"repos_url": "https://api.github.com/users/stas00/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/stas00/subscriptions",
"type": "User",
"url": "https://api.github.com/users/stas00"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-13T18:03:32Z
|
2022-04-01T15:27:10Z
|
2022-04-01T15:27:10Z
|
MEMBER
| null | null | null |
Yesterday morning github wasn't working:
```
:/tmp$ wget https://raw.githubusercontent.com/huggingface/datasets/1.4.1/metrics/sacrebleu/sacrebleu.py--2021-03-12 18:35:59-- https://raw.githubusercontent.com/huggingface/datasets/1.4.1/metrics/sacrebleu/sacrebleu.py
Resolving raw.githubusercontent.com (raw.githubusercontent.com)... 185.199.108.133, 185.199.111.133, 185.199.109.133, ...
Connecting to raw.githubusercontent.com (raw.githubusercontent.com)|185.199.108.133|:443... connected.
HTTP request sent, awaiting response... 500 Internal Server Error
2021-03-12 18:36:11 ERROR 500: Internal Server Error.
```
Suggestion: have a failover system and replicate the data on another system and reach there if gh isn't reachable? perhaps gh can be a master and the replicate a slave - so there is only one true source.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2048/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2048/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2047
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2047/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2047/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2047/events
|
https://github.com/huggingface/datasets/pull/2047
| 830,626,430 |
MDExOlB1bGxSZXF1ZXN0NTkyMTI2NzQ3
| 2,047 |
Multilingual dIalogAct benchMark (miam)
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/1551356?v=4",
"events_url": "https://api.github.com/users/eusip/events{/privacy}",
"followers_url": "https://api.github.com/users/eusip/followers",
"following_url": "https://api.github.com/users/eusip/following{/other_user}",
"gists_url": "https://api.github.com/users/eusip/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/eusip",
"id": 1551356,
"login": "eusip",
"node_id": "MDQ6VXNlcjE1NTEzNTY=",
"organizations_url": "https://api.github.com/users/eusip/orgs",
"received_events_url": "https://api.github.com/users/eusip/received_events",
"repos_url": "https://api.github.com/users/eusip/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/eusip/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/eusip/subscriptions",
"type": "User",
"url": "https://api.github.com/users/eusip"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T23:02:55Z
|
2021-03-23T10:36:34Z
|
2021-03-19T10:47:13Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2047.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2047",
"merged_at": "2021-03-19T10:47:13Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2047.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2047"
}
|
My collaborators (@EmileChapuis, @PierreColombo) and I within the Affective Computing team at Telecom Paris would like to anonymously publish the miam dataset. It is assocated with a publication currently under review. We will update the dataset with full citations once the review period is over.
|
{
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2047/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2047/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2046
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2046/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2046/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2046/events
|
https://github.com/huggingface/datasets/issues/2046
| 830,423,033 |
MDU6SXNzdWU4MzA0MjMwMzM=
| 2,046 |
add_faisis_index gets very slow when doing it interatively
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4",
"events_url": "https://api.github.com/users/shamanez/events{/privacy}",
"followers_url": "https://api.github.com/users/shamanez/followers",
"following_url": "https://api.github.com/users/shamanez/following{/other_user}",
"gists_url": "https://api.github.com/users/shamanez/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/shamanez",
"id": 16892570,
"login": "shamanez",
"node_id": "MDQ6VXNlcjE2ODkyNTcw",
"organizations_url": "https://api.github.com/users/shamanez/orgs",
"received_events_url": "https://api.github.com/users/shamanez/received_events",
"repos_url": "https://api.github.com/users/shamanez/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/shamanez/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shamanez/subscriptions",
"type": "User",
"url": "https://api.github.com/users/shamanez"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T20:27:18Z
|
2021-03-24T22:29:11Z
|
2021-03-24T22:29:11Z
|
NONE
| null | null | null |
As the below code suggests, I want to run add_faisis_index in every nth interaction from the training loop. I have 7.2 million documents. Usually, it takes 2.5 hours (if I run an as a separate process similar to the script given in rag/use_own_knowleldge_dataset.py). Now, this takes usually 5hrs. Is this normal? Any way to make this process faster?
@lhoestq
```
def training_step(self, batch, batch_idx) -> Dict:
if (not batch_idx==0) and (batch_idx%5==0):
print("******************************************************")
ctx_encoder=self.trainer.model.module.module.model.rag.ctx_encoder
model_copy =type(ctx_encoder)(self.config_dpr) # get a new instance #this will be load in the CPU
model_copy.load_state_dict(ctx_encoder.state_dict()) # copy weights and stuff
list_of_gpus = ['cuda:2','cuda:3']
c_dir='/custom/cache/dir'
kb_dataset = load_dataset("csv", data_files=[self.custom_config.csv_path], split="train", delimiter="\t", column_names=["title", "text"],cache_dir=c_dir)
print(kb_dataset)
n=len(list_of_gpus) #nunber of dedicated GPUs
kb_list=[kb_dataset.shard(n, i, contiguous=True) for i in range(n)]
#kb_dataset.save_to_disk('/hpc/gsir059/MY-Test/RAY/transformers/examples/research_projects/rag/haha-dir')
print(self.trainer.global_rank)
dataset_shards = self.re_encode_kb(model_copy.to(device=list_of_gpus[self.trainer.global_rank]),kb_list[self.trainer.global_rank])
output = [None for _ in list_of_gpus]
#self.trainer.accelerator_connector.accelerator.barrier("embedding_process")
dist.all_gather_object(output, dataset_shards)
#This creation and re-initlaization of the new index
if (self.trainer.global_rank==0): #saving will be done in the main process
combined_dataset = concatenate_datasets(output)
passages_path =self.config.passages_path
logger.info("saving the dataset with ")
#combined_dataset.save_to_disk('/hpc/gsir059/MY-Test/RAY/transformers/examples/research_projects/rag/MY-Passage')
combined_dataset.save_to_disk(passages_path)
logger.info("Add faiss index to the dataset that consist of embeddings")
embedding_dataset=combined_dataset
index = faiss.IndexHNSWFlat(768, 128, faiss.METRIC_INNER_PRODUCT)
embedding_dataset.add_faiss_index("embeddings", custom_index=index)
embedding_dataset.get_index("embeddings").save(self.config.index_path)
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2046/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2046/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2045
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2045/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2045/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2045/events
|
https://github.com/huggingface/datasets/pull/2045
| 830,351,527 |
MDExOlB1bGxSZXF1ZXN0NTkxODc2Mjcz
| 2,045 |
Preserve column ordering in Dataset.rename_column
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T18:26:47Z
|
2021-03-16T14:48:05Z
|
2021-03-16T14:35:05Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2045.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2045",
"merged_at": "2021-03-16T14:35:05Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2045.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2045"
}
|
Currently `Dataset.rename_column` doesn't necessarily preserve the order of the columns:
```python
>>> from datasets import Dataset
>>> d = Dataset.from_dict({'sentences': ["s1", "s2"], 'label': [0, 1]})
>>> d
Dataset({
features: ['sentences', 'label'],
num_rows: 2
})
>>> d.rename_column('sentences', 'text')
Dataset({
features: ['label', 'text'],
num_rows: 2
})
```
This PR fixes this.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2045/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2045/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2044
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2044/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2044/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2044/events
|
https://github.com/huggingface/datasets/pull/2044
| 830,339,905 |
MDExOlB1bGxSZXF1ZXN0NTkxODY2NzM1
| 2,044 |
Add CBT dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4",
"events_url": "https://api.github.com/users/gchhablani/events{/privacy}",
"followers_url": "https://api.github.com/users/gchhablani/followers",
"following_url": "https://api.github.com/users/gchhablani/following{/other_user}",
"gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gchhablani",
"id": 29076344,
"login": "gchhablani",
"node_id": "MDQ6VXNlcjI5MDc2MzQ0",
"organizations_url": "https://api.github.com/users/gchhablani/orgs",
"received_events_url": "https://api.github.com/users/gchhablani/received_events",
"repos_url": "https://api.github.com/users/gchhablani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gchhablani"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T18:04:19Z
|
2021-03-19T11:10:13Z
|
2021-03-19T10:29:15Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2044.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2044",
"merged_at": "2021-03-19T10:29:15Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2044.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2044"
}
|
This PR adds the [CBT Dataset](https://arxiv.org/abs/1511.02301).
Note that I have also added the `raw` dataset as a separate configuration. I couldn't find a suitable "task" for it in YAML tags.
The dummy files have one example each, as the examples are slightly big. For `raw` dataset, I just used top few lines, because they are entire books and would take up a lot of space.
Let me know in case of any issues.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2044/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2044/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2043
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2043/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2043/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2043/events
|
https://github.com/huggingface/datasets/pull/2043
| 830,279,098 |
MDExOlB1bGxSZXF1ZXN0NTkxODE1ODAz
| 2,043 |
Support pickle protocol for dataset splits defined as ReadInstruction
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T16:35:11Z
|
2021-03-16T14:25:38Z
|
2021-03-16T14:05:05Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2043.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2043",
"merged_at": "2021-03-16T14:05:05Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2043.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2043"
}
|
Fixes #2022 (+ some style fixes)
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2043/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2043/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2042
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2042/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2042/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2042/events
|
https://github.com/huggingface/datasets/pull/2042
| 830,190,276 |
MDExOlB1bGxSZXF1ZXN0NTkxNzQwNzQ3
| 2,042 |
Fix arrow memory checks issue in tests
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T14:49:52Z
|
2021-03-12T15:04:23Z
|
2021-03-12T15:04:22Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2042.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2042",
"merged_at": "2021-03-12T15:04:22Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2042.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2042"
}
|
The tests currently fail on `master` because the arrow memory verification doesn't return the expected memory evolution when loading an arrow table in memory.
From my experiments, the tests fail only when the full test suite is ran.
This made me think that maybe some arrow objects from other tests were not freeing their memory until they do and cause the memory verifications to fail in other tests.
Collecting the garbage collector before checking the arrow memory usage seems to fix this issue.
I added a context manager `assert_arrow_memory_increases` that we can use in tests and that deals with the gc.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2042/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2042/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2041
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2041/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2041/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2041/events
|
https://github.com/huggingface/datasets/pull/2041
| 830,180,803 |
MDExOlB1bGxSZXF1ZXN0NTkxNzMyNzMw
| 2,041 |
Doc2dial update data_infos and data_loaders
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/2062185?v=4",
"events_url": "https://api.github.com/users/songfeng/events{/privacy}",
"followers_url": "https://api.github.com/users/songfeng/followers",
"following_url": "https://api.github.com/users/songfeng/following{/other_user}",
"gists_url": "https://api.github.com/users/songfeng/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/songfeng",
"id": 2062185,
"login": "songfeng",
"node_id": "MDQ6VXNlcjIwNjIxODU=",
"organizations_url": "https://api.github.com/users/songfeng/orgs",
"received_events_url": "https://api.github.com/users/songfeng/received_events",
"repos_url": "https://api.github.com/users/songfeng/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/songfeng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/songfeng/subscriptions",
"type": "User",
"url": "https://api.github.com/users/songfeng"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T14:39:29Z
|
2021-03-16T11:09:20Z
|
2021-03-16T11:09:20Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2041.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2041",
"merged_at": "2021-03-16T11:09:20Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2041.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2041"
}
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2041/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2041/timeline
| null | null | true |
|
https://api.github.com/repos/huggingface/datasets/issues/2040
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2040/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2040/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2040/events
|
https://github.com/huggingface/datasets/issues/2040
| 830,169,387 |
MDU6SXNzdWU4MzAxNjkzODc=
| 2,040 |
ValueError: datasets' indices [1] come from memory and datasets' indices [0] come from disk
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/53626067?v=4",
"events_url": "https://api.github.com/users/simonschoe/events{/privacy}",
"followers_url": "https://api.github.com/users/simonschoe/followers",
"following_url": "https://api.github.com/users/simonschoe/following{/other_user}",
"gists_url": "https://api.github.com/users/simonschoe/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/simonschoe",
"id": 53626067,
"login": "simonschoe",
"node_id": "MDQ6VXNlcjUzNjI2MDY3",
"organizations_url": "https://api.github.com/users/simonschoe/orgs",
"received_events_url": "https://api.github.com/users/simonschoe/received_events",
"repos_url": "https://api.github.com/users/simonschoe/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/simonschoe/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/simonschoe/subscriptions",
"type": "User",
"url": "https://api.github.com/users/simonschoe"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T14:27:00Z
|
2021-08-04T18:00:43Z
|
2021-08-04T18:00:43Z
|
NONE
| null | null | null |
Hi there,
I am trying to concat two datasets that I've previously saved to disk via `save_to_disk()` like so (note that both are saved as `DataDict`, `PATH_DATA_CLS_*` are `Path`-objects):
```python
concatenate_datasets([load_from_disk(PATH_DATA_CLS_A)['train'], load_from_disk(PATH_DATA_CLS_B)['train']])
```
Yielding the following error:
```python
ValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.
However datasets' indices [1] come from memory and datasets' indices [0] come from disk.
```
Been trying to solve this for quite some time now. Both `DataDict` have been created by reading in a `csv` via `load_dataset` and subsequently processed using the various `datasets` methods (i.e. filter, map, remove col, rename col). Can't figure out tho...
`load_from_disk(PATH_DATA_CLS_A)['train']` yields:
```python
Dataset({
features: ['labels', 'text'],
num_rows: 785
})
```
`load_from_disk(PATH_DATA_CLS_B)['train']` yields:
```python
Dataset({
features: ['labels', 'text'],
num_rows: 3341
})
```
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2040/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2040/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2039
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2039/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2039/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2039/events
|
https://github.com/huggingface/datasets/pull/2039
| 830,047,652 |
MDExOlB1bGxSZXF1ZXN0NTkxNjE3ODY3
| 2,039 |
Doc2dial rc
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/2062185?v=4",
"events_url": "https://api.github.com/users/songfeng/events{/privacy}",
"followers_url": "https://api.github.com/users/songfeng/followers",
"following_url": "https://api.github.com/users/songfeng/following{/other_user}",
"gists_url": "https://api.github.com/users/songfeng/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/songfeng",
"id": 2062185,
"login": "songfeng",
"node_id": "MDQ6VXNlcjIwNjIxODU=",
"organizations_url": "https://api.github.com/users/songfeng/orgs",
"received_events_url": "https://api.github.com/users/songfeng/received_events",
"repos_url": "https://api.github.com/users/songfeng/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/songfeng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/songfeng/subscriptions",
"type": "User",
"url": "https://api.github.com/users/songfeng"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T11:56:28Z
|
2021-03-12T15:32:36Z
|
2021-03-12T15:32:36Z
|
CONTRIBUTOR
| null | true |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2039.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2039",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2039.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2039"
}
|
Added fix to handle the last turn that is a user turn.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2039/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2039/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2038
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2038/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2038/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2038/events
|
https://github.com/huggingface/datasets/issues/2038
| 830,036,875 |
MDU6SXNzdWU4MzAwMzY4NzU=
| 2,038 |
outdated dataset_infos.json might fail verifications
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/2062185?v=4",
"events_url": "https://api.github.com/users/songfeng/events{/privacy}",
"followers_url": "https://api.github.com/users/songfeng/followers",
"following_url": "https://api.github.com/users/songfeng/following{/other_user}",
"gists_url": "https://api.github.com/users/songfeng/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/songfeng",
"id": 2062185,
"login": "songfeng",
"node_id": "MDQ6VXNlcjIwNjIxODU=",
"organizations_url": "https://api.github.com/users/songfeng/orgs",
"received_events_url": "https://api.github.com/users/songfeng/received_events",
"repos_url": "https://api.github.com/users/songfeng/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/songfeng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/songfeng/subscriptions",
"type": "User",
"url": "https://api.github.com/users/songfeng"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T11:41:54Z
|
2021-03-16T16:27:40Z
|
2021-03-16T16:27:40Z
|
CONTRIBUTOR
| null | null | null |
The [doc2dial/dataset_infos.json](https://github.com/huggingface/datasets/blob/master/datasets/doc2dial/dataset_infos.json) is outdated. It would fail data_loader when verifying download checksum etc..
Could you please update this file or point me how to update this file?
Thank you.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2038/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2038/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2037
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2037/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2037/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2037/events
|
https://github.com/huggingface/datasets/pull/2037
| 829,919,685 |
MDExOlB1bGxSZXF1ZXN0NTkxNTA4MTQz
| 2,037 |
Fix: Wikipedia - save memory by replacing root.clear with elem.clear
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/6331508?v=4",
"events_url": "https://api.github.com/users/miyamonz/events{/privacy}",
"followers_url": "https://api.github.com/users/miyamonz/followers",
"following_url": "https://api.github.com/users/miyamonz/following{/other_user}",
"gists_url": "https://api.github.com/users/miyamonz/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/miyamonz",
"id": 6331508,
"login": "miyamonz",
"node_id": "MDQ6VXNlcjYzMzE1MDg=",
"organizations_url": "https://api.github.com/users/miyamonz/orgs",
"received_events_url": "https://api.github.com/users/miyamonz/received_events",
"repos_url": "https://api.github.com/users/miyamonz/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/miyamonz/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/miyamonz/subscriptions",
"type": "User",
"url": "https://api.github.com/users/miyamonz"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T09:22:00Z
|
2021-03-23T06:08:16Z
|
2021-03-16T11:01:22Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2037.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2037",
"merged_at": "2021-03-16T11:01:22Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2037.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2037"
}
|
see: https://github.com/huggingface/datasets/issues/2031
What I did:
- replace root.clear with elem.clear
- remove lines to get root element
- $ make style
- $ make test
- some tests required some pip packages, I installed them.
test results on origin/master and my branch are same. I think it's not related on my modification, isn't it?
```
==================================================================================== short test summary info ====================================================================================
FAILED tests/test_arrow_writer.py::TypedSequenceTest::test_catch_overflow - AssertionError: OverflowError not raised
============================================================= 1 failed, 2332 passed, 5138 skipped, 70 warnings in 91.75s (0:01:31) ==============================================================
make: *** [Makefile:19: test] Error 1
```
Is there anything else I should do?
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2037/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2037/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2036
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2036/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2036/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2036/events
|
https://github.com/huggingface/datasets/issues/2036
| 829,909,258 |
MDU6SXNzdWU4Mjk5MDkyNTg=
| 2,036 |
Cannot load wikitext
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/19349207?v=4",
"events_url": "https://api.github.com/users/Gpwner/events{/privacy}",
"followers_url": "https://api.github.com/users/Gpwner/followers",
"following_url": "https://api.github.com/users/Gpwner/following{/other_user}",
"gists_url": "https://api.github.com/users/Gpwner/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Gpwner",
"id": 19349207,
"login": "Gpwner",
"node_id": "MDQ6VXNlcjE5MzQ5MjA3",
"organizations_url": "https://api.github.com/users/Gpwner/orgs",
"received_events_url": "https://api.github.com/users/Gpwner/received_events",
"repos_url": "https://api.github.com/users/Gpwner/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Gpwner/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Gpwner/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Gpwner"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-12T09:09:39Z
|
2021-03-15T08:45:02Z
|
2021-03-15T08:44:44Z
|
NONE
| null | null | null |
when I execute these codes
```
>>> from datasets import load_dataset
>>> test_dataset = load_dataset("wikitext")
```
I got an error,any help?
```
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/xxx/anaconda3/envs/transformer/lib/python3.7/site-packages/datasets/load.py", line 589, in load_dataset
path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True
File "/home/xxx/anaconda3/envs/transformer/lib/python3.7/site-packages/datasets/load.py", line 267, in prepare_module
local_path = cached_path(file_path, download_config=download_config)
File "/home/xxx/anaconda3/envs/transformer/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 308, in cached_path
use_etag=download_config.use_etag,
File "/home/xxx/anaconda3/envs/transformer/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 487, in get_from_cache
raise ConnectionError("Couldn't reach {}".format(url))
ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/wikitext/wikitext.py
```
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2036/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2036/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2035
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2035/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2035/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2035/events
|
https://github.com/huggingface/datasets/issues/2035
| 829,475,544 |
MDU6SXNzdWU4Mjk0NzU1NDQ=
| 2,035 |
wiki40b/wikipedia for almost all languages cannot be downloaded
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4",
"events_url": "https://api.github.com/users/dorost1234/events{/privacy}",
"followers_url": "https://api.github.com/users/dorost1234/followers",
"following_url": "https://api.github.com/users/dorost1234/following{/other_user}",
"gists_url": "https://api.github.com/users/dorost1234/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/dorost1234",
"id": 79165106,
"login": "dorost1234",
"node_id": "MDQ6VXNlcjc5MTY1MTA2",
"organizations_url": "https://api.github.com/users/dorost1234/orgs",
"received_events_url": "https://api.github.com/users/dorost1234/received_events",
"repos_url": "https://api.github.com/users/dorost1234/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/dorost1234/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dorost1234/subscriptions",
"type": "User",
"url": "https://api.github.com/users/dorost1234"
}
|
[] |
open
| false | null |
[] | null |
[] |
2021-03-11T19:54:54Z
|
2021-03-16T14:53:37Z
| null |
NONE
| null | null | null |
Hi
I am trying to download the data as below:
```
from datasets import load_dataset
dataset = load_dataset("wiki40b", "cs")
print(dataset)
```
I am getting this error. @lhoestq I will be grateful if you could assist me with this error. For almost all languages except english I am getting this error.
I really need majority of languages in this dataset to be able to train my models for a deadline and your great scalable super well-written library is my only hope to train the models at scale while being low on resources.
thank you very much.
```
(fast) dara@vgne046:/user/dara/dev/codes/seq2seq$ python test_data.py
Downloading and preparing dataset wiki40b/cs (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to temp/dara/cache_home_2/datasets/wiki40b/cs/1.1.0/063778187363ffb294896eaa010fc254b42b73e31117c71573a953b0b0bf010f...
Traceback (most recent call last):
File "test_data.py", line 3, in <module>
dataset = load_dataset("wiki40b", "cs")
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/load.py", line 746, in load_dataset
use_auth_token=use_auth_token,
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/builder.py", line 579, in download_and_prepare
dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/datasets/builder.py", line 1105, in _download_and_prepare
import apache_beam as beam
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/apache_beam-2.28.0-py3.7-linux-x86_64.egg/apache_beam/__init__.py", line 96, in <module>
from apache_beam import io
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/apache_beam-2.28.0-py3.7-linux-x86_64.egg/apache_beam/io/__init__.py", line 23, in <module>
from apache_beam.io.avroio import *
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/apache_beam-2.28.0-py3.7-linux-x86_64.egg/apache_beam/io/avroio.py", line 55, in <module>
import avro
File "<frozen importlib._bootstrap>", line 983, in _find_and_load
File "<frozen importlib._bootstrap>", line 967, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 668, in _load_unlocked
File "<frozen importlib._bootstrap>", line 638, in _load_backward_compatible
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/avro_python3-1.9.2.1-py3.7.egg/avro/__init__.py", line 34, in <module>
File "/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/avro_python3-1.9.2.1-py3.7.egg/avro/__init__.py", line 30, in LoadResource
NotADirectoryError: [Errno 20] Not a directory: '/user/dara/libs/anaconda3/envs/fast/lib/python3.7/site-packages/avro_python3-1.9.2.1-py3.7.egg/avro/VERSION.txt'
```
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2035/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2035/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2034
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2034/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2034/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2034/events
|
https://github.com/huggingface/datasets/pull/2034
| 829,381,388 |
MDExOlB1bGxSZXF1ZXN0NTkxMDU2MTEw
| 2,034 |
Fix typo
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/3413464?v=4",
"events_url": "https://api.github.com/users/pcyin/events{/privacy}",
"followers_url": "https://api.github.com/users/pcyin/followers",
"following_url": "https://api.github.com/users/pcyin/following{/other_user}",
"gists_url": "https://api.github.com/users/pcyin/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/pcyin",
"id": 3413464,
"login": "pcyin",
"node_id": "MDQ6VXNlcjM0MTM0NjQ=",
"organizations_url": "https://api.github.com/users/pcyin/orgs",
"received_events_url": "https://api.github.com/users/pcyin/received_events",
"repos_url": "https://api.github.com/users/pcyin/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/pcyin/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pcyin/subscriptions",
"type": "User",
"url": "https://api.github.com/users/pcyin"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-11T17:46:13Z
|
2021-03-11T18:06:25Z
|
2021-03-11T18:06:25Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2034.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2034",
"merged_at": "2021-03-11T18:06:25Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2034.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2034"
}
|
Change `ENV_XDG_CACHE_HOME ` to `XDG_CACHE_HOME `
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2034/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2034/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2033
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2033/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2033/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2033/events
|
https://github.com/huggingface/datasets/pull/2033
| 829,295,339 |
MDExOlB1bGxSZXF1ZXN0NTkwOTgzMDAy
| 2,033 |
Raise an error for outdated sacrebleu versions
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-11T16:08:00Z
|
2021-03-11T17:58:12Z
|
2021-03-11T17:58:12Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2033.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2033",
"merged_at": "2021-03-11T17:58:12Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2033.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2033"
}
|
The `sacrebleu` metric seem to only work for sacrecleu>=1.4.12
For example using sacrebleu==1.2.10, an error is raised (from metric/sacrebleu/sacrebleu.py):
```python
def _compute(
self,
predictions,
references,
smooth_method="exp",
smooth_value=None,
force=False,
lowercase=False,
tokenize=scb.DEFAULT_TOKENIZER,
use_effective_order=False,
):
references_per_prediction = len(references[0])
if any(len(refs) != references_per_prediction for refs in references):
raise ValueError("Sacrebleu requires the same number of references for each prediction")
transformed_references = [[refs[i] for refs in references] for i in range(references_per_prediction)]
> output = scb.corpus_bleu(
sys_stream=predictions,
ref_streams=transformed_references,
smooth_method=smooth_method,
smooth_value=smooth_value,
force=force,
lowercase=lowercase,
tokenize=tokenize,
use_effective_order=use_effective_order,
)
E TypeError: corpus_bleu() got an unexpected keyword argument 'smooth_method'
/mnt/cache/modules/datasets_modules/metrics/sacrebleu/b390045b3d1dd4abf6a95c4a2a11ee3bcc2b7620b076204d0ddc353fa649fd86/sacrebleu.py:114: TypeError
```
I improved the error message when users have an outdated version of sacrebleu.
The new error message tells the user to update sacrebleu.
cc @LysandreJik
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2033/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2033/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2032
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2032/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2032/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2032/events
|
https://github.com/huggingface/datasets/issues/2032
| 829,250,912 |
MDU6SXNzdWU4MjkyNTA5MTI=
| 2,032 |
Use Arrow filtering instead of writing a new arrow file for Dataset.filter
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
}
|
[
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] |
open
| false |
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
|
[
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
] | null |
[] |
2021-03-11T15:18:50Z
|
2021-03-11T17:20:57Z
| null |
MEMBER
| null | null | null |
Currently the filter method reads the dataset batch by batch to write a new, filtered, arrow file on disk. Therefore all the reading + writing can take some time.
Using a mask directly on the arrow table doesn't do any read or write operation therefore it's significantly quicker.
I think there are two cases:
- if the dataset doesn't have an indices mapping, then one can simply use the arrow filtering on the main arrow table `dataset._data.filter(...)`
- if the dataset an indices mapping, then the mask should be applied on the indices mapping table `dataset._indices.filter(...)`
The indices mapping is used to map between the idx at `dataset[idx]` in `__getitem__` and the idx in the actual arrow table.
The new filter method should therefore be faster, and allow users to pass either a filtering function (that returns a boolean given an example), or directly a mask.
Feel free to discuss this idea in this thread :)
One additional note: the refactor at #2025 would make all the pickle-related stuff work directly with the arrow filtering, so that we only need to change the Dataset.filter method without having to deal with pickle.
cc @theo-m @gchhablani
related issues: #1796 #1949
|
{
"+1": 4,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 4,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2032/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2032/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2031
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2031/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2031/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2031/events
|
https://github.com/huggingface/datasets/issues/2031
| 829,122,778 |
MDU6SXNzdWU4MjkxMjI3Nzg=
| 2,031 |
wikipedia.py generator that extracts XML doesn't release memory
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/6331508?v=4",
"events_url": "https://api.github.com/users/miyamonz/events{/privacy}",
"followers_url": "https://api.github.com/users/miyamonz/followers",
"following_url": "https://api.github.com/users/miyamonz/following{/other_user}",
"gists_url": "https://api.github.com/users/miyamonz/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/miyamonz",
"id": 6331508,
"login": "miyamonz",
"node_id": "MDQ6VXNlcjYzMzE1MDg=",
"organizations_url": "https://api.github.com/users/miyamonz/orgs",
"received_events_url": "https://api.github.com/users/miyamonz/received_events",
"repos_url": "https://api.github.com/users/miyamonz/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/miyamonz/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/miyamonz/subscriptions",
"type": "User",
"url": "https://api.github.com/users/miyamonz"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-11T12:51:24Z
|
2021-03-22T08:33:52Z
|
2021-03-22T08:33:52Z
|
CONTRIBUTOR
| null | null | null |
I tried downloading Japanese wikipedia, but it always failed because of out of memory maybe.
I found that the generator function that extracts XML data in wikipedia.py doesn't release memory in the loop.
https://github.com/huggingface/datasets/blob/13a5b7db992ad5cf77895e4c0f76595314390418/datasets/wikipedia/wikipedia.py#L464-L502
`root.clear()` intend to clear memory, but it doesn't.
https://github.com/huggingface/datasets/blob/13a5b7db992ad5cf77895e4c0f76595314390418/datasets/wikipedia/wikipedia.py#L490
https://github.com/huggingface/datasets/blob/13a5b7db992ad5cf77895e4c0f76595314390418/datasets/wikipedia/wikipedia.py#L494
I replaced them with `elem.clear()`, then it seems to work correctly.
here is the notebook to reproduce it.
https://gist.github.com/miyamonz/dc06117302b6e85fa51cbf46dde6bb51#file-xtract_content-ipynb
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2031/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2031/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2030
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2030/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2030/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2030/events
|
https://github.com/huggingface/datasets/pull/2030
| 829,110,803 |
MDExOlB1bGxSZXF1ZXN0NTkwODI4NzQ4
| 2,030 |
Implement Dataset from text
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-11T12:34:50Z
|
2021-03-18T13:29:29Z
|
2021-03-18T13:29:29Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2030.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2030",
"merged_at": "2021-03-18T13:29:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2030.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2030"
}
|
Implement `Dataset.from_text`.
Analogue to #1943, #1946.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2030/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2030/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2029
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2029/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2029/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2029/events
|
https://github.com/huggingface/datasets/issues/2029
| 829,097,290 |
MDU6SXNzdWU4MjkwOTcyOTA=
| 2,029 |
Loading a faiss index KeyError
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/24982805?v=4",
"events_url": "https://api.github.com/users/nbroad1881/events{/privacy}",
"followers_url": "https://api.github.com/users/nbroad1881/followers",
"following_url": "https://api.github.com/users/nbroad1881/following{/other_user}",
"gists_url": "https://api.github.com/users/nbroad1881/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/nbroad1881",
"id": 24982805,
"login": "nbroad1881",
"node_id": "MDQ6VXNlcjI0OTgyODA1",
"organizations_url": "https://api.github.com/users/nbroad1881/orgs",
"received_events_url": "https://api.github.com/users/nbroad1881/received_events",
"repos_url": "https://api.github.com/users/nbroad1881/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/nbroad1881/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nbroad1881/subscriptions",
"type": "User",
"url": "https://api.github.com/users/nbroad1881"
}
|
[
{
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation",
"id": 1935892861,
"name": "documentation",
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation"
}
] |
closed
| false | null |
[] | null |
[] |
2021-03-11T12:16:13Z
|
2021-03-12T00:21:09Z
|
2021-03-12T00:21:09Z
|
NONE
| null | null | null |
I've recently been testing out RAG and DPR embeddings, and I've run into an issue that is not apparent in the documentation.
The basic steps are:
1. Create a dataset (dataset1)
2. Create an embeddings column using DPR
3. Add a faiss index to the dataset
4. Save faiss index to a file
5. Create a new dataset (dataset2) with the same text and label information as dataset1
6. Try to load the faiss index from file to dataset2
7. Get `KeyError: "Column embeddings not in the dataset"`
I've made a colab notebook that should show exactly what I did. Please switch to GPU runtime; I didn't check on CPU.
https://colab.research.google.com/drive/1X0S9ZuZ8k0ybcoei4w7so6dS_WrABmIx?usp=sharing
Ubuntu Version
VERSION="18.04.5 LTS (Bionic Beaver)"
datasets==1.4.1
faiss==1.5.3
faiss-gpu==1.7.0
torch==1.8.0+cu101
transformers==4.3.3
NVIDIA-SMI 460.56
Driver Version: 460.32.03
CUDA Version: 11.2
Tesla K80
I was basically following the steps here: https://huggingface.co/docs/datasets/faiss_and_ea.html#adding-a-faiss-index
I included the exact code from the documentation at the end of the notebook to show that they don't work either.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2029/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2029/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2028
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2028/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2028/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2028/events
|
https://github.com/huggingface/datasets/pull/2028
| 828,721,393 |
MDExOlB1bGxSZXF1ZXN0NTkwNDk1NzEx
| 2,028 |
Adding PersiNLU reading-comprehension
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/2441454?v=4",
"events_url": "https://api.github.com/users/danyaljj/events{/privacy}",
"followers_url": "https://api.github.com/users/danyaljj/followers",
"following_url": "https://api.github.com/users/danyaljj/following{/other_user}",
"gists_url": "https://api.github.com/users/danyaljj/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/danyaljj",
"id": 2441454,
"login": "danyaljj",
"node_id": "MDQ6VXNlcjI0NDE0NTQ=",
"organizations_url": "https://api.github.com/users/danyaljj/orgs",
"received_events_url": "https://api.github.com/users/danyaljj/received_events",
"repos_url": "https://api.github.com/users/danyaljj/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/danyaljj/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/danyaljj/subscriptions",
"type": "User",
"url": "https://api.github.com/users/danyaljj"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-11T04:41:13Z
|
2021-03-15T09:39:57Z
|
2021-03-15T09:39:57Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2028.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2028",
"merged_at": "2021-03-15T09:39:57Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2028.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2028"
}
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2028/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2028/timeline
| null | null | true |
|
https://api.github.com/repos/huggingface/datasets/issues/2027
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2027/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2027/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2027/events
|
https://github.com/huggingface/datasets/pull/2027
| 828,490,444 |
MDExOlB1bGxSZXF1ZXN0NTkwMjkzNDA1
| 2,027 |
Update format columns in Dataset.rename_columns
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-10T23:50:59Z
|
2021-03-11T14:38:40Z
|
2021-03-11T14:38:40Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2027.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2027",
"merged_at": "2021-03-11T14:38:40Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2027.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2027"
}
|
Fixes #2026
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2027/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2027/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2026
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2026/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2026/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2026/events
|
https://github.com/huggingface/datasets/issues/2026
| 828,194,467 |
MDU6SXNzdWU4MjgxOTQ0Njc=
| 2,026 |
KeyError on using map after renaming a column
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4",
"events_url": "https://api.github.com/users/gchhablani/events{/privacy}",
"followers_url": "https://api.github.com/users/gchhablani/followers",
"following_url": "https://api.github.com/users/gchhablani/following{/other_user}",
"gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gchhablani",
"id": 29076344,
"login": "gchhablani",
"node_id": "MDQ6VXNlcjI5MDc2MzQ0",
"organizations_url": "https://api.github.com/users/gchhablani/orgs",
"received_events_url": "https://api.github.com/users/gchhablani/received_events",
"repos_url": "https://api.github.com/users/gchhablani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gchhablani"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-10T18:54:17Z
|
2021-03-11T14:39:34Z
|
2021-03-11T14:38:40Z
|
CONTRIBUTOR
| null | null | null |
Hi,
I'm trying to use `cifar10` dataset. I want to rename the `img` feature to `image` in order to make it consistent with `mnist`, which I'm also planning to use. By doing this, I was trying to avoid modifying `prepare_train_features` function.
Here is what I try:
```python
transform = Compose([ToPILImage(),ToTensor(),Normalize([0.0,0.0,0.0],[1.0,1.0,1.0])])
def prepare_features(examples):
images = []
labels = []
print(examples)
for example_idx, example in enumerate(examples["image"]):
if transform is not None:
images.append(transform(examples["image"][example_idx].permute(2,0,1)))
else:
images.append(examples["image"][example_idx].permute(2,0,1))
labels.append(examples["label"][example_idx])
output = {"label":labels, "image":images}
return output
raw_dataset = load_dataset('cifar10')
raw_dataset.set_format('torch',columns=['img','label'])
raw_dataset = raw_dataset.rename_column('img','image')
features = datasets.Features({
"image": datasets.Array3D(shape=(3,32,32),dtype="float32"),
"label": datasets.features.ClassLabel(names=[
"airplane",
"automobile",
"bird",
"cat",
"deer",
"dog",
"frog",
"horse",
"ship",
"truck",
]),
})
train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)
```
The error:
```python
---------------------------------------------------------------------------
KeyError Traceback (most recent call last)
<ipython-input-54-bf29672c53ee> in <module>()
14 ]),
15 })
---> 16 train_dataset = raw_dataset.map(prepare_features, features = features,batched=True, batch_size=10000)
2 frames
/usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint)
1287 test_inputs = self[:2] if batched else self[0]
1288 test_indices = [0, 1] if batched else 0
-> 1289 update_data = does_function_return_dict(test_inputs, test_indices)
1290 logger.info("Testing finished, running the mapping function on the dataset")
1291
/usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py in does_function_return_dict(inputs, indices)
1258 fn_args = [inputs] if input_columns is None else [inputs[col] for col in input_columns]
1259 processed_inputs = (
-> 1260 function(*fn_args, indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs)
1261 )
1262 does_return_dict = isinstance(processed_inputs, Mapping)
<ipython-input-52-b4dccbafb70d> in prepare_features(examples)
3 labels = []
4 print(examples)
----> 5 for example_idx, example in enumerate(examples["image"]):
6 if transform is not None:
7 images.append(transform(examples["image"][example_idx].permute(2,0,1)))
KeyError: 'image'
```
The print statement inside returns this:
```python
{'label': tensor([6, 9])}
```
Apparently, both `img` and `image` do not exist after renaming.
Note that this code works fine with `img` everywhere.
Notebook: https://colab.research.google.com/drive/1SzESAlz3BnVYrgQeJ838vbMp1OsukiA2?usp=sharing
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2026/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2026/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2025
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2025/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2025/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2025/events
|
https://github.com/huggingface/datasets/pull/2025
| 828,047,476 |
MDExOlB1bGxSZXF1ZXN0NTg5ODk2NjMz
| 2,025 |
[Refactor] Use in-memory/memory-mapped/concatenation tables in Dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-10T17:00:47Z
|
2021-03-30T14:46:53Z
|
2021-03-26T16:51:59Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2025.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2025",
"merged_at": "2021-03-26T16:51:58Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2025.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2025"
}
|
## Intro
Currently there is one assumption that we need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk with memory mapping (using the dataset._data_files).
This assumption is used for pickling for example:
- in-memory dataset can just be pickled/unpickled in-memory
- on-disk dataset can be unloaded to only keep the filepaths when pickling, and then reloaded from the disk when unpickling
## Issues
Because of this assumption, we can't easily implement methods like `Dataset.add_item` to append more rows to a dataset, or `dataset.add_column` to add a column, since we can't mix data from memory and data from the disk.
Moreover, `concatenate_datasets` doesn't work if the datasets to concatenate are not all from memory, or all form the disk.
## Solution provided in this PR
I changed this by allowing several types of Table to be used in the Dataset object.
More specifically I added three pyarrow Table wrappers: InMemoryTable, MemoryMappedTable and ConcatenationTable.
The in-memory and memory-mapped tables implement the pickling behavior described above.
The ConcatenationTable can be made from several tables (either in-memory or memory mapped) called "blocks". Pickling a ConcatenationTable simply pickles the underlying blocks.
## Implementation details
The three tables classes mentioned above all inherit from a `Table` class defined in `table.py`, which is a wrapper of a pyarrow table. The `Table` wrapper implements all the attributes and methods of the underlying pyarrow table.
Regarding the MemoryMappedTable:
Reloading a pyarrow table from the disk makes you lose all the changes you may have applied (slice, rename_columns, drop, cast etc.). Therefore the MemoryMappedTable implements a "replay" mechanism to re-apply the changes when reloading the pyarrow table from the disk.
## Checklist
- [x] add InMemoryTable
- [x] add MemoryMappedTable
- [x] add ConcatenationTable
- [x] Update the ArrowReader to use these new tables depending on the `in_memory` parameter
- [x] Update Dataset.from_xxx methods
- [x] Update load_from_disk and save_to_disk
- [x] Backward compatibility of load_from_disk
- [x] Add tests for the new tables
- [x] Update current tests
- [ ] Documentation
----------
I would be happy to discuss the design of this PR :)
Close #1877
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 3,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 3,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2025/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2025/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2024
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2024/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2024/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2024/events
|
https://github.com/huggingface/datasets/pull/2024
| 827,842,962 |
MDExOlB1bGxSZXF1ZXN0NTg5NzEzNDAy
| 2,024 |
Remove print statement from mnist.py
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4",
"events_url": "https://api.github.com/users/gchhablani/events{/privacy}",
"followers_url": "https://api.github.com/users/gchhablani/followers",
"following_url": "https://api.github.com/users/gchhablani/following{/other_user}",
"gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gchhablani",
"id": 29076344,
"login": "gchhablani",
"node_id": "MDQ6VXNlcjI5MDc2MzQ0",
"organizations_url": "https://api.github.com/users/gchhablani/orgs",
"received_events_url": "https://api.github.com/users/gchhablani/received_events",
"repos_url": "https://api.github.com/users/gchhablani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gchhablani"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-10T14:39:58Z
|
2021-03-11T18:03:52Z
|
2021-03-11T18:03:51Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2024.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2024",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2024.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2024"
}
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2024/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2024/timeline
| null | null | true |
|
https://api.github.com/repos/huggingface/datasets/issues/2023
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2023/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2023/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2023/events
|
https://github.com/huggingface/datasets/pull/2023
| 827,819,608 |
MDExOlB1bGxSZXF1ZXN0NTg5NjkyNDU2
| 2,023 |
Add Romanian to XQuAD
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/9285264?v=4",
"events_url": "https://api.github.com/users/M-Salti/events{/privacy}",
"followers_url": "https://api.github.com/users/M-Salti/followers",
"following_url": "https://api.github.com/users/M-Salti/following{/other_user}",
"gists_url": "https://api.github.com/users/M-Salti/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/M-Salti",
"id": 9285264,
"login": "M-Salti",
"node_id": "MDQ6VXNlcjkyODUyNjQ=",
"organizations_url": "https://api.github.com/users/M-Salti/orgs",
"received_events_url": "https://api.github.com/users/M-Salti/received_events",
"repos_url": "https://api.github.com/users/M-Salti/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/M-Salti/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/M-Salti/subscriptions",
"type": "User",
"url": "https://api.github.com/users/M-Salti"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-10T14:24:32Z
|
2021-03-15T10:08:17Z
|
2021-03-15T10:08:17Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2023.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2023",
"merged_at": "2021-03-15T10:08:17Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2023.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2023"
}
|
On Jan 18, XQuAD was updated with a new Romanian validation file ([xquad commit link](https://github.com/deepmind/xquad/commit/60cac411649156efb6aab9dd4c9cde787a2c0345))
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2023/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2023/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2022
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2022/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2022/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2022/events
|
https://github.com/huggingface/datasets/issues/2022
| 827,435,033 |
MDU6SXNzdWU4Mjc0MzUwMzM=
| 2,022 |
ValueError when rename_column on splitted dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/53626067?v=4",
"events_url": "https://api.github.com/users/simonschoe/events{/privacy}",
"followers_url": "https://api.github.com/users/simonschoe/followers",
"following_url": "https://api.github.com/users/simonschoe/following{/other_user}",
"gists_url": "https://api.github.com/users/simonschoe/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/simonschoe",
"id": 53626067,
"login": "simonschoe",
"node_id": "MDQ6VXNlcjUzNjI2MDY3",
"organizations_url": "https://api.github.com/users/simonschoe/orgs",
"received_events_url": "https://api.github.com/users/simonschoe/received_events",
"repos_url": "https://api.github.com/users/simonschoe/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/simonschoe/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/simonschoe/subscriptions",
"type": "User",
"url": "https://api.github.com/users/simonschoe"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-10T09:40:38Z
|
2021-03-16T14:06:08Z
|
2021-03-16T14:05:05Z
|
NONE
| null | null | null |
Hi there,
I am loading `.tsv` file via `load_dataset` and subsequently split the rows into training and test set via the `ReadInstruction` API like so:
```python
split = {
'train': ReadInstruction('train', to=90, unit='%'),
'test': ReadInstruction('train', from_=-10, unit='%')
}
dataset = load_dataset(
path='csv', # use 'text' loading script to load from local txt-files
delimiter='\t', # xxx
data_files=text_files, # list of paths to local text files
split=split, # xxx
)
dataset
```
Part of output:
```python
DatasetDict({
train: Dataset({
features: ['sentence', 'sentiment'],
num_rows: 900
})
test: Dataset({
features: ['sentence', 'sentiment'],
num_rows: 100
})
})
```
Afterwards I'd like to rename the 'sentence' column to 'text' in order to be compatible with my modelin pipeline. If I run the following code I experience a `ValueError` however:
```python
dataset['train'].rename_column('sentence', 'text')
```
```python
/usr/local/lib/python3.7/dist-packages/datasets/splits.py in __init__(self, name)
353 for split_name in split_names_from_instruction:
354 if not re.match(_split_re, split_name):
--> 355 raise ValueError(f"Split name should match '{_split_re}'' but got '{split_name}'.")
356
357 def __str__(self):
ValueError: Split name should match '^\w+(\.\w+)*$'' but got 'ReadInstruction('.
```
In particular, these behavior does not arise if I use the deprecated `rename_column_` method. Any idea what causes the error? Would assume something in the way I defined the split.
Thanks in advance! :)
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2022/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2022/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2021
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2021/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2021/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2021/events
|
https://github.com/huggingface/datasets/issues/2021
| 826,988,016 |
MDU6SXNzdWU4MjY5ODgwMTY=
| 2,021 |
Interactively doing save_to_disk and load_from_disk corrupts the datasets object?
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4",
"events_url": "https://api.github.com/users/shamanez/events{/privacy}",
"followers_url": "https://api.github.com/users/shamanez/followers",
"following_url": "https://api.github.com/users/shamanez/following{/other_user}",
"gists_url": "https://api.github.com/users/shamanez/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/shamanez",
"id": 16892570,
"login": "shamanez",
"node_id": "MDQ6VXNlcjE2ODkyNTcw",
"organizations_url": "https://api.github.com/users/shamanez/orgs",
"received_events_url": "https://api.github.com/users/shamanez/received_events",
"repos_url": "https://api.github.com/users/shamanez/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/shamanez/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shamanez/subscriptions",
"type": "User",
"url": "https://api.github.com/users/shamanez"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-10T02:48:34Z
|
2021-03-13T10:07:41Z
|
2021-03-13T10:07:41Z
|
NONE
| null | null | null |
dataset_info.json file saved after using save_to_disk gets corrupted as follows.

Is there a way to disable the cache that will save to /tmp/huggiface/datastes ?
I have a feeling there is a serious issue with cashing.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2021/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2021/timeline
| null |
completed
| true |
https://api.github.com/repos/huggingface/datasets/issues/2020
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2020/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2020/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2020/events
|
https://github.com/huggingface/datasets/pull/2020
| 826,961,126 |
MDExOlB1bGxSZXF1ZXN0NTg4OTE3MjYx
| 2,020 |
Remove unnecessary docstart check in conll-like datasets
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-10T02:20:16Z
|
2021-03-11T13:33:37Z
|
2021-03-11T13:33:37Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2020.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2020",
"merged_at": "2021-03-11T13:33:37Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2020.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2020"
}
|
Related to this PR: #1998
Additionally, this PR adds the docstart note to the conll2002 dataset card ([link](https://raw.githubusercontent.com/teropa/nlp/master/resources/corpora/conll2002/ned.train) to the raw data with `DOCSTART` lines).
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2020/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2020/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2019
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2019/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2019/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2019/events
|
https://github.com/huggingface/datasets/pull/2019
| 826,625,706 |
MDExOlB1bGxSZXF1ZXN0NTg4NjEyODgy
| 2,019 |
Replace print with logging in dataset scripts
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-09T20:59:34Z
|
2021-03-12T10:09:01Z
|
2021-03-11T16:14:19Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2019.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2019",
"merged_at": "2021-03-11T16:14:18Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2019.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2019"
}
|
Replaces `print(...)` in the dataset scripts with the library logger.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2019/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2019/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2018
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2018/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2018/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2018/events
|
https://github.com/huggingface/datasets/pull/2018
| 826,473,764 |
MDExOlB1bGxSZXF1ZXN0NTg4NDc0NTQz
| 2,018 |
Md gender card update
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4",
"events_url": "https://api.github.com/users/mcmillanmajora/events{/privacy}",
"followers_url": "https://api.github.com/users/mcmillanmajora/followers",
"following_url": "https://api.github.com/users/mcmillanmajora/following{/other_user}",
"gists_url": "https://api.github.com/users/mcmillanmajora/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mcmillanmajora",
"id": 26722925,
"login": "mcmillanmajora",
"node_id": "MDQ6VXNlcjI2NzIyOTI1",
"organizations_url": "https://api.github.com/users/mcmillanmajora/orgs",
"received_events_url": "https://api.github.com/users/mcmillanmajora/received_events",
"repos_url": "https://api.github.com/users/mcmillanmajora/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mcmillanmajora/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mcmillanmajora/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mcmillanmajora"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-09T18:57:20Z
|
2021-03-12T17:31:00Z
|
2021-03-12T17:31:00Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2018.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2018",
"merged_at": "2021-03-12T17:31:00Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2018.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2018"
}
|
I updated the descriptions of the datasets as they appear in the HF repo and the descriptions of the source datasets according to what I could find from the paper and the references. I'm still a little unclear about some of the fields of the different configs, and there was little info on the word list and name list. I'll contact the authors to see if they have any additional information or suggested changes.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2018/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2018/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2017
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2017/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2017/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2017/events
|
https://github.com/huggingface/datasets/pull/2017
| 826,428,578 |
MDExOlB1bGxSZXF1ZXN0NTg4NDMyNDc2
| 2,017 |
Add TF-based Features to handle different modes of data
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4",
"events_url": "https://api.github.com/users/gchhablani/events{/privacy}",
"followers_url": "https://api.github.com/users/gchhablani/followers",
"following_url": "https://api.github.com/users/gchhablani/following{/other_user}",
"gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gchhablani",
"id": 29076344,
"login": "gchhablani",
"node_id": "MDQ6VXNlcjI5MDc2MzQ0",
"organizations_url": "https://api.github.com/users/gchhablani/orgs",
"received_events_url": "https://api.github.com/users/gchhablani/received_events",
"repos_url": "https://api.github.com/users/gchhablani/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gchhablani"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-09T18:29:52Z
|
2021-03-17T12:32:08Z
|
2021-03-17T12:32:07Z
|
CONTRIBUTOR
| null | true |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2017.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2017",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2017.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2017"
}
|
Hi,
I am creating this draft PR to work on add features similar to [TF datasets](https://github.com/tensorflow/datasets/tree/master/tensorflow_datasets/core/features). I'll be starting with `Tensor` and `FeatureConnector` classes, and build upon them to add other features as well. This is a work in progress.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2017/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2017/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2016
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2016/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2016/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2016/events
|
https://github.com/huggingface/datasets/pull/2016
| 825,965,493 |
MDExOlB1bGxSZXF1ZXN0NTg4MDA5NjEz
| 2,016 |
Not all languages have 2 digit codes.
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/13891775?v=4",
"events_url": "https://api.github.com/users/asiddhant/events{/privacy}",
"followers_url": "https://api.github.com/users/asiddhant/followers",
"following_url": "https://api.github.com/users/asiddhant/following{/other_user}",
"gists_url": "https://api.github.com/users/asiddhant/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/asiddhant",
"id": 13891775,
"login": "asiddhant",
"node_id": "MDQ6VXNlcjEzODkxNzc1",
"organizations_url": "https://api.github.com/users/asiddhant/orgs",
"received_events_url": "https://api.github.com/users/asiddhant/received_events",
"repos_url": "https://api.github.com/users/asiddhant/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/asiddhant/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/asiddhant/subscriptions",
"type": "User",
"url": "https://api.github.com/users/asiddhant"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-09T13:53:39Z
|
2021-03-11T18:01:03Z
|
2021-03-11T18:01:03Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2016.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2016",
"merged_at": "2021-03-11T18:01:03Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2016.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2016"
}
|
.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2016/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2016/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2015
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2015/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2015/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2015/events
|
https://github.com/huggingface/datasets/pull/2015
| 825,942,108 |
MDExOlB1bGxSZXF1ZXN0NTg3OTg4NTQ0
| 2,015 |
Fix ipython function creation in tests
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-09T13:36:59Z
|
2021-03-09T14:06:04Z
|
2021-03-09T14:06:03Z
|
MEMBER
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2015.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2015",
"merged_at": "2021-03-09T14:06:03Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2015.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2015"
}
|
The test at `tests/test_caching.py::RecurseDumpTest::test_dump_ipython_function` was failing in python 3.8 because the ipython function was not properly created.
Fix #2010
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2015/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2015/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2014
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2014/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2014/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2014/events
|
https://github.com/huggingface/datasets/pull/2014
| 825,916,531 |
MDExOlB1bGxSZXF1ZXN0NTg3OTY1NDg3
| 2,014 |
more explicit method parameters
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-09T13:18:29Z
|
2021-03-10T10:08:37Z
|
2021-03-10T10:08:36Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2014.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2014",
"merged_at": "2021-03-10T10:08:36Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2014.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2014"
}
|
re: #2009
not super convinced this is better, and while I usually fight against kwargs here it seems to me that it better conveys the relationship to the `_split_generator` method.
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2014/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2014/timeline
| null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2013
|
https://api.github.com/repos/huggingface/datasets
|
https://api.github.com/repos/huggingface/datasets/issues/2013/labels{/name}
|
https://api.github.com/repos/huggingface/datasets/issues/2013/comments
|
https://api.github.com/repos/huggingface/datasets/issues/2013/events
|
https://github.com/huggingface/datasets/pull/2013
| 825,694,305 |
MDExOlB1bGxSZXF1ZXN0NTg3NzYzMTgx
| 2,013 |
Add Cryptonite dataset
|
{
"avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4",
"events_url": "https://api.github.com/users/theo-m/events{/privacy}",
"followers_url": "https://api.github.com/users/theo-m/followers",
"following_url": "https://api.github.com/users/theo-m/following{/other_user}",
"gists_url": "https://api.github.com/users/theo-m/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/theo-m",
"id": 17948980,
"login": "theo-m",
"node_id": "MDQ6VXNlcjE3OTQ4OTgw",
"organizations_url": "https://api.github.com/users/theo-m/orgs",
"received_events_url": "https://api.github.com/users/theo-m/received_events",
"repos_url": "https://api.github.com/users/theo-m/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/theo-m/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theo-m/subscriptions",
"type": "User",
"url": "https://api.github.com/users/theo-m"
}
|
[] |
closed
| false | null |
[] | null |
[] |
2021-03-09T10:32:11Z
|
2021-03-09T19:27:07Z
|
2021-03-09T19:27:06Z
|
CONTRIBUTOR
| null | false |
{
"diff_url": "https://github.com/huggingface/datasets/pull/2013.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2013",
"merged_at": "2021-03-09T19:27:06Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2013.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2013"
}
|
cc @aviaefrat who's the original author of the dataset & paper, see https://github.com/aviaefrat/cryptonite
|
{
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2013/reactions"
}
|
https://api.github.com/repos/huggingface/datasets/issues/2013/timeline
| null | null | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.