Dataset Viewer
Auto-converted to Parquet
url
stringlengths
61
61
repository_url
stringclasses
1 value
labels_url
stringlengths
75
75
comments_url
stringlengths
70
70
events_url
stringlengths
68
68
html_url
stringlengths
49
51
id
int64
942M
3.2B
node_id
stringlengths
18
32
number
int64
2.63k
7.67k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
int64
0
49
created_at
stringdate
2021-07-12 19:58:31
2025-07-03 11:24:15
updated_at
stringdate
2021-07-13 05:45:26
2025-07-03 18:34:32
closed_at
stringlengths
20
20
author_association
stringclasses
4 values
type
null
active_lock_reason
null
sub_issues_summary
dict
body
stringlengths
0
58.6k
closed_by
dict
reactions
dict
timeline_url
stringlengths
70
70
performed_via_github_app
null
state_reason
stringclasses
4 values
draft
bool
2 classes
pull_request
dict
https://api.github.com/repos/huggingface/datasets/issues/7668
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7668/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7668/comments
https://api.github.com/repos/huggingface/datasets/issues/7668/events
https://github.com/huggingface/datasets/issues/7668
3,199,039,322
I_kwDODunzps6-rXda
7,668
Broken EXIF crash the whole program
{ "avatar_url": "https://avatars.githubusercontent.com/u/30485844?v=4", "events_url": "https://api.github.com/users/Seas0/events{/privacy}", "followers_url": "https://api.github.com/users/Seas0/followers", "following_url": "https://api.github.com/users/Seas0/following{/other_user}", "gists_url": "https://api.github.com/users/Seas0/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Seas0", "id": 30485844, "login": "Seas0", "node_id": "MDQ6VXNlcjMwNDg1ODQ0", "organizations_url": "https://api.github.com/users/Seas0/orgs", "received_events_url": "https://api.github.com/users/Seas0/received_events", "repos_url": "https://api.github.com/users/Seas0/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Seas0/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Seas0/subscriptions", "type": "User", "url": "https://api.github.com/users/Seas0", "user_view_type": "public" }
[]
open
false
null
[]
null
1
2025-07-03T11:24:15Z
2025-07-03T12:27:16Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug When parsing this image in the ImageNet1K dataset, the `datasets` crashs whole training process just because unable to parse an invalid EXIF tag. ![Image](https://github.com/user-attachments/assets/3c840203-ac8c-41a0-9cf7-45f64488037d) ### Steps to reproduce the bug Use the `datasets.Image.decode_example` method to decode the aforementioned image could reproduce the bug. The decoding function will throw an unhandled exception at the `image.getexif()` method call due to invalid utf-8 stream in EXIF tags. ``` File lib/python3.12/site-packages/datasets/features/image.py:188, in Image.decode_example(self, value, token_per_repo_id) 186 image = PIL.Image.open(BytesIO(bytes_)) 187 image.load() # to avoid "Too many open files" errors --> 188 if image.getexif().get(PIL.Image.ExifTags.Base.Orientation) is not None: 189 image = PIL.ImageOps.exif_transpose(image) 190 if self.mode and self.mode != image.mode: File lib/python3.12/site-packages/PIL/Image.py:1542, in Image.getexif(self) 1540 xmp_tags = self.info.get("XML:com.adobe.xmp") 1541 if not xmp_tags and (xmp_tags := self.info.get("xmp")): -> 1542 xmp_tags = xmp_tags.decode("utf-8") 1543 if xmp_tags: 1544 match = re.search(r'tiff:Orientation(="|>)([0-9])', xmp_tags) UnicodeDecodeError: 'utf-8' codec can't decode byte 0xa8 in position 4312: invalid start byte ``` ### Expected behavior The invalid EXIF tag should simply be ignored or issue a warning message, instead of crash the whole program at once. ### Environment info - `datasets` version: 3.6.0 - Platform: Linux-6.5.0-18-generic-x86_64-with-glibc2.35 - Python version: 3.12.11 - `huggingface_hub` version: 0.33.0 - PyArrow version: 20.0.0 - Pandas version: 2.3.0 - `fsspec` version: 2025.3.0
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7668/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7668/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7667
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7667/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7667/comments
https://api.github.com/repos/huggingface/datasets/issues/7667/events
https://github.com/huggingface/datasets/pull/7667
3,196,251,707
PR_kwDODunzps6dGmm8
7,667
Fix infer list of images
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-07-02T15:07:58Z
2025-07-02T15:10:28Z
2025-07-02T15:08:03Z
MEMBER
null
null
null
cc @kashif
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7667/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7667/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7667.diff", "html_url": "https://github.com/huggingface/datasets/pull/7667", "merged_at": "2025-07-02T15:08:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/7667.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7667" }
https://api.github.com/repos/huggingface/datasets/issues/7666
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7666/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7666/comments
https://api.github.com/repos/huggingface/datasets/issues/7666/events
https://github.com/huggingface/datasets/pull/7666
3,196,220,722
PR_kwDODunzps6dGf7E
7,666
Backward compat list feature
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-07-02T14:58:00Z
2025-07-02T15:00:37Z
2025-07-02T14:59:40Z
MEMBER
null
null
null
cc @kashif
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7666/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7666/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7666.diff", "html_url": "https://github.com/huggingface/datasets/pull/7666", "merged_at": "2025-07-02T14:59:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/7666.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7666" }
https://api.github.com/repos/huggingface/datasets/issues/7665
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7665/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7665/comments
https://api.github.com/repos/huggingface/datasets/issues/7665/events
https://github.com/huggingface/datasets/issues/7665
3,193,239,955
I_kwDODunzps6-VPmT
7,665
Function load_dataset() misinterprets string field content as part of dataset schema when dealing with `.jsonl` files
{ "avatar_url": "https://avatars.githubusercontent.com/u/1151198?v=4", "events_url": "https://api.github.com/users/zdzichukowalski/events{/privacy}", "followers_url": "https://api.github.com/users/zdzichukowalski/followers", "following_url": "https://api.github.com/users/zdzichukowalski/following{/other_user}", "gists_url": "https://api.github.com/users/zdzichukowalski/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zdzichukowalski", "id": 1151198, "login": "zdzichukowalski", "node_id": "MDQ6VXNlcjExNTExOTg=", "organizations_url": "https://api.github.com/users/zdzichukowalski/orgs", "received_events_url": "https://api.github.com/users/zdzichukowalski/received_events", "repos_url": "https://api.github.com/users/zdzichukowalski/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zdzichukowalski/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zdzichukowalski/subscriptions", "type": "User", "url": "https://api.github.com/users/zdzichukowalski", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-07-01T17:14:53Z
2025-07-01T17:17:48Z
2025-07-01T17:17:48Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug When loading a `.jsonl` file using `load_dataset("json", data_files="data.jsonl", split="train")`, the function misinterprets the content of a string field as if it were part of the dataset schema. In my case there is a field `body:` with a string value ``` "### Describe the bug (...) ,action: string, datetime: timestamp[s], author: string, (...) Pandas version: 1.3.4" ``` As a result, I got an exception ``` "TypeError: Couldn't cast array of type timestamp[s] to null". ``` Full stack-trace in the attached file below. I also attach a minimized dataset (data.json, a single entry) that reproduces the error. **Observations**(on the minimal example): - if I remove _all fields before_ `body`, a different error appears, - if I remove _all fields after_ `body`, yet another error appears, - if `body` is _the only field_, the error disappears. So this might be one complex bug or several edge cases interacting. I haven’t dug deeper. Also changing the file extension to `.json` or `.txt` avoids the problem. This suggests **a possible workaround** for the general case: convert `.jsonl` to `.json`. Though I haven’t verified correctness of that workaround yet. Anyway my understanding is that `load_dataset` with first argument set to "json" should properly handle `.jsonl` files. Correct me if I'm wrong. [stack_trace.txt](https://github.com/user-attachments/files/21004153/stack_trace.txt) [data.json](https://github.com/user-attachments/files/21004164/data.json) P.S. I discovered this while going through the HuggingFace tutorial. Specifically [this part](https://huggingface.co/learn/llm-course/chapter5/5?fw=pt).I will try to inform the tutorial team about this issue, as it can be a showstopper for young 🤗 adepts. ### Steps to reproduce the bug 1. Download attached [data.json](https://github.com/user-attachments/files/21004164/data.json) file. 2. Run the following code which should work correctly: ``` from datasets import load_dataset load_dataset("json", data_files="data.json", split="train") ``` 3. Change extension of the `data` file to `.jsonl` and run: ``` from datasets import load_dataset load_dataset("json", data_files="data.jsonl", split="train") ``` This will trigger an error like the one in the attached [stack_trace.txt](https://github.com/user-attachments/files/21004153/stack_trace.txt). One can also try removing fields before the `body` field and after it. These actions give different errors. ### Expected behavior Parsing data in `.jsonl` format should yield the same result as parsing the same data in `.json` format. In any case, the content of a string field should never be interpreted as part of the dataset schema. ### Environment info datasets version: _3.6.0_ pyarrow version: _20.0.0_ Python version: _3.11.9_ platform version: _macOS-15.5-arm64-arm-64bit_
{ "avatar_url": "https://avatars.githubusercontent.com/u/1151198?v=4", "events_url": "https://api.github.com/users/zdzichukowalski/events{/privacy}", "followers_url": "https://api.github.com/users/zdzichukowalski/followers", "following_url": "https://api.github.com/users/zdzichukowalski/following{/other_user}", "gists_url": "https://api.github.com/users/zdzichukowalski/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zdzichukowalski", "id": 1151198, "login": "zdzichukowalski", "node_id": "MDQ6VXNlcjExNTExOTg=", "organizations_url": "https://api.github.com/users/zdzichukowalski/orgs", "received_events_url": "https://api.github.com/users/zdzichukowalski/received_events", "repos_url": "https://api.github.com/users/zdzichukowalski/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zdzichukowalski/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zdzichukowalski/subscriptions", "type": "User", "url": "https://api.github.com/users/zdzichukowalski", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7665/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7665/timeline
null
duplicate
null
null
https://api.github.com/repos/huggingface/datasets/issues/7664
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7664/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7664/comments
https://api.github.com/repos/huggingface/datasets/issues/7664/events
https://github.com/huggingface/datasets/issues/7664
3,193,239,035
I_kwDODunzps6-VPX7
7,664
Function load_dataset() misinterprets string field content as part of dataset schema when dealing with `.jsonl` files
{ "avatar_url": "https://avatars.githubusercontent.com/u/1151198?v=4", "events_url": "https://api.github.com/users/zdzichukowalski/events{/privacy}", "followers_url": "https://api.github.com/users/zdzichukowalski/followers", "following_url": "https://api.github.com/users/zdzichukowalski/following{/other_user}", "gists_url": "https://api.github.com/users/zdzichukowalski/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zdzichukowalski", "id": 1151198, "login": "zdzichukowalski", "node_id": "MDQ6VXNlcjExNTExOTg=", "organizations_url": "https://api.github.com/users/zdzichukowalski/orgs", "received_events_url": "https://api.github.com/users/zdzichukowalski/received_events", "repos_url": "https://api.github.com/users/zdzichukowalski/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zdzichukowalski/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zdzichukowalski/subscriptions", "type": "User", "url": "https://api.github.com/users/zdzichukowalski", "user_view_type": "public" }
[]
open
false
null
[]
null
1
2025-07-01T17:14:32Z
2025-07-03T13:01:59Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug When loading a `.jsonl` file using `load_dataset("json", data_files="data.jsonl", split="train")`, the function misinterprets the content of a string field as if it were part of the dataset schema. In my case there is a field `body:` with a string value ``` "### Describe the bug (...) ,action: string, datetime: timestamp[s], author: string, (...) Pandas version: 1.3.4" ``` As a result, I got an exception ``` "TypeError: Couldn't cast array of type timestamp[s] to null". ``` Full stack-trace in the attached file below. I also attach a minimized dataset (data.json, a single entry) that reproduces the error. **Observations**(on the minimal example): - if I remove _all fields before_ `body`, a different error appears, - if I remove _all fields after_ `body`, yet another error appears, - if `body` is _the only field_, the error disappears. So this might be one complex bug or several edge cases interacting. I haven’t dug deeper. Also changing the file extension to `.json` or `.txt` avoids the problem. This suggests **a possible workaround** for the general case: convert `.jsonl` to `.json`. Though I haven’t verified correctness of that workaround yet. Anyway my understanding is that `load_dataset` with first argument set to "json" should properly handle `.jsonl` files. Correct me if I'm wrong. [stack_trace.txt](https://github.com/user-attachments/files/21004153/stack_trace.txt) [data.json](https://github.com/user-attachments/files/21004164/data.json) P.S. I discovered this while going through the HuggingFace tutorial. Specifically [this part](https://huggingface.co/learn/llm-course/chapter5/5?fw=pt). I will try to inform the tutorial team about this issue, as it can be a showstopper for young 🤗 adepts. ### Steps to reproduce the bug 1. Download attached [data.json](https://github.com/user-attachments/files/21004164/data.json) file. 2. Run the following code which should work correctly: ``` from datasets import load_dataset load_dataset("json", data_files="data.json", split="train") ``` 3. Change extension of the `data` file to `.jsonl` and run: ``` from datasets import load_dataset load_dataset("json", data_files="data.jsonl", split="train") ``` This will trigger an error like the one in the attached [stack_trace.txt](https://github.com/user-attachments/files/21004153/stack_trace.txt). One can also try removing fields before the `body` field and after it. These actions give different errors. ### Expected behavior Parsing data in `.jsonl` format should yield the same result as parsing the same data in `.json` format. In any case, the content of a string field should never be interpreted as part of the dataset schema. ### Environment info datasets version: _3.6.0_ pyarrow version: _20.0.0_ Python version: _3.11.9_ platform version: _macOS-15.5-arm64-arm-64bit_
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7664/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7664/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7663
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7663/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7663/comments
https://api.github.com/repos/huggingface/datasets/issues/7663/events
https://github.com/huggingface/datasets/pull/7663
3,192,582,371
PR_kwDODunzps6c6aJF
7,663
Custom metadata filenames
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-07-01T13:50:36Z
2025-07-01T13:58:41Z
2025-07-01T13:58:39Z
MEMBER
null
null
null
example: https://huggingface.co/datasets/lhoestq/overlapping-subsets-imagefolder/tree/main To make multiple subsets for an imagefolder (one metadata file per subset), e.g. ```yaml configs: - config_name: default metadata_filenames: - metadata.csv - config_name: other metadata_filenames: - metadata2.csv ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7663/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7663/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7663.diff", "html_url": "https://github.com/huggingface/datasets/pull/7663", "merged_at": "2025-07-01T13:58:39Z", "patch_url": "https://github.com/huggingface/datasets/pull/7663.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7663" }
https://api.github.com/repos/huggingface/datasets/issues/7662
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7662/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7662/comments
https://api.github.com/repos/huggingface/datasets/issues/7662/events
https://github.com/huggingface/datasets/issues/7662
3,190,805,531
I_kwDODunzps6-L9Qb
7,662
Applying map after transform with multiprocessing will cause OOM
{ "avatar_url": "https://avatars.githubusercontent.com/u/26482910?v=4", "events_url": "https://api.github.com/users/JunjieLl/events{/privacy}", "followers_url": "https://api.github.com/users/JunjieLl/followers", "following_url": "https://api.github.com/users/JunjieLl/following{/other_user}", "gists_url": "https://api.github.com/users/JunjieLl/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JunjieLl", "id": 26482910, "login": "JunjieLl", "node_id": "MDQ6VXNlcjI2NDgyOTEw", "organizations_url": "https://api.github.com/users/JunjieLl/orgs", "received_events_url": "https://api.github.com/users/JunjieLl/received_events", "repos_url": "https://api.github.com/users/JunjieLl/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JunjieLl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JunjieLl/subscriptions", "type": "User", "url": "https://api.github.com/users/JunjieLl", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-07-01T05:45:57Z
2025-07-01T05:45:57Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug I have a 30TB dataset. When I perform add_column and cast_column operations on it and then execute a multiprocessing map, it results in an OOM (Out of Memory) error. However, if I skip the add_column and cast_column steps and directly run the map, there is no OOM. After debugging step by step, I found that the OOM is caused at this point, and I suspect it’s because the add_column and cast_column operations are not cached, which causes the entire dataset to be loaded in each subprocess, leading to the OOM. The critical line of code is: https://github.com/huggingface/datasets/blob/e71b0b19d79c7531f9b9bea7c09916b5f6157f42/src/datasets/utils/py_utils.py#L607 Note num_process=1 would not cause OOM. I'm confused. ### Steps to reproduce the bug For reproduce, you can load dataset and set cache_dir (for caching): amphion/Emilia-Dataset which is a veru large datasets that RAM can not fits. And apply the map with multiprocessing after a transform operation (e.g. add_column, cast_column). As long as num_process>1, it must cause OOM. ### Expected behavior It should not cause OOM. ### Environment info - `datasets` version: 3.6.0 - Platform: Linux-5.10.134-16.101.al8.x86_64-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.33.1 - PyArrow version: 20.0.0 - Pandas version: 2.3.0 - `fsspec` version: 2024.6.1
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7662/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7662/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7661
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7661/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7661/comments
https://api.github.com/repos/huggingface/datasets/issues/7661/events
https://github.com/huggingface/datasets/pull/7661
3,190,408,237
PR_kwDODunzps6czBDi
7,661
fix del tqdm lock error
{ "avatar_url": "https://avatars.githubusercontent.com/u/44766273?v=4", "events_url": "https://api.github.com/users/Hypothesis-Z/events{/privacy}", "followers_url": "https://api.github.com/users/Hypothesis-Z/followers", "following_url": "https://api.github.com/users/Hypothesis-Z/following{/other_user}", "gists_url": "https://api.github.com/users/Hypothesis-Z/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Hypothesis-Z", "id": 44766273, "login": "Hypothesis-Z", "node_id": "MDQ6VXNlcjQ0NzY2Mjcz", "organizations_url": "https://api.github.com/users/Hypothesis-Z/orgs", "received_events_url": "https://api.github.com/users/Hypothesis-Z/received_events", "repos_url": "https://api.github.com/users/Hypothesis-Z/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Hypothesis-Z/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hypothesis-Z/subscriptions", "type": "User", "url": "https://api.github.com/users/Hypothesis-Z", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-07-01T02:04:02Z
2025-07-01T02:33:04Z
null
NONE
null
null
null
for issue https://github.com/huggingface/datasets/issues/7660
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7661/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7661/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7661.diff", "html_url": "https://github.com/huggingface/datasets/pull/7661", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7661.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7661" }
https://api.github.com/repos/huggingface/datasets/issues/7660
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7660/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7660/comments
https://api.github.com/repos/huggingface/datasets/issues/7660/events
https://github.com/huggingface/datasets/issues/7660
3,189,028,251
I_kwDODunzps6-FLWb
7,660
AttributeError: type object 'tqdm' has no attribute '_lock'
{ "avatar_url": "https://avatars.githubusercontent.com/u/44766273?v=4", "events_url": "https://api.github.com/users/Hypothesis-Z/events{/privacy}", "followers_url": "https://api.github.com/users/Hypothesis-Z/followers", "following_url": "https://api.github.com/users/Hypothesis-Z/following{/other_user}", "gists_url": "https://api.github.com/users/Hypothesis-Z/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Hypothesis-Z", "id": 44766273, "login": "Hypothesis-Z", "node_id": "MDQ6VXNlcjQ0NzY2Mjcz", "organizations_url": "https://api.github.com/users/Hypothesis-Z/orgs", "received_events_url": "https://api.github.com/users/Hypothesis-Z/received_events", "repos_url": "https://api.github.com/users/Hypothesis-Z/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Hypothesis-Z/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hypothesis-Z/subscriptions", "type": "User", "url": "https://api.github.com/users/Hypothesis-Z", "user_view_type": "public" }
[]
open
false
null
[]
null
2
2025-06-30T15:57:16Z
2025-07-03T15:14:27Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug `AttributeError: type object 'tqdm' has no attribute '_lock'` It occurs when I'm trying to load datasets in thread pool. Issue https://github.com/huggingface/datasets/issues/6066 and PR https://github.com/huggingface/datasets/pull/6067 https://github.com/huggingface/datasets/pull/6068 tried to fix this. ### Steps to reproduce the bug Will have to try several times to reproduce the error because it is concerned with threads. 1. save some datasets for test ```pythonfrom datasets import Dataset, DatasetDict import os os.makedirs("test_dataset_shards", exist_ok=True) for i in range(10): data = Dataset.from_dict({"text": [f"example {j}" for j in range(1000000)]}) data = DatasetDict({'train': data}) data.save_to_disk(f"test_dataset_shards/shard_{i}") ``` 2. load them in a thread pool ```python from datasets import load_from_disk from tqdm import tqdm from concurrent.futures import ThreadPoolExecutor, as_completed import glob datas = glob.glob('test_dataset_shards/shard_*') with ThreadPoolExecutor(max_workers=10) as pool: futures = [pool.submit(load_from_disk, it) for it in datas] datas = [] for future in tqdm(as_completed(futures), total=len(futures)): datas.append(future.result()) ``` ### Expected behavior no exception raised ### Environment info datasets==2.19.0 python==3.10
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7660/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7660/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7659
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7659/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7659/comments
https://api.github.com/repos/huggingface/datasets/issues/7659/events
https://github.com/huggingface/datasets/pull/7659
3,187,882,217
PR_kwDODunzps6cqkou
7,659
Update the beans dataset link in Preprocess
{ "avatar_url": "https://avatars.githubusercontent.com/u/5434867?v=4", "events_url": "https://api.github.com/users/HJassar/events{/privacy}", "followers_url": "https://api.github.com/users/HJassar/followers", "following_url": "https://api.github.com/users/HJassar/following{/other_user}", "gists_url": "https://api.github.com/users/HJassar/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/HJassar", "id": 5434867, "login": "HJassar", "node_id": "MDQ6VXNlcjU0MzQ4Njc=", "organizations_url": "https://api.github.com/users/HJassar/orgs", "received_events_url": "https://api.github.com/users/HJassar/received_events", "repos_url": "https://api.github.com/users/HJassar/repos", "site_admin": false, "starred_url": "https://api.github.com/users/HJassar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/HJassar/subscriptions", "type": "User", "url": "https://api.github.com/users/HJassar", "user_view_type": "public" }
[]
closed
false
null
[]
null
0
2025-06-30T09:58:44Z
2025-07-01T14:01:42Z
2025-07-01T14:01:42Z
CONTRIBUTOR
null
null
null
In the Preprocess tutorial, the to "the beans dataset" is incorrect. Fixed.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7659/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7659/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7659.diff", "html_url": "https://github.com/huggingface/datasets/pull/7659", "merged_at": "2025-07-01T14:01:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/7659.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7659" }
https://api.github.com/repos/huggingface/datasets/issues/7658
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7658/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7658/comments
https://api.github.com/repos/huggingface/datasets/issues/7658/events
https://github.com/huggingface/datasets/pull/7658
3,187,800,504
PR_kwDODunzps6cqTMs
7,658
Fix: Prevent loss of info.features and column_names in IterableDatasetDict.map when features is None
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
closed
false
null
[]
null
5
2025-06-30T09:31:12Z
2025-07-01T16:26:30Z
2025-07-01T16:26:12Z
CONTRIBUTOR
null
null
null
This PR fixes a bug where calling `IterableDatasetDict.map()` or `IterableDataset.map()` with the default `features=None` argument would overwrite the existing `info.features` attribute with `None`. This, in turn, caused the resulting dataset to lose its schema, breaking downstream usage of attributes like `column_names`. Why Previously, the code would always set `info.features = features`, even if `features` was `None`. When mapping with removal of columns or other transformations, this led to the destruction of the schema and caused failures in code that relied on the dataset schema being present. How We now update `info.features` only if `features` is not `None`. This preserves the original schema unless the user explicitly provides a new one. Reference Fixes #7568
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7658/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7658/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7658.diff", "html_url": "https://github.com/huggingface/datasets/pull/7658", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7658.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7658" }
https://api.github.com/repos/huggingface/datasets/issues/7657
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7657/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7657/comments
https://api.github.com/repos/huggingface/datasets/issues/7657/events
https://github.com/huggingface/datasets/pull/7657
3,186,036,016
PR_kwDODunzps6cks2E
7,657
feat: add subset_name as alias for name in load_dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-29T10:39:00Z
2025-06-29T10:55:11Z
null
CONTRIBUTOR
null
null
null
fixes #7637 This PR introduces subset_name as a user-facing alias for the name (previously `config_name`) argument in load_dataset. It aligns terminology with the Hugging Face Hub UI (which shows “Subset”), reducing confusion for new users. Supports `subset_name` in `load_dataset()` Adds `.subset_name` property to DatasetBuilder Maintains full backward compatibility Raises clear error if name and `subset_name` conflict
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7657/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7657/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7657.diff", "html_url": "https://github.com/huggingface/datasets/pull/7657", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7657.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7657" }
https://api.github.com/repos/huggingface/datasets/issues/7656
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7656/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7656/comments
https://api.github.com/repos/huggingface/datasets/issues/7656/events
https://github.com/huggingface/datasets/pull/7656
3,185,865,686
PR_kwDODunzps6ckPHc
7,656
fix(iterable): ensure MappedExamplesIterable supports state_dict for resume
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-29T07:50:13Z
2025-06-29T07:50:13Z
null
CONTRIBUTOR
null
null
null
Fixes #7630 ### Problem When calling `.map()` on an `IterableDataset`, resuming from a checkpoint skips a large number of samples. This is because `MappedExamplesIterable` did not implement `state_dict()` or `load_state_dict()`, so checkpointing was not properly delegated to the underlying iterable. ### What This PR Does This patch adds: ```python def state_dict(self): return self.ex_iterable.state_dict() def load_state_dict(self, state): self.ex_iterable.load_state_dict(state) ``` to MappedExamplesIterable, so the wrapped base iterable's state can be saved and restored as expected. Result Using .map() no longer causes sample skipping after checkpoint resume. Let me know if a dedicated test case is required — happy to add one!
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7656/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7656/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7656.diff", "html_url": "https://github.com/huggingface/datasets/pull/7656", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7656.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7656" }
https://api.github.com/repos/huggingface/datasets/issues/7655
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7655/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7655/comments
https://api.github.com/repos/huggingface/datasets/issues/7655/events
https://github.com/huggingface/datasets/pull/7655
3,185,382,105
PR_kwDODunzps6ci9oi
7,655
Added specific use cases in Improve Performace
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-28T19:00:32Z
2025-06-28T19:00:32Z
null
CONTRIBUTOR
null
null
null
Fixes #2494
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7655/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7655/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7655.diff", "html_url": "https://github.com/huggingface/datasets/pull/7655", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7655.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7655" }
https://api.github.com/repos/huggingface/datasets/issues/7654
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7654/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7654/comments
https://api.github.com/repos/huggingface/datasets/issues/7654/events
https://github.com/huggingface/datasets/pull/7654
3,184,770,992
PR_kwDODunzps6chPmz
7,654
fix(load): strip deprecated use_auth_token from config_kwargs
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-28T09:20:21Z
2025-06-28T09:20:21Z
null
CONTRIBUTOR
null
null
null
Fixes #7504 This PR resolves a compatibility error when loading datasets via `load_dataset()` using outdated arguments like `use_auth_token`. **What was happening:** Users passing `use_auth_token` in `load_dataset(..., use_auth_token=...)` encountered a `ValueError`: BuilderConfig ParquetConfig(...) doesn't have a 'use_auth_token' key. **Why:** `use_auth_token` has been deprecated and removed from config definitions (replaced by `token`), but the `load_dataset()` function still forwarded it via `**config_kwargs` to BuilderConfigs, leading to unrecognized key errors. **Fix:** We now intercept and strip `use_auth_token` from `config_kwargs` inside `load_dataset`, replacing it with a warning: ```python if "use_auth_token" in config_kwargs: logger.warning("The 'use_auth_token' argument is deprecated. Please use 'token' instead.") config_kwargs.pop("use_auth_token") ``` This ensures legacy compatibility while guiding users to switch to the token argument. Let me know if you'd prefer a deprecation error instead of a warning. Thanks!
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7654/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7654/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7654.diff", "html_url": "https://github.com/huggingface/datasets/pull/7654", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7654.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7654" }
https://api.github.com/repos/huggingface/datasets/issues/7653
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7653/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7653/comments
https://api.github.com/repos/huggingface/datasets/issues/7653/events
https://github.com/huggingface/datasets/pull/7653
3,184,746,093
PR_kwDODunzps6chLmb
7,653
feat(load): fallback to `load_from_disk()` when loading a saved dataset directory
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-28T08:47:36Z
2025-06-28T08:47:36Z
null
CONTRIBUTOR
null
null
null
### Related Issue Fixes #7503 Partially addresses #5044 by allowing `load_dataset()` to auto-detect and gracefully delegate to `load_from_disk()` for locally saved datasets. --- ### What does this PR do? This PR introduces a minimal fallback mechanism in `load_dataset()` that detects when the provided `path` points to a dataset saved using `save_to_disk()`, and automatically redirects to `load_from_disk()`. #### 🐛 Before (unexpected metadata-only rows): ```python ds = load_dataset("/path/to/saved_dataset") # → returns rows with only internal metadata (_data_files, _fingerprint, etc.) ```` #### ✅ After (graceful fallback): ```python ds = load_dataset("/path/to/saved_dataset") # → logs a warning and internally switches to load_from_disk() ``` --- ### Why is this useful? * Prevents confusion when reloading local datasets saved via `save_to_disk()`. * Enables smoother compatibility with frameworks (e.g., TRL, `lighteval`) that rely on `load_dataset()` calls. * Fully backward-compatible — hub-based loading, custom builders, and streaming remain untouched.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7653/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7653/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7653.diff", "html_url": "https://github.com/huggingface/datasets/pull/7653", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7653.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7653" }
https://api.github.com/repos/huggingface/datasets/issues/7652
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7652/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7652/comments
https://api.github.com/repos/huggingface/datasets/issues/7652/events
https://github.com/huggingface/datasets/pull/7652
3,183,372,055
PR_kwDODunzps6cdCnv
7,652
Add columns support to JSON loader for selective key filtering
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
2
2025-06-27T16:18:42Z
2025-07-03T09:52:48Z
null
CONTRIBUTOR
null
null
null
Fixes #7594 This PR adds support for filtering specific columns when loading datasets from .json or .jsonl files — similar to how the columns=... argument works for Parquet. As suggested, support for the `columns=...` argument (previously available for Parquet) has now been extended to **JSON and JSONL** loading via `load_dataset(...)`. You can now load only specific keys/columns and skip the rest — which should help in cases where some fields are unclean, inconsistent, or just unnecessary. ### Example: ```python from datasets import load_dataset dataset = load_dataset("json", data_files="your_data.jsonl", columns=["id", "title"]) print(dataset["train"].column_names) # Output: ['id', 'title'] ``` ### Summary of changes: * Added `columns: Optional[List[str]]` to `JsonConfig` * Updated `_generate_tables()` to filter selected columns * Forwarded `columns` argument from `load_dataset()` to the config * Added test for validation(should be fine!) Let me know if you'd like the same to be added for CSV or others as a follow-up — happy to help.
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7652/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7652/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7652.diff", "html_url": "https://github.com/huggingface/datasets/pull/7652", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7652.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7652" }
https://api.github.com/repos/huggingface/datasets/issues/7651
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7651/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7651/comments
https://api.github.com/repos/huggingface/datasets/issues/7651/events
https://github.com/huggingface/datasets/pull/7651
3,182,792,775
PR_kwDODunzps6cbMmg
7,651
fix: Extended metadata file names for folder_based_builder
{ "avatar_url": "https://avatars.githubusercontent.com/u/6965756?v=4", "events_url": "https://api.github.com/users/iPieter/events{/privacy}", "followers_url": "https://api.github.com/users/iPieter/followers", "following_url": "https://api.github.com/users/iPieter/following{/other_user}", "gists_url": "https://api.github.com/users/iPieter/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/iPieter", "id": 6965756, "login": "iPieter", "node_id": "MDQ6VXNlcjY5NjU3NTY=", "organizations_url": "https://api.github.com/users/iPieter/orgs", "received_events_url": "https://api.github.com/users/iPieter/received_events", "repos_url": "https://api.github.com/users/iPieter/repos", "site_admin": false, "starred_url": "https://api.github.com/users/iPieter/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iPieter/subscriptions", "type": "User", "url": "https://api.github.com/users/iPieter", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-27T13:12:11Z
2025-06-30T08:19:37Z
null
NONE
null
null
null
Fixes #7650. The metadata files generated by the `DatasetDict.save_to_file` function are not included in the folder_based_builder's metadata list, causing issues when only 1 actual data file is present, as described in issue #7650. This PR adds these filenames to the builder, allowing correct loading.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7651/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7651/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7651.diff", "html_url": "https://github.com/huggingface/datasets/pull/7651", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7651.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7651" }
https://api.github.com/repos/huggingface/datasets/issues/7650
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7650/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7650/comments
https://api.github.com/repos/huggingface/datasets/issues/7650/events
https://github.com/huggingface/datasets/issues/7650
3,182,745,315
I_kwDODunzps69tNbj
7,650
`load_dataset` defaults to json file format for datasets with 1 shard
{ "avatar_url": "https://avatars.githubusercontent.com/u/6965756?v=4", "events_url": "https://api.github.com/users/iPieter/events{/privacy}", "followers_url": "https://api.github.com/users/iPieter/followers", "following_url": "https://api.github.com/users/iPieter/following{/other_user}", "gists_url": "https://api.github.com/users/iPieter/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/iPieter", "id": 6965756, "login": "iPieter", "node_id": "MDQ6VXNlcjY5NjU3NTY=", "organizations_url": "https://api.github.com/users/iPieter/orgs", "received_events_url": "https://api.github.com/users/iPieter/received_events", "repos_url": "https://api.github.com/users/iPieter/repos", "site_admin": false, "starred_url": "https://api.github.com/users/iPieter/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iPieter/subscriptions", "type": "User", "url": "https://api.github.com/users/iPieter", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-27T12:54:25Z
2025-06-27T12:54:25Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug I currently have multiple datasets (train+validation) saved as 50MB shards. For one dataset the validation pair is small enough to fit into a single shard and this apparently causes problems when loading the dataset. I created the datasets using a DatasetDict, saved them as 50MB arrow files for streaming and then load each dataset. I have no problem loading any of the other datasets with more than 1 arrow file/shard. The error indicates the training set got loaded in arrow format (correct) and the validation set in json (incorrect). This seems to be because some of the metadata files are considered as dataset files. ``` Error loading /nfs/dataset_pt-uk: Couldn't infer the same data file format for all splits. Got {NamedSplit('train'): ('arrow', {}), NamedSplit('validation'): ('json', {})} ``` ![Image](https://github.com/user-attachments/assets/f6e7596a-dd53-46a9-9a23-4e9cac2ac049) Concretely, there is a mismatch between the metadata created by the `DatasetDict.save_to_file` and the builder for `datasets.load_dataset`: https://github.com/huggingface/datasets/blob/e71b0b19d79c7531f9b9bea7c09916b5f6157f42/src/datasets/data_files.py#L107 The `folder_based_builder` lists all files and with 1 arrow file the json files (that are actually metadata) are in the majority. https://github.com/huggingface/datasets/blob/e71b0b19d79c7531f9b9bea7c09916b5f6157f42/src/datasets/packaged_modules/folder_based_builder/folder_based_builder.py#L58 ### Steps to reproduce the bug Create a dataset with metadata and 1 arrow file in validation and multiple arrow files in the training set, following the above description. In my case, I saved the files via: ```python dataset = DatasetDict({ 'train': train_dataset, 'validation': val_dataset }) dataset.save_to_disk(output_path, max_shard_size="50MB") ``` ### Expected behavior The dataset would get loaded. ### Environment info - `datasets` version: 3.6.0 - Platform: Linux-6.14.0-22-generic-x86_64-with-glibc2.41 - Python version: 3.12.7 - `huggingface_hub` version: 0.31.1 - PyArrow version: 18.1.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.6.1
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7650/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7650/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7649
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7649/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7649/comments
https://api.github.com/repos/huggingface/datasets/issues/7649/events
https://github.com/huggingface/datasets/pull/7649
3,181,481,444
PR_kwDODunzps6cW0sQ
7,649
Enable parallel shard upload in push_to_hub() using num_proc
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-27T05:59:03Z
2025-06-27T06:03:46Z
null
CONTRIBUTOR
null
null
null
Fixes #7591 ### Add num_proc support to `push_to_hub()` for parallel shard upload This PR adds support for parallel upload of dataset shards via the `num_proc` argument in `Dataset.push_to_hub()`. 📌 While the `num_proc` parameter was already present in the `push_to_hub()` signature and correctly passed to `_push_parquet_shards_to_hub()`, it was not being used to parallelize the upload. 🔧 This PR updates the internal `_push_parquet_shards_to_hub()` function to: - Use `multiprocessing.Pool` and `iflatmap_unordered()` for concurrent shard upload when `num_proc > 1` - Preserve original serial upload behavior if `num_proc` is `None` or ≤ 1 - Keep tqdm progress and commit behavior unchanged Let me know if any test coverage or further changes are needed!
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7649/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7649/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7649.diff", "html_url": "https://github.com/huggingface/datasets/pull/7649", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7649.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7649" }
https://api.github.com/repos/huggingface/datasets/issues/7648
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7648/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7648/comments
https://api.github.com/repos/huggingface/datasets/issues/7648/events
https://github.com/huggingface/datasets/pull/7648
3,181,409,736
PR_kwDODunzps6cWmSn
7,648
Fix misleading add_column() usage example in docstring
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-27T05:27:04Z
2025-06-27T05:27:54Z
null
CONTRIBUTOR
null
null
null
Fixes #7611 This PR fixes the usage example in the Dataset.add_column() docstring, which previously implied that add_column() modifies the dataset in-place. Why: The method returns a new dataset with the additional column, and users must assign the result to a variable to preserve the change. This should make the behavior clearer for users. @lhoestq @davanstrien
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7648/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7648/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7648.diff", "html_url": "https://github.com/huggingface/datasets/pull/7648", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7648.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7648" }
https://api.github.com/repos/huggingface/datasets/issues/7647
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7647/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7647/comments
https://api.github.com/repos/huggingface/datasets/issues/7647/events
https://github.com/huggingface/datasets/issues/7647
3,178,952,517
I_kwDODunzps69evdF
7,647
loading mozilla-foundation--common_voice_11_0 fails
{ "avatar_url": "https://avatars.githubusercontent.com/u/5703039?v=4", "events_url": "https://api.github.com/users/pavel-esir/events{/privacy}", "followers_url": "https://api.github.com/users/pavel-esir/followers", "following_url": "https://api.github.com/users/pavel-esir/following{/other_user}", "gists_url": "https://api.github.com/users/pavel-esir/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/pavel-esir", "id": 5703039, "login": "pavel-esir", "node_id": "MDQ6VXNlcjU3MDMwMzk=", "organizations_url": "https://api.github.com/users/pavel-esir/orgs", "received_events_url": "https://api.github.com/users/pavel-esir/received_events", "repos_url": "https://api.github.com/users/pavel-esir/repos", "site_admin": false, "starred_url": "https://api.github.com/users/pavel-esir/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pavel-esir/subscriptions", "type": "User", "url": "https://api.github.com/users/pavel-esir", "user_view_type": "public" }
[]
open
false
null
[]
null
1
2025-06-26T12:23:48Z
2025-06-27T12:29:03Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug Hello everyone, i am trying to load `mozilla-foundation--common_voice_11_0` and it fails. Reproducer ``` import datasets datasets.load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test", streaming=True, trust_remote_code=True) ``` and it fails with ``` File ~/opt/envs/.../lib/python3.10/site-packages/datasets/utils/file_utils.py:827, in _add_retries_to_file_obj_read_method.<locals>.read_with_retries(*args, **kwargs) 825 for retry in range(1, max_retries + 1): 826 try: --> 827 out = read(*args, **kwargs) 828 break 829 except ( 830 _AiohttpClientError, 831 asyncio.TimeoutError, 832 requests.exceptions.ConnectionError, 833 requests.exceptions.Timeout, 834 ) as err: File /usr/lib/python3.10/codecs.py:322, in BufferedIncrementalDecoder.decode(self, input, final) 319 def decode(self, input, final=False): 320 # decode input (taking the buffer into account) 321 data = self.buffer + input --> 322 (result, consumed) = self._buffer_decode(data, self.errors, final) 323 # keep undecoded input until the next call 324 self.buffer = data[consumed:] UnicodeDecodeError: 'utf-8' codec can't decode byte 0x8b in position 1: invalid start byte ``` When i remove streaming then everything is good but i need `streaming=True` ### Steps to reproduce the bug ``` import datasets datasets.load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test", streaming=True, trust_remote_code=True) ``` ### Expected behavior Expected that it will download dataset ### Environment info datasets==3.6.0 python3.10 on all platforms linux/win/mac
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7647/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7647/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7646
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7646/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7646/comments
https://api.github.com/repos/huggingface/datasets/issues/7646/events
https://github.com/huggingface/datasets/pull/7646
3,178,036,854
PR_kwDODunzps6cLhrM
7,646
Introduces automatic subset-level grouping for folder-based dataset builders #7066
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
3
2025-06-26T07:01:37Z
2025-06-27T18:04:04Z
null
CONTRIBUTOR
null
null
null
Fixes #7066 This PR introduces automatic **subset-level grouping** for folder-based dataset builders by: 1. Adding a utility function `group_files_by_subset()` that clusters files by root name (ignoring digits and shard suffixes). 2. Integrating this logic into `FolderBasedBuilder._split_generators()` to yield one split per subset. 3. Adding unit tests for the grouping function. 4. Updating the documentation to describe this new behavior under `docs/source/repository_structure.mdx`. --- ### Motivation Datasets with files like: ``` train0.jsonl train1.jsonl animals.jsonl metadata.jsonl ``` will now be **automatically grouped** as: - `"train"` subset → `train0.jsonl`, `train1.jsonl` - `"animals"` subset → `animals.jsonl` - `"metadata"` subset → `metadata.jsonl` This enables structured multi-subset loading even when the dataset doesn't follow traditional `train/validation/test` split conventions. --- ### Files Changed - `src/datasets/data_files.py`: added `group_files_by_subset()` utility - `src/datasets/packaged_modules/folder_based_builder/folder_based_builder.py`: grouped files before yielding splits - `tests/test_data_files.py`: added unit test `test_group_files_by_subset` - `docs/source/repository_structure.mdx`: documented subset grouping for maintainers and users --- ### Benefits - More flexible and robust dataset split logic - Enables logical grouping of user-uploaded files without nested folder structure - Backward-compatible with all existing folder-based configs --- Ready for review ✅
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7646/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7646/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7646.diff", "html_url": "https://github.com/huggingface/datasets/pull/7646", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7646.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7646" }
https://api.github.com/repos/huggingface/datasets/issues/7645
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7645/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7645/comments
https://api.github.com/repos/huggingface/datasets/issues/7645/events
https://github.com/huggingface/datasets/pull/7645
3,176,810,164
PR_kwDODunzps6cHkp-
7,645
`ClassLabel` docs: Correct value for unknown labels
{ "avatar_url": "https://avatars.githubusercontent.com/u/56924246?v=4", "events_url": "https://api.github.com/users/l-uuz/events{/privacy}", "followers_url": "https://api.github.com/users/l-uuz/followers", "following_url": "https://api.github.com/users/l-uuz/following{/other_user}", "gists_url": "https://api.github.com/users/l-uuz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/l-uuz", "id": 56924246, "login": "l-uuz", "node_id": "MDQ6VXNlcjU2OTI0MjQ2", "organizations_url": "https://api.github.com/users/l-uuz/orgs", "received_events_url": "https://api.github.com/users/l-uuz/received_events", "repos_url": "https://api.github.com/users/l-uuz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/l-uuz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/l-uuz/subscriptions", "type": "User", "url": "https://api.github.com/users/l-uuz", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-25T20:01:35Z
2025-06-25T20:01:35Z
null
NONE
null
null
null
This small change fixes the documentation to to be compliant with what happens in `encode_example`. https://github.com/huggingface/datasets/blob/e71b0b19d79c7531f9b9bea7c09916b5f6157f42/src/datasets/features/features.py#L1126-L1129
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7645/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7645/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7645.diff", "html_url": "https://github.com/huggingface/datasets/pull/7645", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7645.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7645" }
https://api.github.com/repos/huggingface/datasets/issues/7644
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7644/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7644/comments
https://api.github.com/repos/huggingface/datasets/issues/7644/events
https://github.com/huggingface/datasets/pull/7644
3,176,363,492
PR_kwDODunzps6cGGfW
7,644
fix sequence ci
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-25T17:07:55Z
2025-06-25T17:10:30Z
2025-06-25T17:08:01Z
MEMBER
null
null
null
fix error from https://github.com/huggingface/datasets/pull/7643
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7644/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7644/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7644.diff", "html_url": "https://github.com/huggingface/datasets/pull/7644", "merged_at": "2025-06-25T17:08:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/7644.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7644" }
https://api.github.com/repos/huggingface/datasets/issues/7643
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7643/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7643/comments
https://api.github.com/repos/huggingface/datasets/issues/7643/events
https://github.com/huggingface/datasets/pull/7643
3,176,354,431
PR_kwDODunzps6cGEeK
7,643
Backward compat sequence instance
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-25T17:05:09Z
2025-06-25T17:07:40Z
2025-06-25T17:05:44Z
MEMBER
null
null
null
useful to still get `isinstance(Sequence(Value("int64")), Sequence)`for downstream libs like evaluate
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7643/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7643/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7643.diff", "html_url": "https://github.com/huggingface/datasets/pull/7643", "merged_at": "2025-06-25T17:05:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/7643.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7643" }
https://api.github.com/repos/huggingface/datasets/issues/7642
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7642/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7642/comments
https://api.github.com/repos/huggingface/datasets/issues/7642/events
https://github.com/huggingface/datasets/pull/7642
3,176,025,890
PR_kwDODunzps6cE_Wr
7,642
fix length for ci
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
0
2025-06-25T15:10:38Z
2025-06-25T15:11:53Z
2025-06-25T15:11:51Z
MEMBER
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7642/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7642/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7642.diff", "html_url": "https://github.com/huggingface/datasets/pull/7642", "merged_at": "2025-06-25T15:11:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/7642.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7642" }
https://api.github.com/repos/huggingface/datasets/issues/7641
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7641/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7641/comments
https://api.github.com/repos/huggingface/datasets/issues/7641/events
https://github.com/huggingface/datasets/pull/7641
3,175,953,405
PR_kwDODunzps6cEwUl
7,641
update docs and docstrings
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-25T14:48:58Z
2025-06-25T14:51:46Z
2025-06-25T14:49:33Z
MEMBER
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7641/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7641/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7641.diff", "html_url": "https://github.com/huggingface/datasets/pull/7641", "merged_at": "2025-06-25T14:49:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/7641.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7641" }
https://api.github.com/repos/huggingface/datasets/issues/7640
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7640/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7640/comments
https://api.github.com/repos/huggingface/datasets/issues/7640/events
https://github.com/huggingface/datasets/pull/7640
3,175,914,924
PR_kwDODunzps6cEofU
7,640
better features repr
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-25T14:37:32Z
2025-06-25T14:46:47Z
2025-06-25T14:46:45Z
MEMBER
null
null
null
following the addition of List in #7634 before: ```python In [3]: ds.features Out[3]: {'json': {'id': Value(dtype='string', id=None), 'metadata:transcript': [{'end': Value(dtype='float64', id=None), 'start': Value(dtype='float64', id=None), 'transcript': Value(dtype='string', id=None), 'words': [{'end': Value(dtype='float64', id=None), 'score': Value(dtype='float64', id=None), 'start': Value(dtype='float64', id=None), 'word': Value(dtype='string', id=None)}]}], 'metadata:vad': [{'end': Value(dtype='float64', id=None), 'start': Value(dtype='float64', id=None)}]}, 'mp4': Value(dtype='binary', id=None), 'npz': {'boxes_and_keypoints:box': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'boxes_and_keypoints:is_valid_box': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'boxes_and_keypoints:keypoints': Sequence(feature=Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), length=-1, id=None), 'movement:EmotionArousalToken': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:EmotionValenceToken': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:FAUToken': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:FAUValue': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:alignment_head_rotation': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:alignment_translation': Sequence(feature=Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), length=-1, id=None), 'movement:emotion_arousal': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:emotion_scores': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:emotion_valence': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:expression': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:frame_latent': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:gaze_encodings': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:head_encodings': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:hypernet_features': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'movement:is_valid': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'smplh:body_pose': Sequence(feature=Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), length=-1, id=None), 'smplh:global_orient': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), 'smplh:is_valid': Sequence(feature=Value(dtype='bool', id=None), length=-1, id=None), 'smplh:left_hand_pose': Sequence(feature=Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), length=-1, id=None), 'smplh:right_hand_pose': Sequence(feature=Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None), length=-1, id=None), 'smplh:translation': Sequence(feature=Sequence(feature=Value(dtype='float32', id=None), length=-1, id=None), length=-1, id=None)}, 'wav': Audio(sampling_rate=None, mono=True, decode=True, id=None), '__key__': Value(dtype='string', id=None), '__url__': Value(dtype='string', id=None)} ``` after: ```python In [3]: ds.features Out[3]: {'json': {'id': Value('string'), 'metadata:transcript': List({'end': Value('float64'), 'start': Value('float64'), 'transcript': Value('string'), 'words': List({'end': Value('float64'), 'score': Value('float64'), 'start': Value('float64'), 'word': Value('string')})}), 'metadata:vad': List({'end': Value('float64'), 'start': Value('float64')})}, 'mp4': Value('binary'), 'npz': {'boxes_and_keypoints:box': List(List(Value('float32'))), 'boxes_and_keypoints:is_valid_box': List(Value('bool')), 'boxes_and_keypoints:keypoints': List(List(List(Value('float32')))), 'movement:EmotionArousalToken': List(List(Value('float32'))), 'movement:EmotionValenceToken': List(List(Value('float32'))), 'movement:FAUToken': List(List(Value('float32'))), 'movement:FAUValue': List(List(Value('float32'))), 'movement:alignment_head_rotation': List(List(Value('float32'))), 'movement:alignment_translation': List(List(List(Value('float32')))), 'movement:emotion_arousal': List(List(Value('float32'))), 'movement:emotion_scores': List(List(Value('float32'))), 'movement:emotion_valence': List(List(Value('float32'))), 'movement:expression': List(List(Value('float32'))), 'movement:frame_latent': List(List(Value('float32'))), 'movement:gaze_encodings': List(List(Value('float32'))), 'movement:head_encodings': List(List(Value('float32'))), 'movement:hypernet_features': List(List(Value('float32'))), 'movement:is_valid': List(List(Value('float32'))), 'smplh:body_pose': List(List(List(Value('float32')))), 'smplh:global_orient': List(List(Value('float32'))), 'smplh:is_valid': List(Value('bool')), 'smplh:left_hand_pose': List(List(List(Value('float32')))), 'smplh:right_hand_pose': List(List(List(Value('float32')))), 'smplh:translation': List(List(Value('float32')))}, 'wav': Audio(sampling_rate=None, decode=True, stream_index=None), '__key__': Value('string'), '__url__': Value('string')} ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7640/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7640/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7640.diff", "html_url": "https://github.com/huggingface/datasets/pull/7640", "merged_at": "2025-06-25T14:46:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/7640.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7640" }
https://api.github.com/repos/huggingface/datasets/issues/7639
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7639/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7639/comments
https://api.github.com/repos/huggingface/datasets/issues/7639/events
https://github.com/huggingface/datasets/pull/7639
3,175,616,169
PR_kwDODunzps6cDoAf
7,639
fix save_infos
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-25T13:16:26Z
2025-06-25T13:19:33Z
2025-06-25T13:16:33Z
MEMBER
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7639/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7639/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7639.diff", "html_url": "https://github.com/huggingface/datasets/pull/7639", "merged_at": "2025-06-25T13:16:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/7639.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7639" }
https://api.github.com/repos/huggingface/datasets/issues/7638
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7638/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7638/comments
https://api.github.com/repos/huggingface/datasets/issues/7638/events
https://github.com/huggingface/datasets/pull/7638
3,172,645,391
PR_kwDODunzps6b5vpZ
7,638
Add ignore_decode_errors option to Image feature for robust decoding #7612
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
3
2025-06-24T16:47:51Z
2025-07-03T16:37:38Z
null
CONTRIBUTOR
null
null
null
This PR implements support for robust image decoding in the `Image` feature, as discussed in issue #7612. ## 🔧 What was added - A new boolean field: `ignore_decode_errors` (default: `False`) - If set to `True`, any exceptions during decoding will be caught, and `None` will be returned instead of raising an error ```python features = Features({ "image": Image(decode=True, ignore_decode_errors=True), }) ```` This enables robust iteration over potentially corrupted datasets — especially useful when streaming datasets like WebDataset or image-heavy public sets where sample corruption is common. ## 🧪 Behavior * If `ignore_decode_errors=False` (default), decoding behaves exactly as before * If `True`, decoding errors are caught, and a warning is emitted: ``` [Image.decode_example] Skipped corrupted image: ... ``` ## 🧵 Linked issue Closes #7612 Let me know if you'd like a follow-up test PR. Happy to write one!
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7638/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7638/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7638.diff", "html_url": "https://github.com/huggingface/datasets/pull/7638", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7638.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7638" }
https://api.github.com/repos/huggingface/datasets/issues/7637
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7637/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7637/comments
https://api.github.com/repos/huggingface/datasets/issues/7637/events
https://github.com/huggingface/datasets/issues/7637
3,171,883,522
I_kwDODunzps69DxoC
7,637
Introduce subset_name as an alias of config_name
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
4
2025-06-24T12:49:01Z
2025-07-01T16:08:33Z
null
MEMBER
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Feature request Add support for `subset_name` as an alias for `config_name` in the datasets library and related tools (such as loading scripts, documentation, and metadata). ### Motivation The Hugging Face Hub dataset viewer displays a column named **"Subset"**, which refers to what is currently technically called config_name in the datasets library. This inconsistency has caused confusion for many users, especially those unfamiliar with the internal terminology. I have repeatedly received questions from users trying to understand what "config" means, and why it doesn’t match what they see as "subset" on the Hub. Renaming everything to `subset_name` might be too disruptive, but introducing subset_name as a clear alias for config_name could significantly improve user experience without breaking backward compatibility. This change would: - Align terminology across the Hub UI and datasets codebase - Reduce user confusion, especially for newcomers - Make documentation and examples more intuitive
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/7637/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7637/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7636
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7636/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7636/comments
https://api.github.com/repos/huggingface/datasets/issues/7636/events
https://github.com/huggingface/datasets/issues/7636
3,170,878,167
I_kwDODunzps68_8LX
7,636
"open" in globals()["__builtins__"], an error occurs: "TypeError: argument of type 'module' is not iterable"
{ "avatar_url": "https://avatars.githubusercontent.com/u/51187979?v=4", "events_url": "https://api.github.com/users/kuanyan9527/events{/privacy}", "followers_url": "https://api.github.com/users/kuanyan9527/followers", "following_url": "https://api.github.com/users/kuanyan9527/following{/other_user}", "gists_url": "https://api.github.com/users/kuanyan9527/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kuanyan9527", "id": 51187979, "login": "kuanyan9527", "node_id": "MDQ6VXNlcjUxMTg3OTc5", "organizations_url": "https://api.github.com/users/kuanyan9527/orgs", "received_events_url": "https://api.github.com/users/kuanyan9527/received_events", "repos_url": "https://api.github.com/users/kuanyan9527/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kuanyan9527/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kuanyan9527/subscriptions", "type": "User", "url": "https://api.github.com/users/kuanyan9527", "user_view_type": "public" }
[]
closed
false
null
[]
null
3
2025-06-24T08:09:39Z
2025-07-01T01:54:08Z
2025-07-01T01:54:08Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
When I run the following code, an error occurs: "TypeError: argument of type 'module' is not iterable" ```python print("open" in globals()["__builtins__"]) ``` Traceback (most recent call last): File "./main.py", line 2, in <module> print("open" in globals()["__builtins__"]) ^^^^^^^^^^^^^^^^^^^^^^ TypeError: argument of type 'module' is not iterable But this code runs fine in datasets, I don't understand why [src/datasets/utils/patching.py#L96](https://github.com/huggingface/datasets/blob/3.6.0/src/datasets/utils/patching.py#L96)
{ "avatar_url": "https://avatars.githubusercontent.com/u/51187979?v=4", "events_url": "https://api.github.com/users/kuanyan9527/events{/privacy}", "followers_url": "https://api.github.com/users/kuanyan9527/followers", "following_url": "https://api.github.com/users/kuanyan9527/following{/other_user}", "gists_url": "https://api.github.com/users/kuanyan9527/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kuanyan9527", "id": 51187979, "login": "kuanyan9527", "node_id": "MDQ6VXNlcjUxMTg3OTc5", "organizations_url": "https://api.github.com/users/kuanyan9527/orgs", "received_events_url": "https://api.github.com/users/kuanyan9527/received_events", "repos_url": "https://api.github.com/users/kuanyan9527/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kuanyan9527/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kuanyan9527/subscriptions", "type": "User", "url": "https://api.github.com/users/kuanyan9527", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7636/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7636/timeline
null
completed
null
null
https://api.github.com/repos/huggingface/datasets/issues/7635
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7635/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7635/comments
https://api.github.com/repos/huggingface/datasets/issues/7635/events
https://github.com/huggingface/datasets/pull/7635
3,170,486,408
PR_kwDODunzps6bybOe
7,635
Fix: Preserve float columns in JSON loader when values are integer-like (e.g. 0.0, 1.0)
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-24T06:16:48Z
2025-06-24T06:16:48Z
null
CONTRIBUTOR
null
null
null
This PR fixes a bug in the JSON loader where columns containing float values like `[0.0, 1.0, 2.0]` were being implicitly coerced to `int`, due to pandas or Arrow type inference. This caused issues downstream in statistics computation (e.g., dataset-viewer) where such columns were incorrectly labeled as `"int"` instead of `"float"`. ### 🔍 What was happening: When the JSON loader falls back to `pandas_read_json()` (after `pa.read_json()` fails), pandas/Arrow can coerce float values to integers if all values are integer-like (e.g., `0.0 == 0`). ### ✅ What this PR does: - Adds a check in the fallback path of `_generate_tables()` - Ensures that columns made entirely of floats are preserved as `"float64"` even if they are integer-like (e.g. `0.0`, `1.0`) - This prevents loss of float semantics when creating the Arrow table ### 🧪 Reproducible Example: ```json [{"col": 0.0}, {"col": 1.0}, {"col": 2.0}] ```` Previously loaded as: * `int` Now correctly loaded as: * `float` Fixes #6937
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7635/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7635/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7635.diff", "html_url": "https://github.com/huggingface/datasets/pull/7635", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7635.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7635" }
https://api.github.com/repos/huggingface/datasets/issues/7634
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7634/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7634/comments
https://api.github.com/repos/huggingface/datasets/issues/7634/events
https://github.com/huggingface/datasets/pull/7634
3,169,389,653
PR_kwDODunzps6buyij
7,634
Replace Sequence by List
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-23T20:35:48Z
2025-06-25T13:59:13Z
2025-06-25T13:59:11Z
MEMBER
null
null
null
Sequence is just a utility that we need to keep for backward compatibility. And `[ ]` was used instead but doesn't allow passing the length of the list. This PR removes most mentions of Sequence and usage of `[ ]` and defines a proper List type instead. before: `Sequence(Value("int64"))` or `[Value("int64")]` now: `List(Value("int64"))` This PR conserves full backward compatibility. And it's a good occasion with the release of 4.0.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7634/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7634/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7634.diff", "html_url": "https://github.com/huggingface/datasets/pull/7634", "merged_at": "2025-06-25T13:59:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/7634.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7634" }
https://api.github.com/repos/huggingface/datasets/issues/7633
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7633/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7633/comments
https://api.github.com/repos/huggingface/datasets/issues/7633/events
https://github.com/huggingface/datasets/issues/7633
3,168,399,637
I_kwDODunzps682fEV
7,633
Proposal: Small Tamil Discourse Coherence Dataset.
{ "avatar_url": "https://avatars.githubusercontent.com/u/66418501?v=4", "events_url": "https://api.github.com/users/bikkiNitSrinagar/events{/privacy}", "followers_url": "https://api.github.com/users/bikkiNitSrinagar/followers", "following_url": "https://api.github.com/users/bikkiNitSrinagar/following{/other_user}", "gists_url": "https://api.github.com/users/bikkiNitSrinagar/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bikkiNitSrinagar", "id": 66418501, "login": "bikkiNitSrinagar", "node_id": "MDQ6VXNlcjY2NDE4NTAx", "organizations_url": "https://api.github.com/users/bikkiNitSrinagar/orgs", "received_events_url": "https://api.github.com/users/bikkiNitSrinagar/received_events", "repos_url": "https://api.github.com/users/bikkiNitSrinagar/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bikkiNitSrinagar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bikkiNitSrinagar/subscriptions", "type": "User", "url": "https://api.github.com/users/bikkiNitSrinagar", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-23T14:24:40Z
2025-06-23T14:24:40Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I’m a beginner from NIT Srinagar proposing a dataset of 50 Tamil text pairs for discourse coherence (coherent/incoherent labels) to support NLP research in low-resource languages. - Size: 50 samples - Format: CSV with columns (text1, text2, label) - Use case: Training NLP models for coherence I’ll use GitHub’s web editor and Google Colab. Please confirm if this fits.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7633/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7633/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7632
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7632/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7632/comments
https://api.github.com/repos/huggingface/datasets/issues/7632/events
https://github.com/huggingface/datasets/issues/7632
3,168,283,589
I_kwDODunzps682CvF
7,632
Graceful Error Handling for cast_column("image", Image(decode=True)) in Hugging Face Datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/37377515?v=4", "events_url": "https://api.github.com/users/ganiket19/events{/privacy}", "followers_url": "https://api.github.com/users/ganiket19/followers", "following_url": "https://api.github.com/users/ganiket19/following{/other_user}", "gists_url": "https://api.github.com/users/ganiket19/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ganiket19", "id": 37377515, "login": "ganiket19", "node_id": "MDQ6VXNlcjM3Mzc3NTE1", "organizations_url": "https://api.github.com/users/ganiket19/orgs", "received_events_url": "https://api.github.com/users/ganiket19/received_events", "repos_url": "https://api.github.com/users/ganiket19/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ganiket19/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ganiket19/subscriptions", "type": "User", "url": "https://api.github.com/users/ganiket19", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
0
2025-06-23T13:49:24Z
2025-06-23T16:26:53Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Feature request Currently, when using dataset.cast_column("image", Image(decode=True)), the pipeline throws an error and halts if any image in the dataset is invalid or corrupted (e.g., truncated files, incorrect formats, unreachable URLs). This behavior disrupts large-scale processing where a few faulty samples are common. reference : https://discuss.huggingface.co/t/handle-errors-when-loading-images-404-corrupted-etc/50318/5 https://discuss.huggingface.co/t/handling-non-existing-url-in-image-dataset-while-cast-column/69185 Proposed Feature Introduce a mechanism (e.g., a continue_on_error=True flag or global error handling mode) in Image(decode=True) that: Skips invalid images and sets them as None, or Logs the error but allows the rest of the dataset to be processed without interruption. Example Usage from datasets import load_dataset, Image dataset = load_dataset("my_dataset") dataset = dataset.cast_column("image", Image(decode=True, continue_on_error=True)) Benefits Ensures robust large-scale image dataset processing. Improves developer productivity by avoiding custom retry/error-handling code. Aligns with best practices in dataset preprocessing pipelines that tolerate minor data corruption. Potential Implementation Options Internally wrap the decoding in a try/except block. Return None or a placeholder on failure. Optionally allow custom error callbacks or logging. ### Motivation Robustness: Large-scale image datasets often contain a small fraction of corrupt files or unreachable URLs. Halting on the first error forces users to write custom workarounds or preprocess externally. Simplicity: A built-in flag removes boilerplate try/except logic around every decode step. Performance: Skipping invalid samples inline is more efficient than a two-pass approach (filter then decode). ### Your contribution 1. API Change Extend datasets.features.Image(decode=True) to accept continue_on_error: bool = False. 2. Behavior If continue_on_error=False (default), maintain current behavior: any decode error raises an exception. If continue_on_error=True, wrap decode logic in try/except: On success: store the decoded image. On failure: log a warning (e.g., via logging.warning) and set the field to None (or a sentinel value). 3. Optional Enhancements Allow a callback hook: Image(decode=True, continue_on_error=True, on_error=lambda idx, url, exc: ...) Emit metrics or counts of skipped images.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7632/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7632/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7631
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7631/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7631/comments
https://api.github.com/repos/huggingface/datasets/issues/7631/events
https://github.com/huggingface/datasets/pull/7631
3,165,127,657
PR_kwDODunzps6bgwOB
7,631
Pass user-agent from DownloadConfig into fsspec storage_options
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
1
2025-06-21T14:22:25Z
2025-06-21T14:25:28Z
null
CONTRIBUTOR
null
null
null
Fixes part of issue #6046 ### Problem The `user-agent` defined in `DownloadConfig` was not passed down to fsspec-based filesystems like `HfFileSystem`, which prevents proper identification/tracking of client requests. ### Solution Added support for injecting the `user-agent` into `storage_options["headers"]` within `_prepare_single_hop_path_and_storage_options()` based on the `protocol`. Now, when using `hf://`, `http://`, or `https://`, the custom user-agent is passed automatically. ### Code Location Modified: - `src/datasets/utils/file_utils.py` Used `get_datasets_user_agent(...)` to ensure proper formatting and fallback logic.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7631/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7631/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7631.diff", "html_url": "https://github.com/huggingface/datasets/pull/7631", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7631.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7631" }
https://api.github.com/repos/huggingface/datasets/issues/7630
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7630/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7630/comments
https://api.github.com/repos/huggingface/datasets/issues/7630/events
https://github.com/huggingface/datasets/issues/7630
3,164,650,900
I_kwDODunzps68oL2U
7,630
[bug] resume from ckpt skips samples if .map is applied
{ "avatar_url": "https://avatars.githubusercontent.com/u/23004953?v=4", "events_url": "https://api.github.com/users/felipemello1/events{/privacy}", "followers_url": "https://api.github.com/users/felipemello1/followers", "following_url": "https://api.github.com/users/felipemello1/following{/other_user}", "gists_url": "https://api.github.com/users/felipemello1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/felipemello1", "id": 23004953, "login": "felipemello1", "node_id": "MDQ6VXNlcjIzMDA0OTUz", "organizations_url": "https://api.github.com/users/felipemello1/orgs", "received_events_url": "https://api.github.com/users/felipemello1/received_events", "repos_url": "https://api.github.com/users/felipemello1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/felipemello1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/felipemello1/subscriptions", "type": "User", "url": "https://api.github.com/users/felipemello1", "user_view_type": "public" }
[]
open
false
null
[]
null
2
2025-06-21T01:50:03Z
2025-06-29T07:51:32Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
### Describe the bug resume from ckpt skips samples if .map is applied Maybe related: https://github.com/huggingface/datasets/issues/7538 ### Steps to reproduce the bug ```python from datasets import Dataset from datasets.distributed import split_dataset_by_node # Create dataset with map transformation def create_dataset(): ds = Dataset.from_dict({"id": list(range(100))}) ds = ds.to_iterable_dataset(num_shards=4) ds = ds.map(lambda x: x) #comment it out to get desired behavior ds = split_dataset_by_node(ds, rank=0, world_size=2) return ds ds = create_dataset() # Iterate and save checkpoint after 10 samples it = iter(ds) for idx, sample in enumerate(it): if idx == 9: # Checkpoint after 10 samples checkpoint = ds.state_dict() print(f"Checkpoint saved at sample: {sample['id']}") break # Continue with original iterator original_next_samples = [] for idx, sample in enumerate(it): original_next_samples.append(sample["id"]) if idx >= 4: break # Resume from checkpoint ds_new = create_dataset() ds_new.load_state_dict(checkpoint) # Get samples from resumed iterator it_new = iter(ds_new) resumed_next_samples = [] for idx, sample in enumerate(it_new): resumed_next_samples.append(sample["id"]) if idx >= 4: break print(f"\nExpected next samples: {original_next_samples}") print(f"Actual next samples: {resumed_next_samples}") print( f"\n❌ BUG: {resumed_next_samples[0] - original_next_samples[0]} samples were skipped!" ) ``` With map ``` Checkpoint saved at sample: 9 Expected next samples: [10, 11, 12, 13, 14] Actual next samples: [50, 51, 52, 53, 54] ❌ BUG: 40 samples were skipped! ``` ### Expected behavior without map ``` Expected next samples: [10, 11, 12, 13, 14] Actual next samples: [10, 11, 12, 13, 14] ❌ BUG: 0 samples were skipped! ``` ### Environment info datasets == 3.6.0
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7630/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7630/timeline
null
null
null
null
https://api.github.com/repos/huggingface/datasets/issues/7629
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7629/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7629/comments
https://api.github.com/repos/huggingface/datasets/issues/7629/events
https://github.com/huggingface/datasets/pull/7629
3,161,169,782
PR_kwDODunzps6bTc7b
7,629
Add test for `as_iterable_dataset()` method in DatasetBuilder
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-19T19:23:55Z
2025-06-19T19:23:55Z
null
CONTRIBUTOR
null
null
null
This PR adds a test for the new `as_iterable_dataset()` method introduced in PR #7628. The test: - Loads a builder using `load_dataset_builder("c4", "en")` - Runs `download_and_prepare()` - Streams examples using `builder.as_iterable_dataset(split="train[:100]")` - Verifies streamed examples contain the "text" field This ensures that the builder correctly streams data from cached Arrow files.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7629/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7629/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7629.diff", "html_url": "https://github.com/huggingface/datasets/pull/7629", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7629.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7629" }
https://api.github.com/repos/huggingface/datasets/issues/7628
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7628/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7628/comments
https://api.github.com/repos/huggingface/datasets/issues/7628/events
https://github.com/huggingface/datasets/pull/7628
3,161,156,461
PR_kwDODunzps6bTaGk
7,628
Add `as_iterable_dataset()` method to DatasetBuilder for streaming from cached Arrow files
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-19T19:15:41Z
2025-06-19T19:15:41Z
null
CONTRIBUTOR
null
null
null
This PR implements `builder.as_iterable_dataset(split=...)` as discussed in #5481. It allows users to load an `IterableDataset` directly from cached Arrow files (using ArrowReader and ArrowExamplesIterable), without loading the full dataset into memory. This is useful for large-scale training scenarios where memory is constrained. A test has also been added in `test_builder.py`. Related to: #5481
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7628/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7628/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7628.diff", "html_url": "https://github.com/huggingface/datasets/pull/7628", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7628.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7628" }
https://api.github.com/repos/huggingface/datasets/issues/7627
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7627/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7627/comments
https://api.github.com/repos/huggingface/datasets/issues/7627/events
https://github.com/huggingface/datasets/issues/7627
3,160,544,390
I_kwDODunzps68YhSG
7,627
Creating a HF Dataset from lakeFS with S3 storage takes too much time!
{ "avatar_url": "https://avatars.githubusercontent.com/u/118734142?v=4", "events_url": "https://api.github.com/users/Thunderhead-exe/events{/privacy}", "followers_url": "https://api.github.com/users/Thunderhead-exe/followers", "following_url": "https://api.github.com/users/Thunderhead-exe/following{/other_user}", "gists_url": "https://api.github.com/users/Thunderhead-exe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Thunderhead-exe", "id": 118734142, "login": "Thunderhead-exe", "node_id": "U_kgDOBxO9Pg", "organizations_url": "https://api.github.com/users/Thunderhead-exe/orgs", "received_events_url": "https://api.github.com/users/Thunderhead-exe/received_events", "repos_url": "https://api.github.com/users/Thunderhead-exe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Thunderhead-exe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Thunderhead-exe/subscriptions", "type": "User", "url": "https://api.github.com/users/Thunderhead-exe", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-19T14:28:41Z
2025-06-23T12:39:10Z
2025-06-23T12:39:10Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, I’m new to HF dataset and I tried to create datasets based on data versioned in **lakeFS** _(**MinIO** S3 bucket as storage backend)_ Here I’m using ±30000 PIL image from MNIST data however it is taking around 12min to execute, which is a lot! From what I understand, it is loading the images into cache then building the dataset. – Please find bellow the execution screenshot – Is there a way to optimize this or am I doing something wrong? Thanks! ![Image](https://github.com/user-attachments/assets/c79257c8-f023-42a9-9e6f-0898b3ea93fe)
{ "avatar_url": "https://avatars.githubusercontent.com/u/118734142?v=4", "events_url": "https://api.github.com/users/Thunderhead-exe/events{/privacy}", "followers_url": "https://api.github.com/users/Thunderhead-exe/followers", "following_url": "https://api.github.com/users/Thunderhead-exe/following{/other_user}", "gists_url": "https://api.github.com/users/Thunderhead-exe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Thunderhead-exe", "id": 118734142, "login": "Thunderhead-exe", "node_id": "U_kgDOBxO9Pg", "organizations_url": "https://api.github.com/users/Thunderhead-exe/orgs", "received_events_url": "https://api.github.com/users/Thunderhead-exe/received_events", "repos_url": "https://api.github.com/users/Thunderhead-exe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Thunderhead-exe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Thunderhead-exe/subscriptions", "type": "User", "url": "https://api.github.com/users/Thunderhead-exe", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7627/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7627/timeline
null
completed
null
null
https://api.github.com/repos/huggingface/datasets/issues/7626
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7626/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7626/comments
https://api.github.com/repos/huggingface/datasets/issues/7626/events
https://github.com/huggingface/datasets/pull/7626
3,159,322,138
PR_kwDODunzps6bNMuF
7,626
feat(map): reuse unchanged columns when input_columns specified to reduce disk usage (#6013)
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
0
2025-06-19T07:41:45Z
2025-06-26T06:43:16Z
null
CONTRIBUTOR
null
null
null
## Summary This PR addresses [#6013](https://github.com/huggingface/datasets/issues/6013) by reusing unchanged columns from the original dataset in the `map()` method when `input_columns` is specified. ## What’s Implemented - Injected logic at the end of `Dataset.map()` to: - Identify untouched columns not in `input_columns` or `remove_columns` - Select those columns from the original dataset - Concatenate them with the transformed result using `pyarrow.concat_tables` ## Example Behavior ```python ds = Dataset.from_dict({"a": [1, 2], "b": [3, 4]}) ds2 = ds.map(lambda x: {"c": x["a"] + 10}, input_columns=["a"], remove_columns=["a"]) print(ds2.column_names) # Output: ['b', 'c'] ```` Column `b` is reused from the original dataset. ## Notes * This keeps disk usage and caching minimal by avoiding full dataset duplication. * Only triggered when `input_columns` is set. --- cc @lhoestq @mariosasko for review 🙂
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7626/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7626/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7626.diff", "html_url": "https://github.com/huggingface/datasets/pull/7626", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7626.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7626" }
https://api.github.com/repos/huggingface/datasets/issues/7625
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7625/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7625/comments
https://api.github.com/repos/huggingface/datasets/issues/7625/events
https://github.com/huggingface/datasets/pull/7625
3,159,016,001
PR_kwDODunzps6bMKof
7,625
feat: Add h5folder dataset loader for HDF5 support
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
3
2025-06-19T05:39:10Z
2025-06-26T05:44:26Z
null
CONTRIBUTOR
null
null
null
### Related Issue Closes #3113 ### What does this PR do? This PR introduces a new dataset loader module called **`h5folder`** to support loading datasets stored in **HDF5 (.h5)** format. It allows users to do: ```python from datasets import load_dataset dataset = load_dataset("h5folder", data_dir="path/to/") ```` ### 🧩 Design Overview * Implemented inside `datasets/packaged_modules/h5folder/h5folder.py` * Based on the `GeneratorBasedBuilder` API * Uses `h5py` to read HDF5 files and yield examples * Expects datasets such as `id`, `data`, and `label` inside `data.h5` * Converts numpy arrays to Python types before yielding ### 🧪 Example `.h5` Structure (for local testing) ```python import h5py import numpy as np with h5py.File("data.h5", "w") as f: f.create_dataset("id", data=np.arange(100)) f.create_dataset("data", data=np.random.randn(100, 10)) f.create_dataset("label", data=np.random.randint(0, 2, size=100)) ``` ### ✅ Testing - The loader logic follows the structure of existing modules like `imagefolder` - Will rely on Hugging Face CI to validate integration - Manually testing planned once merged or during feedback ### 📁 Files Added * `datasets/src/datasets/packaged_modules/h5folder/h5folder.py` ### 📌 Component(s) Affected * `area/datasets` * `area/load` ### 📦 Release Note Classification * `rn/feature` – Adds support for loading `.h5` datasets via `load_dataset("h5folder", ...)` --- Let me know if any changes or improvements are needed — happy to iterate. Thanks for reviewing!
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/7625/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7625/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7625.diff", "html_url": "https://github.com/huggingface/datasets/pull/7625", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7625.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7625" }
https://api.github.com/repos/huggingface/datasets/issues/7624
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7624/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7624/comments
https://api.github.com/repos/huggingface/datasets/issues/7624/events
https://github.com/huggingface/datasets/issues/7624
3,156,136,624
I_kwDODunzps68HtKw
7,624
#Dataset Make "image" column appear first in dataset preview UI
{ "avatar_url": "https://avatars.githubusercontent.com/u/98875217?v=4", "events_url": "https://api.github.com/users/jcerveto/events{/privacy}", "followers_url": "https://api.github.com/users/jcerveto/followers", "following_url": "https://api.github.com/users/jcerveto/following{/other_user}", "gists_url": "https://api.github.com/users/jcerveto/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jcerveto", "id": 98875217, "login": "jcerveto", "node_id": "U_kgDOBeS3UQ", "organizations_url": "https://api.github.com/users/jcerveto/orgs", "received_events_url": "https://api.github.com/users/jcerveto/received_events", "repos_url": "https://api.github.com/users/jcerveto/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jcerveto/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jcerveto/subscriptions", "type": "User", "url": "https://api.github.com/users/jcerveto", "user_view_type": "public" }
[]
closed
false
null
[]
null
2
2025-06-18T09:25:19Z
2025-06-20T07:46:43Z
2025-06-20T07:46:43Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi! #Dataset I’m currently uploading a dataset that includes an `"image"` column (PNG files), along with some metadata columns. The dataset is loaded from a .jsonl file. My goal is to have the "image" column appear as the first column in the dataset card preview UI on the :hugs: Hub. However, at the moment, the `"image"` column is not the first—in fact, it appears last, which is not ideal for the presentation I’d like to achieve. I have a couple of questions: Is there a way to force the dataset card to display the `"image"` column first? Is there currently any way to control or influence the column order in the dataset preview UI? Does the order of keys in the .jsonl file or the features argument affect the display order? Thanks again for your time and help! :blush:
{ "avatar_url": "https://avatars.githubusercontent.com/u/98875217?v=4", "events_url": "https://api.github.com/users/jcerveto/events{/privacy}", "followers_url": "https://api.github.com/users/jcerveto/followers", "following_url": "https://api.github.com/users/jcerveto/following{/other_user}", "gists_url": "https://api.github.com/users/jcerveto/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jcerveto", "id": 98875217, "login": "jcerveto", "node_id": "U_kgDOBeS3UQ", "organizations_url": "https://api.github.com/users/jcerveto/orgs", "received_events_url": "https://api.github.com/users/jcerveto/received_events", "repos_url": "https://api.github.com/users/jcerveto/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jcerveto/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jcerveto/subscriptions", "type": "User", "url": "https://api.github.com/users/jcerveto", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7624/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7624/timeline
null
completed
null
null
https://api.github.com/repos/huggingface/datasets/issues/7623
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7623/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7623/comments
https://api.github.com/repos/huggingface/datasets/issues/7623/events
https://github.com/huggingface/datasets/pull/7623
3,154,519,684
PR_kwDODunzps6a9Jk5
7,623
fix: raise error in FolderBasedBuilder when data_dir and data_files are missing
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
closed
false
null
[]
null
2
2025-06-17T19:16:34Z
2025-06-18T14:18:41Z
2025-06-18T14:18:41Z
CONTRIBUTOR
null
null
null
### Related Issues/PRs Fixes #6152 --- ### What changes are proposed in this pull request? This PR adds a dedicated validation check in the `_info()` method of the `FolderBasedBuilder` class to ensure that users provide either `data_dir` or `data_files` when loading folder-based datasets (such as `audiofolder`, `imagefolder`, etc.). --- ### Why this change? Previously, when calling: ```python load_dataset("audiofolder") ```` without specifying `data_dir` or `data_files`, the loader would silently fallback to the **current working directory**, leading to: * Long loading times * Unexpected behavior (e.g., scanning unrelated files) This behavior was discussed in issue #6152. As suggested by maintainers, the fix has now been implemented directly inside the `FolderBasedBuilder._info()` method — keeping the logic localized to the specific builder instead of a generic loader function. --- ### How is this PR tested? * ✅ Manually tested by calling `load_dataset("audiofolder")` with no `data_dir` or `data_files` → a `ValueError` is now raised early. * ✅ Existing functionality (with valid input) remains unaffected. --- ### Does this PR require documentation update? * [x] No --- ### Release Notes #### Is this a user-facing change? * [x] Yes > Folder-based datasets now raise an explicit error if neither `data_dir` nor `data_files` are specified, preventing unintended fallback to the current working directory. --- #### What component(s) does this PR affect? * [x] `area/datasets` * [x] `area/load` --- <a name="release-note-category"></a> #### How should the PR be classified? * [x] `rn/bug-fix` - A user-facing bug fix --- #### Should this be included in the next patch release? * [x] Yes
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7623/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7623/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7623.diff", "html_url": "https://github.com/huggingface/datasets/pull/7623", "merged_at": "2025-06-18T14:18:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/7623.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7623" }
https://api.github.com/repos/huggingface/datasets/issues/7622
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7622/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7622/comments
https://api.github.com/repos/huggingface/datasets/issues/7622/events
https://github.com/huggingface/datasets/pull/7622
3,154,398,557
PR_kwDODunzps6a8v6J
7,622
Guard against duplicate builder_kwargs/config_kwargs in load_dataset_…
{ "avatar_url": "https://avatars.githubusercontent.com/u/149825575?v=4", "events_url": "https://api.github.com/users/Shohail-Ismail/events{/privacy}", "followers_url": "https://api.github.com/users/Shohail-Ismail/followers", "following_url": "https://api.github.com/users/Shohail-Ismail/following{/other_user}", "gists_url": "https://api.github.com/users/Shohail-Ismail/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Shohail-Ismail", "id": 149825575, "login": "Shohail-Ismail", "node_id": "U_kgDOCO4oJw", "organizations_url": "https://api.github.com/users/Shohail-Ismail/orgs", "received_events_url": "https://api.github.com/users/Shohail-Ismail/received_events", "repos_url": "https://api.github.com/users/Shohail-Ismail/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Shohail-Ismail/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Shohail-Ismail/subscriptions", "type": "User", "url": "https://api.github.com/users/Shohail-Ismail", "user_view_type": "public" }
[]
open
false
null
[]
null
1
2025-06-17T18:28:35Z
2025-07-02T12:39:20Z
null
NONE
null
null
null
…builder (#4910 ) ### What does this PR do? Fixes edge case in `load_dataset_builder` by raising a `TypeError` if the same key exists in both `builder_kwargs` and `config_kwargs`. ### Implementation details - Added a guard clause in `load_dataset_builder` to detect duplicate keys between `builder_kwargs` and `config_kwargs` - Wrote a unit test in `tests/test_load_duplicate_keys.py` to verify the exception is raised correctly ### Fixes Closes #4910 ### Reviewers @zach-huggingface @SunMarc Would appreciate your review if you have time - thanks!
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7622/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7622/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7622.diff", "html_url": "https://github.com/huggingface/datasets/pull/7622", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7622.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7622" }
https://api.github.com/repos/huggingface/datasets/issues/7621
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7621/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7621/comments
https://api.github.com/repos/huggingface/datasets/issues/7621/events
https://github.com/huggingface/datasets/pull/7621
3,153,780,963
PR_kwDODunzps6a6rAu
7,621
minor docs data aug
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-17T14:46:57Z
2025-06-17T14:50:28Z
2025-06-17T14:47:11Z
MEMBER
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7621/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7621/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7621.diff", "html_url": "https://github.com/huggingface/datasets/pull/7621", "merged_at": "2025-06-17T14:47:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/7621.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7621" }
https://api.github.com/repos/huggingface/datasets/issues/7620
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7620/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7620/comments
https://api.github.com/repos/huggingface/datasets/issues/7620/events
https://github.com/huggingface/datasets/pull/7620
3,153,565,183
PR_kwDODunzps6a58TP
7,620
Fixes in docs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
1
2025-06-17T13:41:54Z
2025-06-17T13:58:26Z
2025-06-17T13:58:24Z
MEMBER
null
null
null
before release 4.0 (I also did minor improvements to `features` to not show their `id=None` in their `__repr__()`)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7620/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7620/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/7620.diff", "html_url": "https://github.com/huggingface/datasets/pull/7620", "merged_at": "2025-06-17T13:58:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/7620.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7620" }
End of preview. Expand in Data Studio

Dataset Card for "github-issues-real"

More Information needed

Downloads last month
0