url
stringlengths 61
61
| repository_url
stringclasses 1
value | labels_url
stringlengths 75
75
| comments_url
stringlengths 70
70
| events_url
stringlengths 68
68
| html_url
stringlengths 49
51
| id
int64 942M
3.2B
| node_id
stringlengths 18
32
| number
int64 2.63k
7.67k
| title
stringlengths 1
290
| user
dict | labels
listlengths 0
4
| state
stringclasses 2
values | locked
bool 1
class | assignee
dict | assignees
listlengths 0
4
| milestone
dict | comments
sequencelengths 0
30
| created_at
stringdate 2021-07-12 19:58:31
2025-07-03 11:24:15
| updated_at
stringdate 2021-07-13 05:45:26
2025-07-03 18:34:32
| closed_at
stringlengths 20
20
β | author_association
stringclasses 4
values | type
null | active_lock_reason
null | sub_issues_summary
dict | body
stringlengths 0
58.6k
β | closed_by
dict | reactions
dict | timeline_url
stringlengths 70
70
| performed_via_github_app
null | state_reason
stringclasses 4
values | draft
bool 2
classes | pull_request
dict | is_pull_request
bool 2
classes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/7566 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7566/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7566/comments | https://api.github.com/repos/huggingface/datasets/issues/7566/events | https://github.com/huggingface/datasets/issues/7566 | 3,055,279,344 | I_kwDODunzps62G9zw | 7,566 | terminate called without an active exception; Aborted (core dumped) | {
"avatar_url": "https://avatars.githubusercontent.com/u/18581488?v=4",
"events_url": "https://api.github.com/users/alexey-milovidov/events{/privacy}",
"followers_url": "https://api.github.com/users/alexey-milovidov/followers",
"following_url": "https://api.github.com/users/alexey-milovidov/following{/other_user}",
"gists_url": "https://api.github.com/users/alexey-milovidov/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/alexey-milovidov",
"id": 18581488,
"login": "alexey-milovidov",
"node_id": "MDQ6VXNlcjE4NTgxNDg4",
"organizations_url": "https://api.github.com/users/alexey-milovidov/orgs",
"received_events_url": "https://api.github.com/users/alexey-milovidov/received_events",
"repos_url": "https://api.github.com/users/alexey-milovidov/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/alexey-milovidov/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/alexey-milovidov/subscriptions",
"type": "User",
"url": "https://api.github.com/users/alexey-milovidov",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"@alexey-milovidov I followed the code snippet, but am able to successfully execute without any error. Could you please verify if the error persists or there is any additional details.",
"@alexey-milovidov else if the problem does not exist please feel free to close this issue.",
"```\nmilovidov@milovidov-pc:~/work/datasets$ \n./main.py \nResolving data files: 100%|βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ| 25868/25868 [00:05<00:00, 4753.90it/s]\nResolving data files: 100%|βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ| 25868/25868 [00:00<00:00, 238798.85it/s]\n{'text': \"How AP reported in all formats from tornado-stricken regionsMarch 8, 2012\\nWhen the first serious bout of tornadoes of 2012 blew through middle America in the middle of the night, they touched down in places hours from any AP bureau. Our closest video journalist was Chicago-based Robert Ray, who dropped his plans to travel to Georgia for Super Tuesday, booked several flights to the cities closest to the strikes and headed for the airport. Heβd decide once there which flight to take.\\nHe never got on board a plane. Instead, he ended up driving toward Harrisburg, Ill., where initial reports suggested a town was destroyed. That decision turned out to be a lucky break for the AP. Twice.\\nRay was among the first journalists to arrive and he confirmed those reports -- in all formats. He shot powerful video, put victims on the phone with AP Radio and played back sound to an editor who transcribed the interviews and put the material on text wires. He then walked around the devastation with the Central Regional Desk on the line, talking to victims with the phone held so close that editors could transcribe his interviews in real time.\\nRay also made a dramatic image of a young girl who found a manβs prosthetic leg in the rubble, propped it up next to her destroyed home and spray-painted an impromptu sign: βFound leg. Seriously.β\\nThe following day, he was back on the road and headed for Georgia and a Super Tuesday date with Newt Gingrichβs campaign. The drive would take him through a stretch of the South that forecasters expected would suffer another wave of tornadoes.\\nTo prevent running into THAT storm, Ray used his iPhone to monitor Doppler radar, zooming in on extreme cells and using Google maps to direct himself to safe routes. And then the journalist took over again.\\nβWhen weather like that occurs, a reporter must seize the opportunity to get the news out and allow people to see, hear and read the power of nature so that they can take proper shelter,β Ray says.\\nSo Ray now started to use his phone to follow the storms. He attached a small GoPro camera to his steering wheel in case a tornado dropped down in front of the car somewhere, and took video of heavy rain and hail with his iPhone. Soon, he spotted a tornado and the chase was on. He followed an unmarked emergency vehicle to Cleveland, Tenn., where he was first on the scene of the storm's aftermath.\\nAgain, the tornadoes had struck in locations that were hours from the nearest AP bureau. Damage and debris, as well as a wickedly violent storm that made travel dangerous, slowed our efforts to get to the news. That wasnβt a problem in Tennessee, where our customers were well served by an all-formats report that included this text story.\\nβCLEVELAND, Tenn. (AP) _ Fierce wind, hail and rain lashed Tennessee for the second time in three days, and at least 15 people were hospitalized Friday in the Chattanooga area.β\\nThe byline? Robert Ray.\\nFor being adept with technology, chasing after news as it literally dropped from the sky and setting a standard for all-formats reporting that put the AP ahead on the most competitive news story of the day, Ray wins this weekβs $300 Best of the States prize.\\nΒ© 2013 The Associated Press. All rights reserved. Terms and conditions apply. See AP.org for details.\", 'id': '<urn:uuid:d66bc6fe-8477-4adf-b430-f6a558ccc8ff>', 'dump': 'CC-MAIN-2013-20', 'url': 'http://%[email protected]/Content/Press-Release/2012/How-AP-reported-in-all-formats-from-tornado-stricken-regions', 'date': '2013-05-18T05:48:54Z', 'file_path': 's3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368696381249/warc/CC-MAIN-20130516092621-00000-ip-10-60-113-184.ec2.internal.warc.gz', 'language': 'en', 'language_score': 0.9721424579620361, 'token_count': 717}\nterminate called without an active exception\nAborted (core dumped)\nmilovidov@milovidov-pc:~/work/datasets$ \npython3 --version\nPython 3.10.12\n```",
"Thank you @alexey-milovidov for the details, was able to reproduce the issue.\n\nFollowing is a preliminary analysis which would help to further isolate the issue:\nOn local: \n- For alternate datasets e.g. `speed/english_quotes_paraphrase` instead of `HuggingFaceFW/fineweb` the code works\n- Multiple calls of `print(next(iter(dataset)))` can be performed successfully before the `terminate` is raised, indicating possibility of issue when connection is closed\n\nOn colab:\n- The above code works properly"
] | 2025-05-11T23:05:54Z | 2025-06-23T17:56:02Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I use it as in the tutorial here: https://huggingface.co/docs/datasets/stream, and it ends up with abort.
### Steps to reproduce the bug
1. `pip install datasets`
2.
```
$ cat main.py
#!/usr/bin/env python3
from datasets import load_dataset
dataset = load_dataset('HuggingFaceFW/fineweb', split='train', streaming=True)
print(next(iter(dataset)))
```
3. `chmod +x main.py`
```
$ ./main.py
README.md: 100%|ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ| 43.1k/43.1k [00:00<00:00, 7.04MB/s]
Resolving data files: 100%|ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ| 25868/25868 [00:05<00:00, 4859.26it/s]
Resolving data files: 100%|βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ| 25868/25868 [00:00<00:00, 54773.56it/s]
{'text': "How AP reported in all formats from tornado-stricken regionsMarch 8, 2012\nWhen the first serious bout of tornadoes of 2012 blew through middle America in the middle of the night, they touched down in places hours from any AP bureau. Our closest video journalist was Chicago-based Robert Ray, who dropped his plans to travel to Georgia for Super Tuesday, booked several flights to the cities closest to the strikes and headed for the airport. Heβd decide once there which flight to take.\nHe never got on board a plane. Instead, he ended up driving toward Harrisburg, Ill., where initial reports suggested a town was destroyed. That decision turned out to be a lucky break for the AP. Twice.\nRay was among the first journalists to arrive and he confirmed those reports -- in all formats. He shot powerful video, put victims on the phone with AP Radio and played back sound to an editor who transcribed the interviews and put the material on text wires. He then walked around the devastation with the Central Regional Desk on the line, talking to victims with the phone held so close that editors could transcribe his interviews in real time.\nRay also made a dramatic image of a young girl who found a manβs prosthetic leg in the rubble, propped it up next to her destroyed home and spray-painted an impromptu sign: βFound leg. Seriously.β\nThe following day, he was back on the road and headed for Georgia and a Super Tuesday date with Newt Gingrichβs campaign. The drive would take him through a stretch of the South that forecasters expected would suffer another wave of tornadoes.\nTo prevent running into THAT storm, Ray used his iPhone to monitor Doppler radar, zooming in on extreme cells and using Google maps to direct himself to safe routes. And then the journalist took over again.\nβWhen weather like that occurs, a reporter must seize the opportunity to get the news out and allow people to see, hear and read the power of nature so that they can take proper shelter,β Ray says.\nSo Ray now started to use his phone to follow the storms. He attached a small GoPro camera to his steering wheel in case a tornado dropped down in front of the car somewhere, and took video of heavy rain and hail with his iPhone. Soon, he spotted a tornado and the chase was on. He followed an unmarked emergency vehicle to Cleveland, Tenn., where he was first on the scene of the storm's aftermath.\nAgain, the tornadoes had struck in locations that were hours from the nearest AP bureau. Damage and debris, as well as a wickedly violent storm that made travel dangerous, slowed our efforts to get to the news. That wasnβt a problem in Tennessee, where our customers were well served by an all-formats report that included this text story.\nβCLEVELAND, Tenn. (AP) _ Fierce wind, hail and rain lashed Tennessee for the second time in three days, and at least 15 people were hospitalized Friday in the Chattanooga area.β\nThe byline? Robert Ray.\nFor being adept with technology, chasing after news as it literally dropped from the sky and setting a standard for all-formats reporting that put the AP ahead on the most competitive news story of the day, Ray wins this weekβs $300 Best of the States prize.\nΒ© 2013 The Associated Press. All rights reserved. Terms and conditions apply. See AP.org for details.", 'id': '<urn:uuid:d66bc6fe-8477-4adf-b430-f6a558ccc8ff>', 'dump': 'CC-MAIN-2013-20', 'url': 'http://%[email protected]/Content/Press-Release/2012/How-AP-reported-in-all-formats-from-tornado-stricken-regions', 'date': '2013-05-18T05:48:54Z', 'file_path': 's3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368696381249/warc/CC-MAIN-20130516092621-00000-ip-10-60-113-184.ec2.internal.warc.gz', 'language': 'en', 'language_score': 0.9721424579620361, 'token_count': 717}
terminate called without an active exception
Aborted (core dumped)
```
### Expected behavior
I'm not a proficient Python user, so it might be my own error, but even in that case, the error message should be better.
### Environment info
`Successfully installed datasets-3.6.0 dill-0.3.8 hf-xet-1.1.0 huggingface-hub-0.31.1 multiprocess-0.70.16 requests-2.32.3 xxhash-3.5.0`
```
$ cat /etc/lsb-release
DISTRIB_ID=Ubuntu
DISTRIB_RELEASE=22.04
DISTRIB_CODENAME=jammy
DISTRIB_DESCRIPTION="Ubuntu 22.04.4 LTS"
``` | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7566/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7566/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7565 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7565/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7565/comments | https://api.github.com/repos/huggingface/datasets/issues/7565/events | https://github.com/huggingface/datasets/pull/7565 | 3,051,731,207 | PR_kwDODunzps6VkFBm | 7,565 | add check if repo exists for dataset uploading | {
"avatar_url": "https://avatars.githubusercontent.com/u/36135455?v=4",
"events_url": "https://api.github.com/users/Samoed/events{/privacy}",
"followers_url": "https://api.github.com/users/Samoed/followers",
"following_url": "https://api.github.com/users/Samoed/following{/other_user}",
"gists_url": "https://api.github.com/users/Samoed/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Samoed",
"id": 36135455,
"login": "Samoed",
"node_id": "MDQ6VXNlcjM2MTM1NDU1",
"organizations_url": "https://api.github.com/users/Samoed/orgs",
"received_events_url": "https://api.github.com/users/Samoed/received_events",
"repos_url": "https://api.github.com/users/Samoed/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Samoed/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Samoed/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Samoed",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7565). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"@lhoestq Can you review, please? I don't think that errors in CI are related to my changes"
] | 2025-05-09T10:27:00Z | 2025-06-09T14:39:23Z | null | NONE | null | null | null | Currently, I'm reuploading datasets for [`MTEB`](https://github.com/embeddings-benchmark/mteb/). Some of them have many splits (more than 20), and I'm encountering the error:
`Too many requests for https://huggingface.co/datasets/repo/create`.
It seems that this issue occurs because the dataset tries to recreate itself every time a split is uploaded. To resolve this, I've added a check to ensure that if the dataset already exists, it won't attempt to recreate it. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7565/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7565/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7565.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7565",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7565.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7565"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7564 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7564/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7564/comments | https://api.github.com/repos/huggingface/datasets/issues/7564/events | https://github.com/huggingface/datasets/pull/7564 | 3,049,275,226 | PR_kwDODunzps6VczLS | 7,564 | Implementation of iteration over values of a column in an IterableDataset object | {
"avatar_url": "https://avatars.githubusercontent.com/u/47208659?v=4",
"events_url": "https://api.github.com/users/TopCoder2K/events{/privacy}",
"followers_url": "https://api.github.com/users/TopCoder2K/followers",
"following_url": "https://api.github.com/users/TopCoder2K/following{/other_user}",
"gists_url": "https://api.github.com/users/TopCoder2K/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/TopCoder2K",
"id": 47208659,
"login": "TopCoder2K",
"node_id": "MDQ6VXNlcjQ3MjA4NjU5",
"organizations_url": "https://api.github.com/users/TopCoder2K/orgs",
"received_events_url": "https://api.github.com/users/TopCoder2K/received_events",
"repos_url": "https://api.github.com/users/TopCoder2K/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/TopCoder2K/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/TopCoder2K/subscriptions",
"type": "User",
"url": "https://api.github.com/users/TopCoder2K",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"A couple of questions:\r\n1. I've noticed two strange things: 1) \"Around 80% of the final dataset is made of the `en_dataset`\" in https://huggingface.co/docs/datasets/stream, 2) \"Click on \"Pull request\" to send your to the project maintainers\" in https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md Are `en_dataset` and \"your [???]\" typos? If so, I can fix them in this PR.\r\n2. Should I update https://huggingface.co/docs/datasets/stream or https://huggingface.co/docs/datasets/access#iterabledataset to include the new feature?",
"Great ! and chained indexing was easy indeed, thanks :)\r\n\r\nregarding your questions:\r\n\r\n> I've noticed two strange things: 1) \"Around 80% of the final dataset is made of the en_dataset\" in https://huggingface.co/docs/datasets/stream, 2) \"Click on \"Pull request\" to send your to the project maintainers\" in https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md Are en_dataset and \"your [???]\" typos? If so, I can fix them in this PR.\r\n\r\nOh good catch, both should be fixed indeed. Feel free to open a new PR for those docs fixes\r\n\r\n> Should I update https://huggingface.co/docs/datasets/stream or https://huggingface.co/docs/datasets/access#iterabledataset to include the new feature?\r\n\r\nYep good idea, I think in both places, since /stream is supposed to be exhaustive, and /access already mentions accessing a specific column for `Dataset`",
"@lhoestq, thank you for the answers!\r\n\r\n> Yep good idea, I think in both places, since /stream is supposed to be exhaustive, and /access already mentions accessing a specific column for Dataset\r\n\r\nπ, I'll try to add something.\r\n\r\nBy the way, do you have any ideas about why the CI pipelines have failed? Essentially, I've already encountered these problems [here](https://github.com/huggingface/datasets/issues/7381#issuecomment-2863421974).\r\nI think `check_code_quality` has failed due to the usage of `pre-commit`. The problem seems to be the old version of the ruff hook. I've tried `v0.11.8` (the one that was installed with `pip install -e \".[quality]\"`) and `pre-commit` seems to work like `make style` now. However, I don't have any ideas about `pyav` since I don't know what it is...",
"I've updated /stream and /access, please check the style and clarity. By the way, I would like to add `IterableDataset.skip` near `IterableDataset.take` to mimic [slicing](https://huggingface.co/docs/datasets/access/#slicing). What do you think?",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7564). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-05-08T14:59:22Z | 2025-05-19T12:15:02Z | 2025-05-19T12:15:02Z | CONTRIBUTOR | null | null | null | Refers to [this issue](https://github.com/huggingface/datasets/issues/7381). | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7564/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7564/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7564.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7564",
"merged_at": "2025-05-19T12:15:02Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7564.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7564"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7563 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7563/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7563/comments | https://api.github.com/repos/huggingface/datasets/issues/7563/events | https://github.com/huggingface/datasets/pull/7563 | 3,046,351,253 | PR_kwDODunzps6VS0QL | 7,563 | set dev version | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7563). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-05-07T15:18:29Z | 2025-05-07T15:21:05Z | 2025-05-07T15:18:36Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7563/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7563/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7563.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7563",
"merged_at": "2025-05-07T15:18:36Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7563.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7563"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7562 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7562/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7562/comments | https://api.github.com/repos/huggingface/datasets/issues/7562/events | https://github.com/huggingface/datasets/pull/7562 | 3,046,339,430 | PR_kwDODunzps6VSxmx | 7,562 | release: 3.6.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7562). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-05-07T15:15:13Z | 2025-05-07T15:17:46Z | 2025-05-07T15:15:21Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 1,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7562/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7562/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7562.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7562",
"merged_at": "2025-05-07T15:15:20Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7562.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7562"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7561 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7561/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7561/comments | https://api.github.com/repos/huggingface/datasets/issues/7561/events | https://github.com/huggingface/datasets/issues/7561 | 3,046,302,653 | I_kwDODunzps61kuO9 | 7,561 | NotImplementedError: <class 'datasets.iterable_dataset.RepeatExamplesIterable'> doesn't implement num_shards yet | {
"avatar_url": "https://avatars.githubusercontent.com/u/32219669?v=4",
"events_url": "https://api.github.com/users/cyanic-selkie/events{/privacy}",
"followers_url": "https://api.github.com/users/cyanic-selkie/followers",
"following_url": "https://api.github.com/users/cyanic-selkie/following{/other_user}",
"gists_url": "https://api.github.com/users/cyanic-selkie/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/cyanic-selkie",
"id": 32219669,
"login": "cyanic-selkie",
"node_id": "MDQ6VXNlcjMyMjE5NjY5",
"organizations_url": "https://api.github.com/users/cyanic-selkie/orgs",
"received_events_url": "https://api.github.com/users/cyanic-selkie/received_events",
"repos_url": "https://api.github.com/users/cyanic-selkie/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/cyanic-selkie/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/cyanic-selkie/subscriptions",
"type": "User",
"url": "https://api.github.com/users/cyanic-selkie",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [] | 2025-05-07T15:05:42Z | 2025-06-05T12:41:30Z | 2025-06-05T12:41:30Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When using `.repeat()` on an `IterableDataset`, this error gets thrown. There is [this thread](https://discuss.huggingface.co/t/making-an-infinite-iterabledataset/146192/5) that seems to imply the fix is trivial, but I don't know anything about this codebase, so I'm opening this issue rather than attempting to open a PR.
### Steps to reproduce the bug
1. Create an `IterableDataset`.
2. Call `.repeat(None)` on it.
3. Wrap it in a pytorch `DataLoader`
4. Iterate over it.
### Expected behavior
This should work normally.
### Environment info
datasets: 3.5.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7561/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7561/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7560 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7560/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7560/comments | https://api.github.com/repos/huggingface/datasets/issues/7560/events | https://github.com/huggingface/datasets/pull/7560 | 3,046,265,500 | PR_kwDODunzps6VShIc | 7,560 | fix decoding tests | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7560). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-05-07T14:56:14Z | 2025-05-07T14:59:02Z | 2025-05-07T14:56:20Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7560/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7560/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7560.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7560",
"merged_at": "2025-05-07T14:56:20Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7560.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7560"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7559 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7559/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7559/comments | https://api.github.com/repos/huggingface/datasets/issues/7559/events | https://github.com/huggingface/datasets/pull/7559 | 3,046,177,078 | PR_kwDODunzps6VSNiX | 7,559 | fix aiohttp import | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7559). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-05-07T14:31:32Z | 2025-05-07T14:34:34Z | 2025-05-07T14:31:38Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7559/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7559/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7559.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7559",
"merged_at": "2025-05-07T14:31:38Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7559.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7559"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7558 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7558/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7558/comments | https://api.github.com/repos/huggingface/datasets/issues/7558/events | https://github.com/huggingface/datasets/pull/7558 | 3,046,066,628 | PR_kwDODunzps6VR1gN | 7,558 | fix regression | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7558). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-05-07T13:56:03Z | 2025-05-07T13:58:52Z | 2025-05-07T13:56:18Z | MEMBER | null | null | null | reported in https://github.com/huggingface/datasets/pull/7557 (I just reorganized the condition)
wanted to apply this change to the original PR but github didn't let me apply it directly - merging this one instead | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7558/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7558/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7558.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7558",
"merged_at": "2025-05-07T13:56:18Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7558.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7558"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7557 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7557/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7557/comments | https://api.github.com/repos/huggingface/datasets/issues/7557/events | https://github.com/huggingface/datasets/pull/7557 | 3,045,962,076 | PR_kwDODunzps6VRenr | 7,557 | check for empty _formatting | {
"avatar_url": "https://avatars.githubusercontent.com/u/381258?v=4",
"events_url": "https://api.github.com/users/winglian/events{/privacy}",
"followers_url": "https://api.github.com/users/winglian/followers",
"following_url": "https://api.github.com/users/winglian/following{/other_user}",
"gists_url": "https://api.github.com/users/winglian/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/winglian",
"id": 381258,
"login": "winglian",
"node_id": "MDQ6VXNlcjM4MTI1OA==",
"organizations_url": "https://api.github.com/users/winglian/orgs",
"received_events_url": "https://api.github.com/users/winglian/received_events",
"repos_url": "https://api.github.com/users/winglian/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/winglian/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/winglian/subscriptions",
"type": "User",
"url": "https://api.github.com/users/winglian",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Thanks for reporting and for the fix ! I tried to reorganize the condition in your PR but didn't get the right permission so. I ended up merging https://github.com/huggingface/datasets/pull/7558 directly so I can make a release today - I hope you don't mind"
] | 2025-05-07T13:22:37Z | 2025-05-07T13:57:12Z | 2025-05-07T13:57:12Z | CONTRIBUTOR | null | null | null | Fixes a regression from #7553 breaking shuffling of iterable datasets
<img width="884" alt="Screenshot 2025-05-07 at 9 16 52β―AM" src="https://github.com/user-attachments/assets/d2f43c5f-4092-4efe-ac31-a32cbd025fe3" />
| {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7557/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7557/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7557.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7557",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7557.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7557"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7556 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7556/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7556/comments | https://api.github.com/repos/huggingface/datasets/issues/7556/events | https://github.com/huggingface/datasets/pull/7556 | 3,043,615,210 | PR_kwDODunzps6VJlTR | 7,556 | Add `--merge-pull-request` option for `convert_to_parquet` | {
"avatar_url": "https://avatars.githubusercontent.com/u/17013474?v=4",
"events_url": "https://api.github.com/users/klamike/events{/privacy}",
"followers_url": "https://api.github.com/users/klamike/followers",
"following_url": "https://api.github.com/users/klamike/following{/other_user}",
"gists_url": "https://api.github.com/users/klamike/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/klamike",
"id": 17013474,
"login": "klamike",
"node_id": "MDQ6VXNlcjE3MDEzNDc0",
"organizations_url": "https://api.github.com/users/klamike/orgs",
"received_events_url": "https://api.github.com/users/klamike/received_events",
"repos_url": "https://api.github.com/users/klamike/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/klamike/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/klamike/subscriptions",
"type": "User",
"url": "https://api.github.com/users/klamike",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"This is ready for a review, happy to make any changes. The main question for maintainers is how this should interact with #7555. If my suggestion there is accepted, this PR can be kept as is. If not, more changes are required to merge all the PR parts."
] | 2025-05-06T18:05:05Z | 2025-05-07T17:41:16Z | null | NONE | null | null | null | Closes #7527
Note that this implementation **will only merge the last PR in the case that they get split up by `push_to_hub`**. See https://github.com/huggingface/datasets/discussions/7555 for more details. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7556/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7556/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7556.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7556",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7556.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7556"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7554 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7554/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7554/comments | https://api.github.com/repos/huggingface/datasets/issues/7554/events | https://github.com/huggingface/datasets/issues/7554 | 3,043,089,844 | I_kwDODunzps61Yd20 | 7,554 | datasets downloads and generates all splits, even though a single split is requested (for dataset with loading script) | {
"avatar_url": "https://avatars.githubusercontent.com/u/50171988?v=4",
"events_url": "https://api.github.com/users/sei-eschwartz/events{/privacy}",
"followers_url": "https://api.github.com/users/sei-eschwartz/followers",
"following_url": "https://api.github.com/users/sei-eschwartz/following{/other_user}",
"gists_url": "https://api.github.com/users/sei-eschwartz/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sei-eschwartz",
"id": 50171988,
"login": "sei-eschwartz",
"node_id": "MDQ6VXNlcjUwMTcxOTg4",
"organizations_url": "https://api.github.com/users/sei-eschwartz/orgs",
"received_events_url": "https://api.github.com/users/sei-eschwartz/received_events",
"repos_url": "https://api.github.com/users/sei-eschwartz/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sei-eschwartz/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sei-eschwartz/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sei-eschwartz",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! there has been some effort on allowing to download only a subset of splits in https://github.com/huggingface/datasets/pull/6832 but no one has been continuing this work so far. This would be a welcomed contribution though\n\nAlso note that loading script are often unoptimized, and we recommend using datasets in standard formats like Parquet instead.\n\nBtw there is a CLI tool to convert a loading script to parquet:\n\n```\ndatasets-cli convert_to_parquet <dataset-name> --trust_remote_code\n```",
"Closing in favor of #6832 "
] | 2025-05-06T14:43:38Z | 2025-05-07T14:53:45Z | 2025-05-07T14:53:44Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
`datasets` downloads and generates all splits, even though a single split is requested. [This](https://huggingface.co/datasets/jordiae/exebench) is the dataset in question. It uses a loading script. I am not 100% sure that this is a bug, because maybe with loading scripts `datasets` must actually process all the splits? But I thought loading scripts were designed to avoid this.
### Steps to reproduce the bug
See [this notebook](https://colab.research.google.com/drive/14kcXp_hgcdj-kIzK0bCG6taE-CLZPVvq?usp=sharing)
Or:
```python
from datasets import load_dataset
dataset = load_dataset('jordiae/exebench', split='test_synth', trust_remote_code=True)
```
### Expected behavior
I expected only the `test_synth` split to be downloaded and processed.
### Environment info
- `datasets` version: 3.5.1
- Platform: Linux-6.1.123+-x86_64-with-glibc2.35
- Python version: 3.11.12
- `huggingface_hub` version: 0.30.2
- PyArrow version: 18.1.0
- Pandas version: 2.2.2
- `fsspec` version: 2025.3.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/50171988?v=4",
"events_url": "https://api.github.com/users/sei-eschwartz/events{/privacy}",
"followers_url": "https://api.github.com/users/sei-eschwartz/followers",
"following_url": "https://api.github.com/users/sei-eschwartz/following{/other_user}",
"gists_url": "https://api.github.com/users/sei-eschwartz/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sei-eschwartz",
"id": 50171988,
"login": "sei-eschwartz",
"node_id": "MDQ6VXNlcjUwMTcxOTg4",
"organizations_url": "https://api.github.com/users/sei-eschwartz/orgs",
"received_events_url": "https://api.github.com/users/sei-eschwartz/received_events",
"repos_url": "https://api.github.com/users/sei-eschwartz/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sei-eschwartz/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sei-eschwartz/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sei-eschwartz",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7554/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7554/timeline | null | duplicate | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7553 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7553/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7553/comments | https://api.github.com/repos/huggingface/datasets/issues/7553/events | https://github.com/huggingface/datasets/pull/7553 | 3,042,953,907 | PR_kwDODunzps6VHUNW | 7,553 | Rebatch arrow iterables before formatted iterable | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7553). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"@lhoestq Our CI found an issue with this changeset causing a regression with shuffling iterable datasets \r\n<img width=\"884\" alt=\"Screenshot 2025-05-07 at 9 16 52β―AM\" src=\"https://github.com/user-attachments/assets/bf7d9c7e-cc14-47da-8da6-d1a345992d7c\" />\r\n"
] | 2025-05-06T13:59:58Z | 2025-05-07T13:17:41Z | 2025-05-06T14:03:42Z | MEMBER | null | null | null | close https://github.com/huggingface/datasets/issues/7538 and https://github.com/huggingface/datasets/issues/7475 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7553/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7553/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7553.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7553",
"merged_at": "2025-05-06T14:03:41Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7553.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7553"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7552 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7552/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7552/comments | https://api.github.com/repos/huggingface/datasets/issues/7552/events | https://github.com/huggingface/datasets/pull/7552 | 3,040,258,084 | PR_kwDODunzps6U-BUv | 7,552 | Enable xet in push to hub | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7552). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-05-05T17:02:09Z | 2025-05-06T12:42:51Z | 2025-05-06T12:42:48Z | MEMBER | null | null | null | follows https://github.com/huggingface/huggingface_hub/pull/3035
related to https://github.com/huggingface/datasets/issues/7526 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7552/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7552/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7552.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7552",
"merged_at": "2025-05-06T12:42:48Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7552.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7552"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7551 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7551/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7551/comments | https://api.github.com/repos/huggingface/datasets/issues/7551/events | https://github.com/huggingface/datasets/issues/7551 | 3,038,114,928 | I_kwDODunzps61FfRw | 7,551 | Issue with offline mode and partial dataset cached | {
"avatar_url": "https://avatars.githubusercontent.com/u/353245?v=4",
"events_url": "https://api.github.com/users/nrv/events{/privacy}",
"followers_url": "https://api.github.com/users/nrv/followers",
"following_url": "https://api.github.com/users/nrv/following{/other_user}",
"gists_url": "https://api.github.com/users/nrv/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/nrv",
"id": 353245,
"login": "nrv",
"node_id": "MDQ6VXNlcjM1MzI0NQ==",
"organizations_url": "https://api.github.com/users/nrv/orgs",
"received_events_url": "https://api.github.com/users/nrv/received_events",
"repos_url": "https://api.github.com/users/nrv/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/nrv/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nrv/subscriptions",
"type": "User",
"url": "https://api.github.com/users/nrv",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"It seems the problem comes from builder.py / create_config_id()\n\nOn the first call, when the cache is empty we have\n```\nconfig_kwargs = {'data_files': {'train': ['hf://datasets/uonlp/CulturaX@6a8734bc69fefcbb7735f4f9250f43e4cd7a442e/fr/fr_part_00038.parquet']}}\n```\nleading to config_id beeing 'default-2935e8cdcc21c613'\n\nthen, on the second call, \n```\nconfig_kwargs = {'data_files': 'fr/fr_part_00038.parquet'}\n```\nthus explaining why the hash is not the same, despite having the same parameter when calling load_dataset : data_files=\"fr/fr_part_00038.parquet\"",
"Same behavior with version 3.5.1",
"Same issue when loading `google/IndicGenBench_flores_in` with `dataset==2.21.0` and `dataset==3.6.0` .",
"\n\n\n> It seems the problem comes from builder.py / create_config_id()\n> \n> On the first call, when the cache is empty we have\n> \n> ```\n> config_kwargs = {'data_files': {'train': ['hf://datasets/uonlp/CulturaX@6a8734bc69fefcbb7735f4f9250f43e4cd7a442e/fr/fr_part_00038.parquet']}}\n> ```\n> \n> leading to config_id beeing 'default-2935e8cdcc21c613'\n> \n> then, on the second call,\n> \n> ```\n> config_kwargs = {'data_files': 'fr/fr_part_00038.parquet'}\n> ```\n> \n> thus explaining why the hash is not the same, despite having the same parameter when calling load_dataset : data_files=\"fr/fr_part_00038.parquet\"\n\n\nI have identified that the issue indeed lies in the `data_files` within `config_kwargs`. \nThe format and prefix of `data_files` differ depending on whether `HF_HUB_OFFLINE` is set, leading to different final `config_id` values. \nWhen I use other datasets without passing the `data_files` parameter, this issue does not occur.\n\nA possible solution might be to standardize the formatting of `data_files` within the `create_config_id` function."
] | 2025-05-04T16:49:37Z | 2025-05-13T03:18:43Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Hi,
a issue related to #4760 here when loading a single file from a dataset, unable to access it in offline mode afterwards
### Steps to reproduce the bug
```python
import os
# os.environ["HF_HUB_OFFLINE"] = "1"
os.environ["HF_TOKEN"] = "xxxxxxxxxxxxxx"
import datasets
dataset_name = "uonlp/CulturaX"
data_files = "fr/fr_part_00038.parquet"
ds = datasets.load_dataset(dataset_name, split='train', data_files=data_files)
print(f"Dataset loaded : {ds}")
```
Once the file has been cached, I rerun with the HF_HUB_OFFLINE activated an get this error :
```
ValueError: Couldn't find cache for uonlp/CulturaX for config 'default-1e725f978350254e'
Available configs in the cache: ['default-2935e8cdcc21c613']
```
### Expected behavior
Should be able to access the previously cached files
### Environment info
- `datasets` version: 3.2.0
- Platform: Linux-5.4.0-215-generic-x86_64-with-glibc2.31
- Python version: 3.12.0
- `huggingface_hub` version: 0.27.0
- PyArrow version: 19.0.0
- Pandas version: 2.2.2
- `fsspec` version: 2024.3.1
| null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7551/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7551/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7550 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7550/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7550/comments | https://api.github.com/repos/huggingface/datasets/issues/7550/events | https://github.com/huggingface/datasets/pull/7550 | 3,037,017,367 | PR_kwDODunzps6UzksN | 7,550 | disable aiohttp depend for python 3.13t free-threading compat | {
"avatar_url": "https://avatars.githubusercontent.com/u/417764?v=4",
"events_url": "https://api.github.com/users/Qubitium/events{/privacy}",
"followers_url": "https://api.github.com/users/Qubitium/followers",
"following_url": "https://api.github.com/users/Qubitium/following{/other_user}",
"gists_url": "https://api.github.com/users/Qubitium/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Qubitium",
"id": 417764,
"login": "Qubitium",
"node_id": "MDQ6VXNlcjQxNzc2NA==",
"organizations_url": "https://api.github.com/users/Qubitium/orgs",
"received_events_url": "https://api.github.com/users/Qubitium/received_events",
"repos_url": "https://api.github.com/users/Qubitium/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Qubitium/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Qubitium/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Qubitium",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [] | 2025-05-03T00:28:18Z | 2025-05-03T00:28:24Z | 2025-05-03T00:28:24Z | NONE | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/417764?v=4",
"events_url": "https://api.github.com/users/Qubitium/events{/privacy}",
"followers_url": "https://api.github.com/users/Qubitium/followers",
"following_url": "https://api.github.com/users/Qubitium/following{/other_user}",
"gists_url": "https://api.github.com/users/Qubitium/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Qubitium",
"id": 417764,
"login": "Qubitium",
"node_id": "MDQ6VXNlcjQxNzc2NA==",
"organizations_url": "https://api.github.com/users/Qubitium/orgs",
"received_events_url": "https://api.github.com/users/Qubitium/received_events",
"repos_url": "https://api.github.com/users/Qubitium/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Qubitium/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Qubitium/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Qubitium",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7550/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7550/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7550.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7550",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7550.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7550"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7549 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7549/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7549/comments | https://api.github.com/repos/huggingface/datasets/issues/7549/events | https://github.com/huggingface/datasets/issues/7549 | 3,036,272,015 | I_kwDODunzps60-dWP | 7,549 | TypeError: Couldn't cast array of type string to null on webdataset format dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/117186571?v=4",
"events_url": "https://api.github.com/users/narugo1992/events{/privacy}",
"followers_url": "https://api.github.com/users/narugo1992/followers",
"following_url": "https://api.github.com/users/narugo1992/following{/other_user}",
"gists_url": "https://api.github.com/users/narugo1992/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/narugo1992",
"id": 117186571,
"login": "narugo1992",
"node_id": "U_kgDOBvwgCw",
"organizations_url": "https://api.github.com/users/narugo1992/orgs",
"received_events_url": "https://api.github.com/users/narugo1992/received_events",
"repos_url": "https://api.github.com/users/narugo1992/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/narugo1992/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/narugo1992/subscriptions",
"type": "User",
"url": "https://api.github.com/users/narugo1992",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"seems to get fixed by explicitly adding `dataset_infos.json` like this\n\n```json\n{\n \"default\": {\n \"description\": \"Image dataset with tags and ratings\",\n \"citation\": \"\",\n \"homepage\": \"\",\n \"license\": \"\",\n \"features\": {\n \"image\": {\n \"dtype\": \"image\",\n \"_type\": \"Image\"\n },\n \"json\": {\n \"id\": {\n \"dtype\": \"int32\",\n \"_type\": \"Value\"\n },\n \"width\": {\n \"dtype\": \"int32\",\n \"_type\": \"Value\"\n },\n \"height\": {\n \"dtype\": \"int32\",\n \"_type\": \"Value\"\n },\n \"rating\": {\n \"feature\": {\n \"dtype\": \"string\",\n \"_type\": \"Value\"\n },\n \"_type\": \"Sequence\"\n },\n \"general_tags\": {\n \"feature\": {\n \"dtype\": \"string\",\n \"_type\": \"Value\"\n },\n \"_type\": \"Sequence\"\n },\n \"character_tags\": {\n \"feature\": {\n \"dtype\": \"string\",\n \"_type\": \"Value\"\n },\n \"_type\": \"Sequence\"\n }\n }\n },\n \"builder_name\": \"webdataset\",\n \"config_name\": \"default\",\n \"version\": {\n \"version_str\": \"1.0.0\",\n \"description\": null,\n \"major\": 1,\n \"minor\": 0,\n \"patch\": 0\n }\n }\n}\n\n```\n\nwill close this issue if no further issues found"
] | 2025-05-02T15:18:07Z | 2025-05-02T15:37:05Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
```python
from datasets import load_dataset
dataset = load_dataset("animetimm/danbooru-wdtagger-v4-w640-ws-30k")
```
got
```
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/arrow_writer.py", line 626, in write_batch
arrays.append(pa.array(typed_sequence))
File "pyarrow/array.pxi", line 255, in pyarrow.lib.array
File "pyarrow/array.pxi", line 117, in pyarrow.lib._handle_arrow_array_protocol
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/arrow_writer.py", line 258, in __arrow_array__
out = cast_array_to_feature(
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 1798, in wrapper
return func(array, *args, **kwargs)
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 2006, in cast_array_to_feature
arrays = [
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 2007, in <listcomp>
_c(array.field(name) if name in array_fields else null_array, subfeature)
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 1798, in wrapper
return func(array, *args, **kwargs)
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 2066, in cast_array_to_feature
casted_array_values = _c(array.values, feature.feature)
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 1798, in wrapper
return func(array, *args, **kwargs)
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 2103, in cast_array_to_feature
return array_cast(
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 1798, in wrapper
return func(array, *args, **kwargs)
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/table.py", line 1949, in array_cast
raise TypeError(f"Couldn't cast array of type {_short_str(array.type)} to {_short_str(pa_type)}")
TypeError: Couldn't cast array of type string to null
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/load.py", line 2084, in load_dataset
builder_instance.download_and_prepare(
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/builder.py", line 925, in download_and_prepare
self._download_and_prepare(
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/builder.py", line 1649, in _download_and_prepare
super()._download_and_prepare(
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/builder.py", line 1001, in _download_and_prepare
self._prepare_split(split_generator, **prepare_split_kwargs)
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/builder.py", line 1487, in _prepare_split
for job_id, done, content in self._prepare_split_single(
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/datasets/builder.py", line 1644, in _prepare_split_single
raise DatasetGenerationError("An error occurred while generating the dataset") from e
datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
```
`datasets==3.5.1` whats wrong
its inner json structure is like
```yaml
features:
- name: "image"
dtype: "image"
- name: "json.id"
dtype: "string"
- name: "json.width"
dtype: "int32"
- name: "json.height"
dtype: "int32"
- name: "json.rating"
sequence:
dtype: "string"
- name: "json.general_tags"
sequence:
dtype: "string"
- name: "json.character_tags"
sequence:
dtype: "string"
```
i'm 100% sure all the jsons satisfies the abovementioned format.
### Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset("animetimm/danbooru-wdtagger-v4-w640-ws-30k")
```
### Expected behavior
load the dataset successfully, with the abovementioned json format and webp images
### Environment info
Copy-and-paste the text below in your GitHub issue.
- `datasets` version: 3.5.1
- Platform: Linux-6.8.0-52-generic-x86_64-with-glibc2.35
- Python version: 3.10.16
- `huggingface_hub` version: 0.30.2
- PyArrow version: 20.0.0
- Pandas version: 2.2.3
- `fsspec` version: 2025.3.0
| null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7549/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7549/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7548 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7548/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7548/comments | https://api.github.com/repos/huggingface/datasets/issues/7548/events | https://github.com/huggingface/datasets/issues/7548 | 3,035,568,851 | I_kwDODunzps607xrT | 7,548 | Python 3.13t (free threads) Compat | {
"avatar_url": "https://avatars.githubusercontent.com/u/417764?v=4",
"events_url": "https://api.github.com/users/Qubitium/events{/privacy}",
"followers_url": "https://api.github.com/users/Qubitium/followers",
"following_url": "https://api.github.com/users/Qubitium/following{/other_user}",
"gists_url": "https://api.github.com/users/Qubitium/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Qubitium",
"id": 417764,
"login": "Qubitium",
"node_id": "MDQ6VXNlcjQxNzc2NA==",
"organizations_url": "https://api.github.com/users/Qubitium/orgs",
"received_events_url": "https://api.github.com/users/Qubitium/received_events",
"repos_url": "https://api.github.com/users/Qubitium/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Qubitium/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Qubitium/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Qubitium",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Update: `datasets` use `aiohttp` for data streaming and from what I understand data streaming is useful for large datasets that do not fit in memory and/or multi-modal datasets like image/audio where you only what the actual binary bits to fed in as needed. \n\nHowever, there are also many cases where aiohttp will never be used. Text datasets that are not huge, relative to machine spec, and non-multi-modal datasets. \n\nGetting `aiohttp` fixed for `free threading` appeals to be a large task that is not going to be get done in a quick manner. It may be faster to make `aiohttp` optional and not forced build. Otherwise, testing python 3.13t is going to be a painful install. \n\nI have created a fork/branch that temp disables aiohttp import so non-streaming usage of datasets can be tested under python 3.13.t:\n\nhttps://github.com/Qubitium/datasets/tree/disable-aiohttp-depend",
"We are mostly relying on `huggingface_hub` which uses `requests` to stream files from Hugging Face, so maybe we can move aiohttp to optional dependencies now. Would it solve your issue ? Btw what do you think of `datasets` in the free-threading setting ?",
"> We are mostly relying on `huggingface_hub` which uses `requests` to stream files from Hugging Face, so maybe we can move aiohttp to optional dependencies now. Would it solve your issue ? Btw what do you think of `datasets` in the free-threading setting ?\n\nI am testing transformers + dataset (simple text dataset usage) + GPTQModel for quantization and there were no issues encountered with python 3.13t but my test-case is the base-bare minimal test-case since dataset is not sharded, fully in-memory, text-only, small, not used for training. \n\nOn the technical side, dataset is almost always 100% read-only so there should be zero locking issues but I have not checked the dataset internals so there may be cases where streaming, sharding, and/or cases where datset memory/states are updated needs a per dataset `threading.lock`. \n\nSo yes, making `aiohttp` optional will definitely solve my issue. There is also a companion (datasets and tokenizers usually go hand-in-hand) issue with `Tokenizers` as well but that's simple enough with package version update: https://github.com/huggingface/tokenizers/pull/1774\n",
"Ok I see ! Anyway feel free to edit the setup.py to move aiohttp to optional (tests) dependencies and open a PR, we can run the CI to see if it's ok as a change",
"actually there is https://github.com/huggingface/datasets/pull/7294/ already, let's see if we can merge it",
"wouldn't it be the good reason to switch to `httpx`? π (would require slightly more work, short term agree with https://github.com/huggingface/datasets/issues/7548#issuecomment-2854405923)",
"I made `aiohttp` optional in `datasets` 3.6.0 :)\n\n`datasets` doesn't use it directly anyway, it's only used when someone wants to download files from HTTP URLs outside of HF"
] | 2025-05-02T09:20:09Z | 2025-05-12T15:11:32Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Cannot install `datasets` under `python 3.13t` due to dependency on `aiohttp` and aiohttp cannot be built for free-threading python.
The `free threading` support issue in `aiothttp` is active since August 2024! Ouch.
https://github.com/aio-libs/aiohttp/issues/8796#issue-2475941784
`pip install dataset`
```bash
(vm313t) root@gpu-base:~/GPTQModel# pip install datasets
WARNING: Retrying (Retry(total=4, connect=None, read=None, redirect=None, status=None)) after connection broken by 'ReadTimeoutError("HTTPSConnectionPool(host='pypi.org', port=443): Read timed out. (read timeout=15)")': /simple/datasets/
Collecting datasets
Using cached datasets-3.5.1-py3-none-any.whl.metadata (19 kB)
Requirement already satisfied: filelock in /root/vm313t/lib/python3.13t/site-packages (from datasets) (3.18.0)
Requirement already satisfied: numpy>=1.17 in /root/vm313t/lib/python3.13t/site-packages (from datasets) (2.2.5)
Collecting pyarrow>=15.0.0 (from datasets)
Using cached pyarrow-20.0.0-cp313-cp313t-manylinux_2_28_x86_64.whl.metadata (3.3 kB)
Collecting dill<0.3.9,>=0.3.0 (from datasets)
Using cached dill-0.3.8-py3-none-any.whl.metadata (10 kB)
Collecting pandas (from datasets)
Using cached pandas-2.2.3-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (89 kB)
Requirement already satisfied: requests>=2.32.2 in /root/vm313t/lib/python3.13t/site-packages (from datasets) (2.32.3)
Requirement already satisfied: tqdm>=4.66.3 in /root/vm313t/lib/python3.13t/site-packages (from datasets) (4.67.1)
Collecting xxhash (from datasets)
Using cached xxhash-3.5.0-cp313-cp313t-linux_x86_64.whl
Collecting multiprocess<0.70.17 (from datasets)
Using cached multiprocess-0.70.16-py312-none-any.whl.metadata (7.2 kB)
Collecting fsspec<=2025.3.0,>=2023.1.0 (from fsspec[http]<=2025.3.0,>=2023.1.0->datasets)
Using cached fsspec-2025.3.0-py3-none-any.whl.metadata (11 kB)
Collecting aiohttp (from datasets)
Using cached aiohttp-3.11.18.tar.gz (7.7 MB)
Installing build dependencies ... done
Getting requirements to build wheel ... done
Preparing metadata (pyproject.toml) ... done
Requirement already satisfied: huggingface-hub>=0.24.0 in /root/vm313t/lib/python3.13t/site-packages (from datasets) (0.30.2)
Requirement already satisfied: packaging in /root/vm313t/lib/python3.13t/site-packages (from datasets) (25.0)
Requirement already satisfied: pyyaml>=5.1 in /root/vm313t/lib/python3.13t/site-packages (from datasets) (6.0.2)
Collecting aiohappyeyeballs>=2.3.0 (from aiohttp->datasets)
Using cached aiohappyeyeballs-2.6.1-py3-none-any.whl.metadata (5.9 kB)
Collecting aiosignal>=1.1.2 (from aiohttp->datasets)
Using cached aiosignal-1.3.2-py2.py3-none-any.whl.metadata (3.8 kB)
Collecting attrs>=17.3.0 (from aiohttp->datasets)
Using cached attrs-25.3.0-py3-none-any.whl.metadata (10 kB)
Collecting frozenlist>=1.1.1 (from aiohttp->datasets)
Using cached frozenlist-1.6.0-cp313-cp313t-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (16 kB)
Collecting multidict<7.0,>=4.5 (from aiohttp->datasets)
Using cached multidict-6.4.3-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.3 kB)
Collecting propcache>=0.2.0 (from aiohttp->datasets)
Using cached propcache-0.3.1-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (10 kB)
Collecting yarl<2.0,>=1.17.0 (from aiohttp->datasets)
Using cached yarl-1.20.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (72 kB)
Requirement already satisfied: idna>=2.0 in /root/vm313t/lib/python3.13t/site-packages (from yarl<2.0,>=1.17.0->aiohttp->datasets) (3.10)
Requirement already satisfied: typing-extensions>=3.7.4.3 in /root/vm313t/lib/python3.13t/site-packages (from huggingface-hub>=0.24.0->datasets) (4.13.2)
Requirement already satisfied: charset-normalizer<4,>=2 in /root/vm313t/lib/python3.13t/site-packages (from requests>=2.32.2->datasets) (3.4.1)
Requirement already satisfied: urllib3<3,>=1.21.1 in /root/vm313t/lib/python3.13t/site-packages (from requests>=2.32.2->datasets) (2.4.0)
Requirement already satisfied: certifi>=2017.4.17 in /root/vm313t/lib/python3.13t/site-packages (from requests>=2.32.2->datasets) (2025.4.26)
Collecting python-dateutil>=2.8.2 (from pandas->datasets)
Using cached python_dateutil-2.9.0.post0-py2.py3-none-any.whl.metadata (8.4 kB)
Collecting pytz>=2020.1 (from pandas->datasets)
Using cached pytz-2025.2-py2.py3-none-any.whl.metadata (22 kB)
Collecting tzdata>=2022.7 (from pandas->datasets)
Using cached tzdata-2025.2-py2.py3-none-any.whl.metadata (1.4 kB)
Collecting six>=1.5 (from python-dateutil>=2.8.2->pandas->datasets)
Using cached six-1.17.0-py2.py3-none-any.whl.metadata (1.7 kB)
Using cached datasets-3.5.1-py3-none-any.whl (491 kB)
Using cached dill-0.3.8-py3-none-any.whl (116 kB)
Using cached fsspec-2025.3.0-py3-none-any.whl (193 kB)
Using cached multiprocess-0.70.16-py312-none-any.whl (146 kB)
Using cached multidict-6.4.3-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (220 kB)
Using cached yarl-1.20.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (404 kB)
Using cached aiohappyeyeballs-2.6.1-py3-none-any.whl (15 kB)
Using cached aiosignal-1.3.2-py2.py3-none-any.whl (7.6 kB)
Using cached attrs-25.3.0-py3-none-any.whl (63 kB)
Using cached frozenlist-1.6.0-cp313-cp313t-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (385 kB)
Using cached propcache-0.3.1-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (282 kB)
Using cached pyarrow-20.0.0-cp313-cp313t-manylinux_2_28_x86_64.whl (42.2 MB)
Using cached pandas-2.2.3-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (11.9 MB)
Using cached python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB)
Using cached pytz-2025.2-py2.py3-none-any.whl (509 kB)
Using cached six-1.17.0-py2.py3-none-any.whl (11 kB)
Using cached tzdata-2025.2-py2.py3-none-any.whl (347 kB)
Building wheels for collected packages: aiohttp
Building wheel for aiohttp (pyproject.toml) ... error
error: subprocess-exited-with-error
Γ Building wheel for aiohttp (pyproject.toml) did not run successfully.
β exit code: 1
β°β> [156 lines of output]
*********************
* Accelerated build *
*********************
/tmp/pip-build-env-wjqi8_7w/overlay/lib/python3.13t/site-packages/setuptools/dist.py:759: SetuptoolsDeprecationWarning: License classifiers are deprecated.
!!
********************************************************************************
Please consider removing the following classifiers in favor of a SPDX license expression:
License :: OSI Approved :: Apache Software License
See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details.
********************************************************************************
!!
self._finalize_license_expression()
running bdist_wheel
running build
running build_py
creating build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/typedefs.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/http_parser.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/client_reqrep.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/client_ws.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_app.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/http_websocket.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/resolver.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/tracing.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/http_writer.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/http_exceptions.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/log.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/__init__.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_runner.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/worker.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/connector.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/client_exceptions.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_middlewares.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/tcp_helpers.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_response.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_server.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_request.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_urldispatcher.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_exceptions.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/formdata.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/streams.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/multipart.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_routedef.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_ws.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/payload.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/client_proto.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_log.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/base_protocol.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/payload_streamer.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/http.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_fileresponse.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/test_utils.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/client.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/cookiejar.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/compression_utils.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/hdrs.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/helpers.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/pytest_plugin.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/web_protocol.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/abc.py -> build/lib.linux-x86_64-cpython-313t/aiohttp
creating build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/__init__.py -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/writer.py -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/models.py -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/reader.py -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/reader_c.py -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/helpers.py -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/reader_py.py -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
running egg_info
writing aiohttp.egg-info/PKG-INFO
writing dependency_links to aiohttp.egg-info/dependency_links.txt
writing requirements to aiohttp.egg-info/requires.txt
writing top-level names to aiohttp.egg-info/top_level.txt
reading manifest file 'aiohttp.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
warning: no files found matching 'aiohttp' anywhere in distribution
warning: no files found matching '*.pyi' anywhere in distribution
warning: no previously-included files matching '*.pyc' found anywhere in distribution
warning: no previously-included files matching '*.pyd' found anywhere in distribution
warning: no previously-included files matching '*.so' found anywhere in distribution
warning: no previously-included files matching '*.lib' found anywhere in distribution
warning: no previously-included files matching '*.dll' found anywhere in distribution
warning: no previously-included files matching '*.a' found anywhere in distribution
warning: no previously-included files matching '*.obj' found anywhere in distribution
warning: no previously-included files found matching 'aiohttp/*.html'
no previously-included directories found matching 'docs/_build'
adding license file 'LICENSE.txt'
writing manifest file 'aiohttp.egg-info/SOURCES.txt'
copying aiohttp/_cparser.pxd -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/_find_header.pxd -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/_headers.pxi -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/_http_parser.pyx -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/_http_writer.pyx -> build/lib.linux-x86_64-cpython-313t/aiohttp
copying aiohttp/py.typed -> build/lib.linux-x86_64-cpython-313t/aiohttp
creating build/lib.linux-x86_64-cpython-313t/aiohttp/.hash
copying aiohttp/.hash/_cparser.pxd.hash -> build/lib.linux-x86_64-cpython-313t/aiohttp/.hash
copying aiohttp/.hash/_find_header.pxd.hash -> build/lib.linux-x86_64-cpython-313t/aiohttp/.hash
copying aiohttp/.hash/_http_parser.pyx.hash -> build/lib.linux-x86_64-cpython-313t/aiohttp/.hash
copying aiohttp/.hash/_http_writer.pyx.hash -> build/lib.linux-x86_64-cpython-313t/aiohttp/.hash
copying aiohttp/.hash/hdrs.py.hash -> build/lib.linux-x86_64-cpython-313t/aiohttp/.hash
copying aiohttp/_websocket/mask.pxd -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/mask.pyx -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
copying aiohttp/_websocket/reader_c.pxd -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket
creating build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket/.hash
copying aiohttp/_websocket/.hash/mask.pxd.hash -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket/.hash
copying aiohttp/_websocket/.hash/mask.pyx.hash -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket/.hash
copying aiohttp/_websocket/.hash/reader_c.pxd.hash -> build/lib.linux-x86_64-cpython-313t/aiohttp/_websocket/.hash
running build_ext
building 'aiohttp._websocket.mask' extension
creating build/temp.linux-x86_64-cpython-313t/aiohttp/_websocket
x86_64-linux-gnu-gcc -fno-strict-overflow -Wsign-compare -DNDEBUG -g -O2 -Wall -g -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -fPIC -I/root/vm313t/include -I/usr/include/python3.13t -c aiohttp/_websocket/mask.c -o build/temp.linux-x86_64-cpython-313t/aiohttp/_websocket/mask.o
aiohttp/_websocket/mask.c:1864:80: error: unknown type name β__pyx_vectorcallfuncβ; did you mean βvectorcallfuncβ?
1864 | static CYTHON_INLINE PyObject *__Pyx_PyVectorcall_FastCallDict(PyObject *func, __pyx_vectorcallfunc vc, PyObject *const *args, size_t nargs, PyObject *kw);
| ^~~~~~~~~~~~~~~~~~~~
| vectorcallfunc
aiohttp/_websocket/mask.c: In function β__pyx_f_7aiohttp_10_websocket_4mask__websocket_mask_cythonβ:
aiohttp/_websocket/mask.c:2905:3: warning: βPy_OptimizeFlagβ is deprecated [-Wdeprecated-declarations]
2905 | if (unlikely(__pyx_assertions_enabled())) {
| ^~
In file included from /usr/include/python3.13t/Python.h:76,
from aiohttp/_websocket/mask.c:16:
/usr/include/python3.13t/cpython/pydebug.h:13:37: note: declared here
13 | Py_DEPRECATED(3.12) PyAPI_DATA(int) Py_OptimizeFlag;
| ^~~~~~~~~~~~~~~
aiohttp/_websocket/mask.c: At top level:
aiohttp/_websocket/mask.c:4846:69: error: unknown type name β__pyx_vectorcallfuncβ; did you mean βvectorcallfuncβ?
4846 | static PyObject *__Pyx_PyVectorcall_FastCallDict_kw(PyObject *func, __pyx_vectorcallfunc vc, PyObject *const *args, size_t nargs, PyObject *kw)
| ^~~~~~~~~~~~~~~~~~~~
| vectorcallfunc
aiohttp/_websocket/mask.c:4891:80: error: unknown type name β__pyx_vectorcallfuncβ; did you mean βvectorcallfuncβ?
4891 | static CYTHON_INLINE PyObject *__Pyx_PyVectorcall_FastCallDict(PyObject *func, __pyx_vectorcallfunc vc, PyObject *const *args, size_t nargs, PyObject *kw)
| ^~~~~~~~~~~~~~~~~~~~
| vectorcallfunc
aiohttp/_websocket/mask.c: In function β__Pyx_CyFunction_CallAsMethodβ:
aiohttp/_websocket/mask.c:5580:6: error: unknown type name β__pyx_vectorcallfuncβ; did you mean βvectorcallfuncβ?
5580 | __pyx_vectorcallfunc vc = __Pyx_CyFunction_func_vectorcall(cyfunc);
| ^~~~~~~~~~~~~~~~~~~~
| vectorcallfunc
aiohttp/_websocket/mask.c:1954:45: warning: initialization of βintβ from βvectorcallfuncβ {aka βstruct _object * (*)(struct _object *, struct _object * const*, long unsigned int, struct _object *)β} makes integer from pointer without a cast [-Wint-conversion]
1954 | #define __Pyx_CyFunction_func_vectorcall(f) (((PyCFunctionObject*)f)->vectorcall)
| ^
aiohttp/_websocket/mask.c:5580:32: note: in expansion of macro β__Pyx_CyFunction_func_vectorcallβ
5580 | __pyx_vectorcallfunc vc = __Pyx_CyFunction_func_vectorcall(cyfunc);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
aiohttp/_websocket/mask.c:5583:16: warning: implicit declaration of function β__Pyx_PyVectorcall_FastCallDictβ [-Wimplicit-function-declaration]
5583 | return __Pyx_PyVectorcall_FastCallDict(func, vc, &PyTuple_GET_ITEM(args, 0), (size_t)PyTuple_GET_SIZE(args), kw);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
aiohttp/_websocket/mask.c:5583:16: warning: returning βintβ from a function with return type βPyObject *β {aka βstruct _object *β} makes pointer from integer without a cast [-Wint-conversion]
5583 | return __Pyx_PyVectorcall_FastCallDict(func, vc, &PyTuple_GET_ITEM(args, 0), (size_t)PyTuple_GET_SIZE(args), kw);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
error: command '/usr/bin/x86_64-linux-gnu-gcc' failed with exit code 1
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for aiohttp
Failed to build aiohttp
ERROR: Failed to build installable wheels for some pyproject.toml based projects (aiohttp)
```
### Steps to reproduce the bug
See above
### Expected behavior
Install
### Environment info
Ubuntu 24.04 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7548/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7548/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7547 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7547/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7547/comments | https://api.github.com/repos/huggingface/datasets/issues/7547/events | https://github.com/huggingface/datasets/pull/7547 | 3,034,830,291 | PR_kwDODunzps6UsTuF | 7,547 | Avoid global umask for setting file mode. | {
"avatar_url": "https://avatars.githubusercontent.com/u/1282383?v=4",
"events_url": "https://api.github.com/users/ryan-clancy/events{/privacy}",
"followers_url": "https://api.github.com/users/ryan-clancy/followers",
"following_url": "https://api.github.com/users/ryan-clancy/following{/other_user}",
"gists_url": "https://api.github.com/users/ryan-clancy/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ryan-clancy",
"id": 1282383,
"login": "ryan-clancy",
"node_id": "MDQ6VXNlcjEyODIzODM=",
"organizations_url": "https://api.github.com/users/ryan-clancy/orgs",
"received_events_url": "https://api.github.com/users/ryan-clancy/received_events",
"repos_url": "https://api.github.com/users/ryan-clancy/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ryan-clancy/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ryan-clancy/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ryan-clancy",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7547). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-05-01T22:24:24Z | 2025-05-06T13:05:00Z | 2025-05-06T13:05:00Z | CONTRIBUTOR | null | null | null | This PR updates the method for setting the permissions on `cache_path` after calling `shutil.move`. The call to `shutil.move` may not preserve permissions if the source and destination are on different filesystems. Reading and resetting umask can cause race conditions, so directly read what permissions were set for the `temp_file` instead.
This fixes https://github.com/huggingface/datasets/issues/7536. | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7547/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7547/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7547.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7547",
"merged_at": "2025-05-06T13:05:00Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7547.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7547"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7546 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7546/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7546/comments | https://api.github.com/repos/huggingface/datasets/issues/7546/events | https://github.com/huggingface/datasets/issues/7546 | 3,034,018,298 | I_kwDODunzps6013H6 | 7,546 | Large memory use when loading large datasets to a ZFS pool | {
"avatar_url": "https://avatars.githubusercontent.com/u/6875946?v=4",
"events_url": "https://api.github.com/users/FredHaa/events{/privacy}",
"followers_url": "https://api.github.com/users/FredHaa/followers",
"following_url": "https://api.github.com/users/FredHaa/following{/other_user}",
"gists_url": "https://api.github.com/users/FredHaa/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/FredHaa",
"id": 6875946,
"login": "FredHaa",
"node_id": "MDQ6VXNlcjY4NzU5NDY=",
"organizations_url": "https://api.github.com/users/FredHaa/orgs",
"received_events_url": "https://api.github.com/users/FredHaa/received_events",
"repos_url": "https://api.github.com/users/FredHaa/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/FredHaa/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/FredHaa/subscriptions",
"type": "User",
"url": "https://api.github.com/users/FredHaa",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! datasets are memory mapped from disk, so they don't fill out your RAM. Not sure what's the source of your memory issue.\n\nWhat kind of system are you using ? and what kind of disk ?",
"Well, the fact of the matter is that my RAM is getting filled out by running the given example, as shown in [this video](https://streamable.com/usb0ql).\n\nMy system is a GPU server running Ubuntu. The disk is a SATA SSD attached to the server using a backplane. It is formatted with ZFS, mounted in /cache, and my HF_HOME is set to /cache/hf\n\nI really need this fixed, so I am more than willing to test out various suggestions you might have, or write a PR if we can figure out what is going on.",
"I'm not super familiar with ZFS, but it looks like it loads the data in memory when the files are memory mapped, which is an issue.\n\nMaybe it's a caching mechanism ? Since `datasets` accesses every memory mapped file to read a small part (the metadata of the arrow record batches), maybe ZFS brings the whole files in memory for quicker subsequent reads. This is an antipattern when it comes to lazy loading datasets of that size though",
"This is the answer.\n\nI tried changing my HF_HOME to an NFS share, and no RAM is then consumed loading the dataset.\n\nI will try to see if I can find a way to configure the ZFS pool to not cache the files (disabling the ARC/primary cache didn't work), and if I do write the solution in this issue. If I can't I guess I have to reformat my cache drive."
] | 2025-05-01T14:43:47Z | 2025-05-13T13:30:09Z | 2025-05-13T13:29:53Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When I load large parquet based datasets from the hub like `MLCommons/peoples_speech` using `load_dataset`, all my memory (500GB) is used and isn't released after loading, meaning that the process is terminated by the kernel if I try to load an additional dataset. This makes it impossible to train models using multiple large datasets.
### Steps to reproduce the bug
`uv run --with datasets==3.5.1 python`
```python
from datasets import load_dataset
load_dataset('MLCommons/peoples_speech', 'clean')
load_dataset('mozilla-foundation/common_voice_17_0', 'en')
```
### Expected behavior
I would expect that a lot less than 500GB of RAM would be required to load the dataset, or at least that the RAM usage would be cleared as soon as the dataset is loaded (and thus reside as a memory mapped file) such that other datasets can be loaded.
### Environment info
I am currently using the latest datasets==3.5.1 but I have had the same problem with multiple other versions. | {
"avatar_url": "https://avatars.githubusercontent.com/u/6875946?v=4",
"events_url": "https://api.github.com/users/FredHaa/events{/privacy}",
"followers_url": "https://api.github.com/users/FredHaa/followers",
"following_url": "https://api.github.com/users/FredHaa/following{/other_user}",
"gists_url": "https://api.github.com/users/FredHaa/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/FredHaa",
"id": 6875946,
"login": "FredHaa",
"node_id": "MDQ6VXNlcjY4NzU5NDY=",
"organizations_url": "https://api.github.com/users/FredHaa/orgs",
"received_events_url": "https://api.github.com/users/FredHaa/received_events",
"repos_url": "https://api.github.com/users/FredHaa/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/FredHaa/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/FredHaa/subscriptions",
"type": "User",
"url": "https://api.github.com/users/FredHaa",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7546/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7546/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7545 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7545/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7545/comments | https://api.github.com/repos/huggingface/datasets/issues/7545/events | https://github.com/huggingface/datasets/issues/7545 | 3,031,617,547 | I_kwDODunzps60stAL | 7,545 | Networked Pull Through Cache | {
"avatar_url": "https://avatars.githubusercontent.com/u/8764173?v=4",
"events_url": "https://api.github.com/users/wrmedford/events{/privacy}",
"followers_url": "https://api.github.com/users/wrmedford/followers",
"following_url": "https://api.github.com/users/wrmedford/following{/other_user}",
"gists_url": "https://api.github.com/users/wrmedford/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/wrmedford",
"id": 8764173,
"login": "wrmedford",
"node_id": "MDQ6VXNlcjg3NjQxNzM=",
"organizations_url": "https://api.github.com/users/wrmedford/orgs",
"received_events_url": "https://api.github.com/users/wrmedford/received_events",
"repos_url": "https://api.github.com/users/wrmedford/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/wrmedford/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/wrmedford/subscriptions",
"type": "User",
"url": "https://api.github.com/users/wrmedford",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | [] | null | [] | 2025-04-30T15:16:33Z | 2025-04-30T15:16:33Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
Introduce a HF_DATASET_CACHE_NETWORK_LOCATION configuration (e.g. an environment variable) together with a companion network cache service.
Enable a three-tier cache lookup for datasets:
1. Local on-disk cache
2. Configurable network cache proxy
3. Official Hugging Face Hub
### Motivation
- Distributed training & ephemeral jobs: In high-performance or containerized clusters, relying solely on a local disk cache either becomes a streaming bottleneck or incurs a heavy cold-start penalty as each job must re-download datasets.
- Traffic & cost reduction: A pull-through network cache lets multiple consumers share a common cache layer, reducing duplicate downloads from the Hub and lowering egress costs.
- Better streaming adoption: By offloading repeat dataset pulls to a locally managed cache proxy, streaming workloads can achieve higher throughput and more predictable latency.
- Proven pattern: Similar proxy-cache solutions (e.g. Harborβs Proxy Cache for Docker images) have demonstrated reliability and performance at scale: https://goharbor.io/docs/2.1.0/administration/configure-proxy-cache/
### Your contribution
Iβm happy to draft the initial PR for adding HF_DATASET_CACHE_NETWORK_LOCATION support in datasets and sketch out a minimal cache-service prototype.
I have limited bandwidth so I would be looking for collaborators if anyone else is interested. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7545/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7545/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7544 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7544/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7544/comments | https://api.github.com/repos/huggingface/datasets/issues/7544/events | https://github.com/huggingface/datasets/pull/7544 | 3,027,024,285 | PR_kwDODunzps6UR4Nn | 7,544 | Add try_original_type to DatasetDict.map | {
"avatar_url": "https://avatars.githubusercontent.com/u/11156001?v=4",
"events_url": "https://api.github.com/users/yoshitomo-matsubara/events{/privacy}",
"followers_url": "https://api.github.com/users/yoshitomo-matsubara/followers",
"following_url": "https://api.github.com/users/yoshitomo-matsubara/following{/other_user}",
"gists_url": "https://api.github.com/users/yoshitomo-matsubara/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yoshitomo-matsubara",
"id": 11156001,
"login": "yoshitomo-matsubara",
"node_id": "MDQ6VXNlcjExMTU2MDAx",
"organizations_url": "https://api.github.com/users/yoshitomo-matsubara/orgs",
"received_events_url": "https://api.github.com/users/yoshitomo-matsubara/received_events",
"repos_url": "https://api.github.com/users/yoshitomo-matsubara/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yoshitomo-matsubara/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yoshitomo-matsubara/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yoshitomo-matsubara",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7544). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Sure! I just committed the changes",
"@lhoestq \r\nLet me know if there are other things to do before merge or other places to add `try_original_type` argument "
] | 2025-04-29T04:39:44Z | 2025-05-05T14:42:49Z | 2025-05-05T14:42:49Z | CONTRIBUTOR | null | null | null | This PR resolves #7472 for DatasetDict
The previously merged PR #7483 added `try_original_type` to ArrowDataset, but DatasetDict misses `try_original_type`
Cc: @lhoestq | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7544/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7544/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7544.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7544",
"merged_at": "2025-05-05T14:42:49Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7544.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7544"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7543 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7543/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7543/comments | https://api.github.com/repos/huggingface/datasets/issues/7543/events | https://github.com/huggingface/datasets/issues/7543 | 3,026,867,706 | I_kwDODunzps60alX6 | 7,543 | The memory-disk mapping failure issue of the map functionοΌresolved, but there are some suggestions.οΌ | {
"avatar_url": "https://avatars.githubusercontent.com/u/76415358?v=4",
"events_url": "https://api.github.com/users/jxma20/events{/privacy}",
"followers_url": "https://api.github.com/users/jxma20/followers",
"following_url": "https://api.github.com/users/jxma20/following{/other_user}",
"gists_url": "https://api.github.com/users/jxma20/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jxma20",
"id": 76415358,
"login": "jxma20",
"node_id": "MDQ6VXNlcjc2NDE1MzU4",
"organizations_url": "https://api.github.com/users/jxma20/orgs",
"received_events_url": "https://api.github.com/users/jxma20/received_events",
"repos_url": "https://api.github.com/users/jxma20/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jxma20/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jxma20/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jxma20",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [] | 2025-04-29T03:04:59Z | 2025-04-30T02:22:17Z | 2025-04-30T02:22:17Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
## bug
When the map function processes a large dataset, it temporarily stores the data in a cache file on the disk. After the data is stored, the memory occupied by it is released. Therefore, when using the map function to process a large-scale dataset, only a dataset space of the size of `writer_batch_size` will be occupied in memory.
However, I found that the map function does not actually reduce memory usage when I used it. At first, I thought there was a bug in the program, causing a memory leakβmeaning the memory was not released after the data was stored in the cache. But later, I used a Linux command to check for recently modified files during program execution and found that no new files were created or modified. This indicates that the program did not store the dataset in the disk cache.
## bug solved
After modifying the parameters of the map function multiple times, I discovered the `cache_file_name` parameter. By changing it, the cache file can be stored in the specified directory. After making this change, I noticed that the cache file appeared. Initially, I found this quite incredible, but then I wondered if the cache file might have failed to be stored in a certain folder. This could be related to the fact that I don't have root privileges.
So, I delved into the source code of the map function to find out where the cache file would be stored by default. Eventually, I found the function `def _get_cache_file_path(self, fingerprint):`, which automatically generates the storage path for the cache file. The output was as follows: `/tmp/hf_datasets-j5qco9ug/cache-f2830487643b9cc2.arrow`. My hypothesis was confirmed: the lack of root privileges indeed prevented the cache file from being stored, which in turn prevented the release of memory. Therefore, changing the storage location to a folder where I have write access resolved the issue.
### Steps to reproduce the bug
my code
`train_data = train_data.map(process_fun, remove_columns=['image_name', 'question_type', 'concern', 'question', 'candidate_answers', 'answer'])`
### Expected behavior
Although my bug has been resolved, it still took me nearly a week to search for relevant information and debug the program. However, if a warning or error message about insufficient cache file write permissions could be provided during program execution, I might have been able to identify the cause more quickly. Therefore, I hope this aspect can be improved. I am documenting this bug here so that friends who encounter similar issues can solve their problems in a timely manner.
### Environment info
python: 3.10.15
datasets: 3.5.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/76415358?v=4",
"events_url": "https://api.github.com/users/jxma20/events{/privacy}",
"followers_url": "https://api.github.com/users/jxma20/followers",
"following_url": "https://api.github.com/users/jxma20/following{/other_user}",
"gists_url": "https://api.github.com/users/jxma20/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jxma20",
"id": 76415358,
"login": "jxma20",
"node_id": "MDQ6VXNlcjc2NDE1MzU4",
"organizations_url": "https://api.github.com/users/jxma20/orgs",
"received_events_url": "https://api.github.com/users/jxma20/received_events",
"repos_url": "https://api.github.com/users/jxma20/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jxma20/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jxma20/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jxma20",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7543/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7543/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7542 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7542/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7542/comments | https://api.github.com/repos/huggingface/datasets/issues/7542/events | https://github.com/huggingface/datasets/pull/7542 | 3,025,054,630 | PR_kwDODunzps6ULHxo | 7,542 | set dev version | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7542). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-04-28T14:03:48Z | 2025-04-28T14:08:37Z | 2025-04-28T14:04:00Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7542/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7542/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7542.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7542",
"merged_at": "2025-04-28T14:04:00Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7542.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7542"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7541 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7541/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7541/comments | https://api.github.com/repos/huggingface/datasets/issues/7541/events | https://github.com/huggingface/datasets/pull/7541 | 3,025,045,919 | PR_kwDODunzps6ULF7F | 7,541 | release: 3.5.1 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7541). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-04-28T14:00:59Z | 2025-04-28T14:03:38Z | 2025-04-28T14:01:54Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7541/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7541/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7541.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7541",
"merged_at": "2025-04-28T14:01:54Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7541.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7541"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7540 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7540/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7540/comments | https://api.github.com/repos/huggingface/datasets/issues/7540/events | https://github.com/huggingface/datasets/pull/7540 | 3,024,862,966 | PR_kwDODunzps6UKe6T | 7,540 | support pyarrow 20 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7540). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-04-28T13:01:11Z | 2025-04-28T13:23:53Z | 2025-04-28T13:23:52Z | MEMBER | null | null | null | fix
```
TypeError: ArrayExtensionArray.to_pylist() got an unexpected keyword argument 'maps_as_pydicts'
``` | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7540/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7540/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7540.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7540",
"merged_at": "2025-04-28T13:23:52Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7540.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7540"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7539 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7539/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7539/comments | https://api.github.com/repos/huggingface/datasets/issues/7539/events | https://github.com/huggingface/datasets/pull/7539 | 3,023,311,163 | PR_kwDODunzps6UFQ0W | 7,539 | Fix IterableDataset state_dict shard_example_idx counting | {
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7539). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Hi ! FYI I made a PR to fix https://github.com/huggingface/datasets/issues/7538 and it also fixed https://github.com/huggingface/datasets/issues/7475, so if I'm not mistaken this PR is not needed anymore"
] | 2025-04-27T20:41:18Z | 2025-05-06T14:24:25Z | 2025-05-06T14:24:24Z | CONTRIBUTOR | null | null | null | # Fix IterableDataset's state_dict shard_example_idx reporting
## Description
This PR fixes issue #7475 where the `shard_example_idx` value in `IterableDataset`'s `state_dict()` always equals the number of samples in a shard, even if only a few examples have been consumed.
The issue is in the `_iter_arrow` method of the `ArrowExamplesIterable` class where it updates the `shard_example_idx` state by the full length of the batch (`len(pa_table)`) even when we're only partway through processing the examples.
## Changes
Modified the `_iter_arrow` method of `ArrowExamplesIterable` to:
1. Track the actual number of examples processed
2. Only increment the `shard_example_idx` by the number of examples actually yielded
3. Handle partial batches correctly
## How to Test
I've included a simple test case that demonstrates the fix:
```python
from datasets import Dataset
# Create a test dataset
ds = Dataset.from_dict({"a": range(6)}).to_iterable_dataset(num_shards=1)
# Iterate through part of the dataset
for idx, example in enumerate(ds):
print(example)
if idx == 2: # Stop after 3 examples (0, 1, 2)
state_dict = ds.state_dict()
print("Checkpoint state_dict:", state_dict)
break
# Before the fix, the output would show shard_example_idx: 6
# After the fix, it shows shard_example_idx: 3, correctly reflecting the 3 processed examples
```
## Implementation Details
1. Added logic to track the number of examples actually seen in the current shard
2. Modified the state update to only count examples actually yielded
3. Improved handling of partial batches and skipped examples
This fix ensures that checkpointing and resuming works correctly with exactly the expected number of examples, rather than skipping ahead to the end of the batch. | {
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7539/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7539/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7539.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7539",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7539.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7539"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7538 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7538/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7538/comments | https://api.github.com/repos/huggingface/datasets/issues/7538/events | https://github.com/huggingface/datasets/issues/7538 | 3,023,280,056 | I_kwDODunzps60M5e4 | 7,538 | `IterableDataset` drops samples when resuming from a checkpoint | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mariosasko",
"id": 47462742,
"login": "mariosasko",
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"organizations_url": "https://api.github.com/users/mariosasko/orgs",
"received_events_url": "https://api.github.com/users/mariosasko/received_events",
"repos_url": "https://api.github.com/users/mariosasko/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mariosasko",
"user_view_type": "public"
} | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"Thanks for reporting ! I fixed the issue using RebatchedArrowExamplesIterable before the formatted iterable"
] | 2025-04-27T19:34:49Z | 2025-05-06T14:04:05Z | 2025-05-06T14:03:42Z | COLLABORATOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | When resuming from a checkpoint, `IterableDataset` will drop samples if `num_shards % world_size == 0` and the underlying example supports `iter_arrow` and needs to be formatted.
In that case, the `FormattedExamplesIterable` fetches a batch of samples from the child iterable's `iter_arrow` and yields them one by one (after formatting). However, the child increments the `shard_example_idx` counter (in its `iter_arrow`) before returning the batch for the whole batch size, which leads to a portion of samples being skipped if the iteration (of the parent iterable) is stopped mid-batch.
Perhaps one way to avoid this would be by signalling the child iterable which samples (within the chunk) are processed by the parent and which are not, so that it can adjust the `shard_example_idx` counter accordingly. This would also mean the chunk needs to be sliced when resuming, but this is straightforward to implement.
The following is a minimal reproducer of the bug:
```python
from datasets import Dataset
from datasets.distributed import split_dataset_by_node
ds = Dataset.from_dict({"n": list(range(24))})
ds = ds.to_iterable_dataset(num_shards=4)
world_size = 4
rank = 0
ds_rank = split_dataset_by_node(ds, rank, world_size)
it = iter(ds_rank)
examples = []
for idx, example in enumerate(it):
examples.append(example)
if idx == 2:
state_dict = ds_rank.state_dict()
break
ds_rank.load_state_dict(state_dict)
it_resumed = iter(ds_rank)
examples_resumed = examples[:]
for example in it:
examples.append(example)
for example in it_resumed:
examples_resumed.append(example)
print("ORIGINAL ITER EXAMPLES:", examples)
print("RESUMED ITER EXAMPLES:", examples_resumed)
``` | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7538/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7538/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7537 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7537/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7537/comments | https://api.github.com/repos/huggingface/datasets/issues/7537/events | https://github.com/huggingface/datasets/issues/7537 | 3,018,792,966 | I_kwDODunzps6z7yAG | 7,537 | `datasets.map(..., num_proc=4)` multi-processing fails | {
"avatar_url": "https://avatars.githubusercontent.com/u/24477841?v=4",
"events_url": "https://api.github.com/users/faaany/events{/privacy}",
"followers_url": "https://api.github.com/users/faaany/followers",
"following_url": "https://api.github.com/users/faaany/following{/other_user}",
"gists_url": "https://api.github.com/users/faaany/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/faaany",
"id": 24477841,
"login": "faaany",
"node_id": "MDQ6VXNlcjI0NDc3ODQx",
"organizations_url": "https://api.github.com/users/faaany/orgs",
"received_events_url": "https://api.github.com/users/faaany/received_events",
"repos_url": "https://api.github.com/users/faaany/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/faaany/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/faaany/subscriptions",
"type": "User",
"url": "https://api.github.com/users/faaany",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"related: https://github.com/huggingface/datasets/issues/7510\n\nwe need to do more tests to see if latest `dill` is deterministic"
] | 2025-04-25T01:53:47Z | 2025-05-06T13:12:08Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | The following code fails in python 3.11+
```python
tokenized_datasets = datasets.map(tokenize_function, batched=True, num_proc=4, remove_columns=["text"])
```
Error log:
```bash
Traceback (most recent call last):
File "/usr/local/lib/python3.12/dist-packages/multiprocess/process.py", line 315, in _bootstrap
self.run()
File "/usr/local/lib/python3.12/dist-packages/multiprocess/process.py", line 108, in run
self._target(*self._args, **self._kwargs)
File "/usr/local/lib/python3.12/dist-packages/multiprocess/pool.py", line 114, in worker
task = get()
^^^^^
File "/usr/local/lib/python3.12/dist-packages/multiprocess/queues.py", line 371, in get
return _ForkingPickler.loads(res)
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/dill/_dill.py", line 327, in loads
return load(file, ignore, **kwds)
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/dill/_dill.py", line 313, in load
return Unpickler(file, ignore=ignore, **kwds).load()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/dill/_dill.py", line 525, in load
obj = StockUnpickler.load(self)
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/dill/_dill.py", line 659, in _create_code
if len(args) == 16: return CodeType(*args)
^^^^^^^^^^^^^^^
TypeError: code() argument 13 must be str, not int
```
After upgrading dill to the latest 0.4.0 with "pip install --upgrade dill", it can pass. So it seems that there is a compatibility issue between dill 0.3.4 and python 3.11+, because python 3.10 works fine.
Is the dill deterministic issue mentioned in https://github.com/huggingface/datasets/blob/main/setup.py#L117) still valid? Any plan to unpin?
| null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7537/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7537/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7536 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7536/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7536/comments | https://api.github.com/repos/huggingface/datasets/issues/7536/events | https://github.com/huggingface/datasets/issues/7536 | 3,018,425,549 | I_kwDODunzps6z6YTN | 7,536 | [Errno 13] Permission denied: on `.incomplete` file | {
"avatar_url": "https://avatars.githubusercontent.com/u/1282383?v=4",
"events_url": "https://api.github.com/users/ryan-clancy/events{/privacy}",
"followers_url": "https://api.github.com/users/ryan-clancy/followers",
"following_url": "https://api.github.com/users/ryan-clancy/following{/other_user}",
"gists_url": "https://api.github.com/users/ryan-clancy/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ryan-clancy",
"id": 1282383,
"login": "ryan-clancy",
"node_id": "MDQ6VXNlcjEyODIzODM=",
"organizations_url": "https://api.github.com/users/ryan-clancy/orgs",
"received_events_url": "https://api.github.com/users/ryan-clancy/received_events",
"repos_url": "https://api.github.com/users/ryan-clancy/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ryan-clancy/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ryan-clancy/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ryan-clancy",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"It must be an issue with umask being used by multiple threads indeed. Maybe we can try to make a thread safe function to apply the umask (using filelock for example)",
"> It must be an issue with umask being used by multiple threads indeed. Maybe we can try to make a thread safe function to apply the umask (using filelock for example)\n\n@lhoestq is this something which can go in a 3.5.1 release?",
"Yes for sure",
"@lhoestq - can you take a look at https://github.com/huggingface/datasets/pull/7547/?"
] | 2025-04-24T20:52:45Z | 2025-05-06T13:05:01Z | 2025-05-06T13:05:01Z | CONTRIBUTOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When downloading a dataset, we frequently hit the below Permission Denied error. This looks to happen (at least) across datasets in HF, S3, and GCS.
It looks like the `temp_file` being passed [here](https://github.com/huggingface/datasets/blob/main/src/datasets/utils/file_utils.py#L412) can sometimes be created with `000` permissions leading to the permission denied error (the user running the code is still the owner of the file). Deleting that particular file and re-running the code with 0 changes will usually succeed.
Is there some race condition happening with the [umask](https://github.com/huggingface/datasets/blob/main/src/datasets/utils/file_utils.py#L416), which is process global, and the [file creation](https://github.com/huggingface/datasets/blob/main/src/datasets/utils/file_utils.py#L404)?
```
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.venv/lib/python3.12/site-packages/datasets/load.py:2084: in load_dataset
builder_instance.download_and_prepare(
.venv/lib/python3.12/site-packages/datasets/builder.py:925: in download_and_prepare
self._download_and_prepare(
.venv/lib/python3.12/site-packages/datasets/builder.py:1649: in _download_and_prepare
super()._download_and_prepare(
.venv/lib/python3.12/site-packages/datasets/builder.py:979: in _download_and_prepare
split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
.venv/lib/python3.12/site-packages/datasets/packaged_modules/folder_based_builder/folder_based_builder.py:120: in _split_generators
downloaded_files = dl_manager.download(files)
.venv/lib/python3.12/site-packages/datasets/download/download_manager.py:159: in download
downloaded_path_or_paths = map_nested(
.venv/lib/python3.12/site-packages/datasets/utils/py_utils.py:514: in map_nested
_single_map_nested((function, obj, batched, batch_size, types, None, True, None))
.venv/lib/python3.12/site-packages/datasets/utils/py_utils.py:382: in _single_map_nested
return [mapped_item for batch in iter_batched(data_struct, batch_size) for mapped_item in function(batch)]
.venv/lib/python3.12/site-packages/datasets/download/download_manager.py:206: in _download_batched
return thread_map(
.venv/lib/python3.12/site-packages/tqdm/contrib/concurrent.py:69: in thread_map
return _executor_map(ThreadPoolExecutor, fn, *iterables, **tqdm_kwargs)
.venv/lib/python3.12/site-packages/tqdm/contrib/concurrent.py:51: in _executor_map
return list(tqdm_class(ex.map(fn, *iterables, chunksize=chunksize), **kwargs))
.venv/lib/python3.12/site-packages/tqdm/std.py:1181: in __iter__
for obj in iterable:
../../../_tool/Python/3.12.10/x64/lib/python3.12/concurrent/futures/_base.py:619: in result_iterator
yield _result_or_cancel(fs.pop())
../../../_tool/Python/3.12.10/x64/lib/python3.12/concurrent/futures/_base.py:317: in _result_or_cancel
return fut.result(timeout)
../../../_tool/Python/3.12.10/x64/lib/python3.12/concurrent/futures/_base.py:449: in result
return self.__get_result()
../../../_tool/Python/3.12.10/x64/lib/python3.12/concurrent/futures/_base.py:401: in __get_result
raise self._exception
../../../_tool/Python/3.12.10/x64/lib/python3.12/concurrent/futures/thread.py:59: in run
result = self.fn(*self.args, **self.kwargs)
.venv/lib/python3.12/site-packages/datasets/download/download_manager.py:229: in _download_single
out = cached_path(url_or_filename, download_config=download_config)
.venv/lib/python3.12/site-packages/datasets/utils/file_utils.py:206: in cached_path
output_path = get_from_cache(
.venv/lib/python3.12/site-packages/datasets/utils/file_utils.py:412: in get_from_cache
fsspec_get(url, temp_file, storage_options=storage_options, desc=download_desc, disable_tqdm=disable_tqdm)
.venv/lib/python3.12/site-packages/datasets/utils/file_utils.py:331: in fsspec_get
fs.get_file(path, temp_file.name, callback=callback)
.venv/lib/python3.12/site-packages/fsspec/asyn.py:118: in wrapper
return sync(self.loop, func, *args, **kwargs)
.venv/lib/python3.12/site-packages/fsspec/asyn.py:103: in sync
raise return_result
.venv/lib/python3.12/site-packages/fsspec/asyn.py:56: in _runner
result[0] = await coro
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <s3fs.core.S3FileSystem object at 0x7f27c18b2e70>
rpath = '<my-bucket>/<my-prefix>/img_1.jpg'
lpath = '/home/runner/_work/_temp/hf_cache/downloads/6c97983efa4e24e534557724655df8247a0bd04326cdfc4a95b638c11e78222d.incomplete'
callback = <datasets.utils.file_utils.TqdmCallback object at 0x7f27c00cdbe0>
version_id = None, kwargs = {}
_open_file = <function S3FileSystem._get_file.<locals>._open_file at 0x7f27628d1120>
body = <StreamingBody at 0x7f276344fa80 for ClientResponse at 0x7f27c015fce0>
content_length = 521923, failed_reads = 0, bytes_read = 0
async def _get_file(
self, rpath, lpath, callback=_DEFAULT_CALLBACK, version_id=None, **kwargs
):
if os.path.isdir(lpath):
return
bucket, key, vers = self.split_path(rpath)
async def _open_file(range: int):
kw = self.req_kw.copy()
if range:
kw["Range"] = f"bytes={range}-"
resp = await self._call_s3(
"get_object",
Bucket=bucket,
Key=key,
**version_id_kw(version_id or vers),
**kw,
)
return resp["Body"], resp.get("ContentLength", None)
body, content_length = await _open_file(range=0)
callback.set_size(content_length)
failed_reads = 0
bytes_read = 0
try:
> with open(lpath, "wb") as f0:
E PermissionError: [Errno 13] Permission denied: '/home/runner/_work/_temp/hf_cache/downloads/6c97983efa4e24e534557724655df8247a0bd04326cdfc4a95b638c11e78222d.incomplete'
.venv/lib/python3.12/site-packages/s3fs/core.py:1355: PermissionError
```
### Steps to reproduce the bug
I believe this is a race condition and cannot reliably re-produce it, but it happens fairly frequently in our GitHub Actions tests and can also be re-produced (with lesser frequency) on cloud VMs.
### Expected behavior
The dataset loads properly with no permission denied error.
### Environment info
- `datasets` version: 3.5.0
- Platform: Linux-5.10.0-34-cloud-amd64-x86_64-with-glibc2.31
- Python version: 3.12.10
- `huggingface_hub` version: 0.30.2
- PyArrow version: 19.0.1
- Pandas version: 2.2.3
- `fsspec` version: 2024.12.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7536/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7536/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7535 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7535/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7535/comments | https://api.github.com/repos/huggingface/datasets/issues/7535/events | https://github.com/huggingface/datasets/pull/7535 | 3,018,289,872 | PR_kwDODunzps6T0lm3 | 7,535 | Change dill version in requirements | {
"avatar_url": "https://avatars.githubusercontent.com/u/98061329?v=4",
"events_url": "https://api.github.com/users/JGrel/events{/privacy}",
"followers_url": "https://api.github.com/users/JGrel/followers",
"following_url": "https://api.github.com/users/JGrel/following{/other_user}",
"gists_url": "https://api.github.com/users/JGrel/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/JGrel",
"id": 98061329,
"login": "JGrel",
"node_id": "U_kgDOBdhMEQ",
"organizations_url": "https://api.github.com/users/JGrel/orgs",
"received_events_url": "https://api.github.com/users/JGrel/received_events",
"repos_url": "https://api.github.com/users/JGrel/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/JGrel/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/JGrel/subscriptions",
"type": "User",
"url": "https://api.github.com/users/JGrel",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7535). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-04-24T19:44:28Z | 2025-05-19T14:51:29Z | null | NONE | null | null | null | Change dill version to >=0.3.9,<0.4.5 and check for errors | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7535/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7535/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7535.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7535",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7535.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7535"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7534 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7534/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7534/comments | https://api.github.com/repos/huggingface/datasets/issues/7534/events | https://github.com/huggingface/datasets/issues/7534 | 3,017,259,407 | I_kwDODunzps6z17mP | 7,534 | TensorFlow RaggedTensor Support (batch-level) | {
"avatar_url": "https://avatars.githubusercontent.com/u/7490199?v=4",
"events_url": "https://api.github.com/users/Lundez/events{/privacy}",
"followers_url": "https://api.github.com/users/Lundez/followers",
"following_url": "https://api.github.com/users/Lundez/following{/other_user}",
"gists_url": "https://api.github.com/users/Lundez/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Lundez",
"id": 7490199,
"login": "Lundez",
"node_id": "MDQ6VXNlcjc0OTAxOTk=",
"organizations_url": "https://api.github.com/users/Lundez/orgs",
"received_events_url": "https://api.github.com/users/Lundez/received_events",
"repos_url": "https://api.github.com/users/Lundez/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Lundez/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Lundez/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Lundez",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | [] | null | [
"Keras doesn't support other inputs other than tf.data.Dataset objects ? it's a bit painful to have to support and maintain this kind of integration\n\nIs there a way to use a `datasets.Dataset` with outputs formatted as tensors / ragged tensors instead ? like in https://huggingface.co/docs/datasets/use_with_tensorflow#dataset-format",
"I'll give it a try when I get the time. But quite sure I already tested the `with_format` approach.\n\nKeras when using TF as backend converts the datasets into `tf.data.Dataset`, much like you do.",
"Hi @Lundez! Thanks for raising this β very valid point, especially for Object Detection use-cases.\n\nYou're right that np_get_batch currently enforces numpy batching, which breaks RaggedTensor support due to its inability to handle nested structures. This likely needs a redesign to allow TensorFlow-native batching in specific formats.\n\nBefore diving into a code change though, could you confirm:\n\nDoes `.with_format(\"tensorflow\")` (without batching) return a `tf.data.Dataset` that works if batching is deferred to `model.fit()`?\n\nHave you tried something like:\n\n```python\ntf_dataset = dataset.with_format(\"tensorflow\").to_tf_dataset(\n columns=[\"image\", \"labels\"],\n label_cols=None,\n batch_size=None # No batching here\n)\nmodel.fit(tf_dataset.batch(BATCH_SIZE)) # Use RaggedTensor batching here\n```\n\nIf this works, it might be worth updating the documentation rather than changing batching logic inside datasets itself.\n\nThat said, happy to explore changes if batching needs to be supported natively for RaggedTensor. Just flagging that itβd require some careful design due to existing numpy assumptions.",
"Hi, we've had to move on for now. \n\nWe have actually also moved to dense tensors to make it possible to xla complie the training. \n\nBut I'll check when I'm back from vacation which is far into the future. \n\nThanks"
] | 2025-04-24T13:14:52Z | 2025-06-30T17:03:39Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
Hi,
Currently datasets does not support RaggedTensor output on batch-level.
When building a Object Detection Dataset (with TensorFlow) I need to enable RaggedTensors as that's how BBoxes & classes are expected from the Keras Model POV.
Currently there's a error thrown saying that "Nested Data is not supported".
It'd be very helpful if this was fixed! :)
### Motivation
Enabling Object Detection pipelines for TensorFlow.
### Your contribution
With guidance I'd happily help making the PR.
The current implementation with DataCollator and later enforcing `np.array` is the problematic part (at the end of `np_get_batch` in `tf_utils.py`). As `numpy` don't support "Raggednes" | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7534/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7534/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7533 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7533/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7533/comments | https://api.github.com/repos/huggingface/datasets/issues/7533/events | https://github.com/huggingface/datasets/pull/7533 | 3,015,075,086 | PR_kwDODunzps6TpraJ | 7,533 | Add custom fingerprint support to `from_generator` | {
"avatar_url": "https://avatars.githubusercontent.com/u/43753582?v=4",
"events_url": "https://api.github.com/users/simonreise/events{/privacy}",
"followers_url": "https://api.github.com/users/simonreise/followers",
"following_url": "https://api.github.com/users/simonreise/following{/other_user}",
"gists_url": "https://api.github.com/users/simonreise/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/simonreise",
"id": 43753582,
"login": "simonreise",
"node_id": "MDQ6VXNlcjQzNzUzNTgy",
"organizations_url": "https://api.github.com/users/simonreise/orgs",
"received_events_url": "https://api.github.com/users/simonreise/received_events",
"repos_url": "https://api.github.com/users/simonreise/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/simonreise/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/simonreise/subscriptions",
"type": "User",
"url": "https://api.github.com/users/simonreise",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"This is great !\r\n\r\nWhat do you think of passing `config_id=` directly to the builder instead of just the suffix ? This would be a power user argument though, or for internal use. And in from_generator the new argument can be `fingerprint=` as in `Dataset.__init__()`\r\n\r\nThe `config_id` can be defined using something like `config_id = \"default-fingerprint=\" + fingerprint`\r\n\r\nI feel ike this could make the Dataset API more coherent if we avoid introducing a new argument while we can juste use `fingerprint=`",
"I looked into this issue and the original cause makes total sense β hashing a large generator is clearly inefficient and fragile for big datasets.\r\n\r\nPR #7533 looks like a robust and flexible solution! It cleanly separates the fingerprinting responsibility by letting users pass `fingerprint=` (now `config_id=`), which avoids hashing heavy objects like generators but still preserves caching logic.\r\n"
] | 2025-04-23T19:31:35Z | 2025-06-29T06:17:17Z | null | NONE | null | null | null | This PR adds `dataset_id_suffix` parameter to 'Dataset.from_generator' function.
`Dataset.from_generator` function passes all of its arguments to `BuilderConfig.create_config_id`, including generator function itself. `BuilderConfig.create_config_id` function tries to hash all the args, which can take a large amount of time or even cause MemoryError if the dataset processed in a generator function is large enough.
This PR allows user to pass a custom fingerprint (`dataset_id_suffix`) to be used as a suffix in a dataset name instead of the one generated by hashing the args.
This PR is a possible solution of #7513 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7533/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7533/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7533.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7533",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7533.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7533"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7532 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7532/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7532/comments | https://api.github.com/repos/huggingface/datasets/issues/7532/events | https://github.com/huggingface/datasets/pull/7532 | 3,009,546,204 | PR_kwDODunzps6TW8Ss | 7,532 | Document the HF_DATASETS_CACHE environment variable in the datasets cache documentation | {
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7532). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Your clarification in your comment at https://github.com/huggingface/datasets/issues/7480#issuecomment-2833640084 sounds great, would you like to update this PR to include it ?",
"Hi @lhoestq, Iβve updated the documentation to reflect the clarifications discussed in #7480. Let me know if anything else is needed!\r\n"
] | 2025-04-22T00:23:13Z | 2025-05-06T15:54:38Z | 2025-05-06T15:54:38Z | CONTRIBUTOR | null | null | null |
This pull request updates the Datasets documentation to include the `HF_DATASETS_CACHE` environment variable. While the current documentation only mentions `HF_HOME` for overriding the default cache directory, `HF_DATASETS_CACHE` is also a supported and useful option for specifying a custom cache location for datasets stored in Arrow format.
This addition is based on the discussion in (https://github.com/huggingface/datasets/issues/7457), where users noted the absence of this variable in the documentation despite its functionality. The update adds a new section to `cache.mdx` that explains how to use `HF_DATASETS_CACHE` with an example.
This change aims to improve clarity and help users better manage their cache directories when working in shared environments or with limited local storage.
Closes #7457. | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7532/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7532/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7532.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7532",
"merged_at": "2025-05-06T15:54:38Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7532.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7532"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7531 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7531/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7531/comments | https://api.github.com/repos/huggingface/datasets/issues/7531/events | https://github.com/huggingface/datasets/issues/7531 | 3,008,914,887 | I_kwDODunzps6zWGXH | 7,531 | Deepspeed reward training hangs at end of training with Dataset.from_list | {
"avatar_url": "https://avatars.githubusercontent.com/u/60710414?v=4",
"events_url": "https://api.github.com/users/Matt00n/events{/privacy}",
"followers_url": "https://api.github.com/users/Matt00n/followers",
"following_url": "https://api.github.com/users/Matt00n/following{/other_user}",
"gists_url": "https://api.github.com/users/Matt00n/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Matt00n",
"id": 60710414,
"login": "Matt00n",
"node_id": "MDQ6VXNlcjYwNzEwNDE0",
"organizations_url": "https://api.github.com/users/Matt00n/orgs",
"received_events_url": "https://api.github.com/users/Matt00n/received_events",
"repos_url": "https://api.github.com/users/Matt00n/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Matt00n/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Matt00n/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Matt00n",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! How big is the dataset ? if you load it using `from_list`, the dataset lives in memory and has to be copied to every gpu process, which can be slow.\n\nIt's fasted if you load it from JSON files from disk, because in that case the dataset in converted to Arrow and loaded from disk using memory mapping. Memory mapping allows to quickly reload the dataset in other processes.\n\nMaybe we can change `from_list` and other methods to always use the disk though, instead of loading in memory, WDYT ?",
"Thanks for raising this! As lhoestq mentioned, the root cause seems to be that `Dataset.from_list()` creates an in-memory dataset, which causes issues with DeepSpeed across multiple GPUs due to the cost of copying that memory to all processes.\n\nUsing `load_dataset(\"json\", ...)` works because Hugging Face datasets then convert the data to Apache Arrow and use **memory mapping**, which avoids this copying overhead.\n\nPossible improvement could be to add an option like `use_disk=True` to `Dataset.from_list()` to allow users to write to Arrow + memory-map the dataset, enabling compatibility with multi-process settings like DeepSpeed, while keeping the current fast behavior by default.\n\nWould love to hear if this direction sounds acceptable before attempting a PR.\n"
] | 2025-04-21T17:29:20Z | 2025-06-29T06:20:45Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | There seems to be a weird interaction between Deepspeed, the Dataset.from_list method and trl's RewardTrainer. On a multi-GPU setup (10 A100s), training always hangs at the very end of training until it times out. The training itself works fine until the end of training and running the same script with Deepspeed on a single GPU works without hangig. The issue persisted across a wide range of Deepspeed configs and training arguments. The issue went away when storing the exact same dataset as a JSON and using `dataset = load_dataset("json", ...)`. Here is my training script:
```python
import pickle
import os
import random
import warnings
import torch
from datasets import load_dataset, Dataset
from transformers import AutoModelForSequenceClassification, AutoTokenizer
from trl import RewardConfig, RewardTrainer, ModelConfig
####################################### Reward model #################################################
# Explicitly set arguments
model_name_or_path = "Qwen/Qwen2.5-1.5B"
output_dir = "Qwen2-0.5B-Reward-LoRA"
per_device_train_batch_size = 2
num_train_epochs = 5
gradient_checkpointing = True
learning_rate = 1.0e-4
logging_steps = 25
eval_strategy = "steps"
eval_steps = 50
max_length = 2048
torch_dtype = "auto"
trust_remote_code = False
model_args = ModelConfig(
model_name_or_path=model_name_or_path,
model_revision=None,
trust_remote_code=trust_remote_code,
torch_dtype=torch_dtype,
lora_task_type="SEQ_CLS", # Make sure task type is seq_cls
)
training_args = RewardConfig(
output_dir=output_dir,
per_device_train_batch_size=per_device_train_batch_size,
num_train_epochs=num_train_epochs,
gradient_checkpointing=gradient_checkpointing,
learning_rate=learning_rate,
logging_steps=logging_steps,
eval_strategy=eval_strategy,
eval_steps=eval_steps,
max_length=max_length,
gradient_checkpointing_kwargs=dict(use_reentrant=False),
center_rewards_coefficient = 0.01,
fp16=False,
bf16=True,
save_strategy="no",
dataloader_num_workers=0,
# deepspeed="./configs/deepspeed_config.json",
)
################
# Model & Tokenizer
################
model_kwargs = dict(
revision=model_args.model_revision,
use_cache=False if training_args.gradient_checkpointing else True,
torch_dtype=model_args.torch_dtype,
)
tokenizer = AutoTokenizer.from_pretrained(
model_args.model_name_or_path, use_fast=True
)
model = AutoModelForSequenceClassification.from_pretrained(
model_args.model_name_or_path, num_labels=1, trust_remote_code=model_args.trust_remote_code, **model_kwargs
)
# Align padding tokens between tokenizer and model
model.config.pad_token_id = tokenizer.pad_token_id
# If post-training a base model, use ChatML as the default template
if tokenizer.chat_template is None:
model, tokenizer = setup_chat_format(model, tokenizer)
if model_args.use_peft and model_args.lora_task_type != "SEQ_CLS":
warnings.warn(
"You are using a `task_type` that is different than `SEQ_CLS` for PEFT. This will lead to silent bugs"
" Make sure to pass --lora_task_type SEQ_CLS when using this script with PEFT.",
UserWarning,
)
##############
# Load dataset
##############
with open('./prefs.pkl', 'rb') as fh:
loaded_data = pickle.load(fh)
random.shuffle(loaded_data)
dataset = []
for a_wins, a, b in loaded_data:
if a_wins == 0:
a, b = b, a
dataset.append({'chosen': a, 'rejected': b})
dataset = Dataset.from_list(dataset)
# Split the dataset into training and evaluation sets
train_eval_split = dataset.train_test_split(test_size=0.15, shuffle=True, seed=42)
# Access the training and evaluation datasets
train_dataset = train_eval_split['train']
eval_dataset = train_eval_split['test']
##########
# Training
##########
trainer = RewardTrainer(
model=model,
processing_class=tokenizer,
args=training_args,
train_dataset=train_dataset,
eval_dataset=eval_dataset,
)
trainer.train()
```
Replacing `dataset = Dataset.from_list(dataset)` with
```python
with open('./prefs.json', 'w') as fh:
json.dump(dataset, fh)
dataset = load_dataset("json", data_files="./prefs.json", split='train')
```
resolves the issue. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7531/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7531/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7530 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7530/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7530/comments | https://api.github.com/repos/huggingface/datasets/issues/7530/events | https://github.com/huggingface/datasets/issues/7530 | 3,007,452,499 | I_kwDODunzps6zQhVT | 7,530 | How to solve "Spaces stuck in Building" problems | {
"avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4",
"events_url": "https://api.github.com/users/ghost/events{/privacy}",
"followers_url": "https://api.github.com/users/ghost/followers",
"following_url": "https://api.github.com/users/ghost/following{/other_user}",
"gists_url": "https://api.github.com/users/ghost/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ghost",
"id": 10137,
"login": "ghost",
"node_id": "MDQ6VXNlcjEwMTM3",
"organizations_url": "https://api.github.com/users/ghost/orgs",
"received_events_url": "https://api.github.com/users/ghost/received_events",
"repos_url": "https://api.github.com/users/ghost/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ghost/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ghost/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ghost",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"I'm facing the same issueβSpace stuck in \"Building\" even after restart and Factory rebuild. Any fix?\n",
"> I'm facing the same issueβSpace stuck in \"Building\" even after restart and Factory rebuild. Any fix?\n\nAlso see https://github.com/huggingface/huggingface_hub/issues/3019",
"I'm facing the same issue. The build fails with the same error, and restarting won't help. Is there a fix or ETA? "
] | 2025-04-21T03:08:38Z | 2025-04-22T07:49:52Z | 2025-04-22T07:49:52Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Public spaces may stuck in Building after restarting, error log as follows:
build error
Unexpected job error
ERROR: failed to push spaces-registry.huggingface.tech/spaces/*:cpu-*-*: unexpected status from HEAD request to https://spaces-registry.huggingface.tech/v2/spaces/*/manifests/cpu-*-*: 401 Unauthorized
### Steps to reproduce the bug
Restart space / Factory rebuild cannot avoid it
### Expected behavior
Fix this problem
### Environment info
no requirements.txt can still happen
python gradio spaces | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7530/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7530/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7529 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7529/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7529/comments | https://api.github.com/repos/huggingface/datasets/issues/7529/events | https://github.com/huggingface/datasets/issues/7529 | 3,007,118,969 | I_kwDODunzps6zPP55 | 7,529 | audio folder builder cannot detect custom split name | {
"avatar_url": "https://avatars.githubusercontent.com/u/37548991?v=4",
"events_url": "https://api.github.com/users/phineas-pta/events{/privacy}",
"followers_url": "https://api.github.com/users/phineas-pta/followers",
"following_url": "https://api.github.com/users/phineas-pta/following{/other_user}",
"gists_url": "https://api.github.com/users/phineas-pta/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/phineas-pta",
"id": 37548991,
"login": "phineas-pta",
"node_id": "MDQ6VXNlcjM3NTQ4OTkx",
"organizations_url": "https://api.github.com/users/phineas-pta/orgs",
"received_events_url": "https://api.github.com/users/phineas-pta/received_events",
"repos_url": "https://api.github.com/users/phineas-pta/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/phineas-pta/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/phineas-pta/subscriptions",
"type": "User",
"url": "https://api.github.com/users/phineas-pta",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-04-20T16:53:21Z | 2025-04-20T16:53:21Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
when using audio folder builder (`load_dataset("audiofolder", data_dir="/path/to/folder")`), it cannot detect custom split name other than train/validation/test
### Steps to reproduce the bug
i have the following folder structure
```
my_dataset/
βββ train/
β βββ lorem.wav
β βββ β¦
β βββ metadata.csv
βββ test/
β βββ ipsum.wav
β βββ β¦
β βββ metadata.csv
βββ validation/
β βββ dolor.wav
β βββ β¦
β βββ metadata.csv
βββ custom/
βββ sit.wav
βββ β¦
βββ metadata.csv
```
using `ds = load_dataset("audiofolder", data_dir="/path/to/my_dataset")`
### Expected behavior
i got `ds` with only 3 splits train/validation/test, whenever i rename train/validation/test folder it also disappear if i re-create `ds`
### Environment info
- `datasets` version: 3.5.0
- Platform: Windows-11-10.0.26100-SP0
- Python version: 3.12.8
- `huggingface_hub` version: 0.30.2
- PyArrow version: 18.1.0
- Pandas version: 2.2.3
- `fsspec` version: 2024.9.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7529/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7529/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7528 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7528/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7528/comments | https://api.github.com/repos/huggingface/datasets/issues/7528/events | https://github.com/huggingface/datasets/issues/7528 | 3,006,433,485 | I_kwDODunzps6zMojN | 7,528 | Data Studio Error: Convert JSONL incorrectly | {
"avatar_url": "https://avatars.githubusercontent.com/u/144962041?v=4",
"events_url": "https://api.github.com/users/zxccade/events{/privacy}",
"followers_url": "https://api.github.com/users/zxccade/followers",
"following_url": "https://api.github.com/users/zxccade/following{/other_user}",
"gists_url": "https://api.github.com/users/zxccade/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zxccade",
"id": 144962041,
"login": "zxccade",
"node_id": "U_kgDOCKPx-Q",
"organizations_url": "https://api.github.com/users/zxccade/orgs",
"received_events_url": "https://api.github.com/users/zxccade/received_events",
"repos_url": "https://api.github.com/users/zxccade/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zxccade/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zxccade/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zxccade",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! Your JSONL file is incompatible with Arrow / Parquet. Indeed in Arrow / Parquet every dict should have the same keys, while in your dataset the bboxes have varying keys.\n\nThis causes the Data Studio to treat the bboxes as if each row was missing the keys from other rows.\n\nFeel free to take a look at the docs on object segmentation to see how to format a dataset with bboxes: https://huggingface.co/docs/datasets/object_detection"
] | 2025-04-19T13:21:44Z | 2025-05-06T13:18:38Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Hi there,
I uploaded a dataset here https://huggingface.co/datasets/V-STaR-Bench/V-STaR, but I found that Data Studio incorrectly convert the "bboxes" value for the whole dataset. Therefore, anyone who downloaded the dataset via the API would get the wrong "bboxes" value in the data file.
Could you help me address the issue?
Many thanks,
### Steps to reproduce the bug
The JSONL file of [V_STaR_test_release.jsonl](https://huggingface.co/datasets/V-STaR-Bench/V-STaR/blob/main/V_STaR_test_release.jsonl) has the correct values of every "bboxes" for each sample.
But in the Data Studio, we can see that the values of "bboxes" have changed, and load the dataset via API will also get the wrong values.
### Expected behavior
Fix the bug to correctly download my dataset.
### Environment info
- `datasets` version: 2.16.1
- Platform: Linux-5.14.0-427.22.1.el9_4.x86_64-x86_64-with-glibc2.34
- Python version: 3.10.16
- `huggingface_hub` version: 0.29.3
- PyArrow version: 19.0.0
- Pandas version: 2.2.3
- `fsspec` version: 2023.10.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7528/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7528/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7527 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7527/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7527/comments | https://api.github.com/repos/huggingface/datasets/issues/7527/events | https://github.com/huggingface/datasets/issues/7527 | 3,005,242,422 | I_kwDODunzps6zIFw2 | 7,527 | Auto-merge option for `convert-to-parquet` | {
"avatar_url": "https://avatars.githubusercontent.com/u/17013474?v=4",
"events_url": "https://api.github.com/users/klamike/events{/privacy}",
"followers_url": "https://api.github.com/users/klamike/followers",
"following_url": "https://api.github.com/users/klamike/following{/other_user}",
"gists_url": "https://api.github.com/users/klamike/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/klamike",
"id": 17013474,
"login": "klamike",
"node_id": "MDQ6VXNlcjE3MDEzNDc0",
"organizations_url": "https://api.github.com/users/klamike/orgs",
"received_events_url": "https://api.github.com/users/klamike/received_events",
"repos_url": "https://api.github.com/users/klamike/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/klamike/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/klamike/subscriptions",
"type": "User",
"url": "https://api.github.com/users/klamike",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/17013474?v=4",
"events_url": "https://api.github.com/users/klamike/events{/privacy}",
"followers_url": "https://api.github.com/users/klamike/followers",
"following_url": "https://api.github.com/users/klamike/following{/other_user}",
"gists_url": "https://api.github.com/users/klamike/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/klamike",
"id": 17013474,
"login": "klamike",
"node_id": "MDQ6VXNlcjE3MDEzNDc0",
"organizations_url": "https://api.github.com/users/klamike/orgs",
"received_events_url": "https://api.github.com/users/klamike/received_events",
"repos_url": "https://api.github.com/users/klamike/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/klamike/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/klamike/subscriptions",
"type": "User",
"url": "https://api.github.com/users/klamike",
"user_view_type": "public"
} | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/17013474?v=4",
"events_url": "https://api.github.com/users/klamike/events{/privacy}",
"followers_url": "https://api.github.com/users/klamike/followers",
"following_url": "https://api.github.com/users/klamike/following{/other_user}",
"gists_url": "https://api.github.com/users/klamike/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/klamike",
"id": 17013474,
"login": "klamike",
"node_id": "MDQ6VXNlcjE3MDEzNDc0",
"organizations_url": "https://api.github.com/users/klamike/orgs",
"received_events_url": "https://api.github.com/users/klamike/received_events",
"repos_url": "https://api.github.com/users/klamike/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/klamike/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/klamike/subscriptions",
"type": "User",
"url": "https://api.github.com/users/klamike",
"user_view_type": "public"
}
] | null | [
"Alternatively, there could be an option to switch from submitting PRs to just committing changes directly to `main`.",
"Why not, I'd be in favor of `--merge-pull-request` to call `HfApi().merge_pull_request()` at the end of the conversion :) feel free to open a PR if you'd like",
"#self-assign"
] | 2025-04-18T16:03:22Z | 2025-05-07T12:47:02Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
Add a command-line option, e.g. `--auto-merge-pull-request` that enables automatic merging of the commits created by the `convert-to-parquet` tool.
### Motivation
Large datasets may result in dozens of PRs due to the splitting mechanism. Each of these has to be manually accepted via the website.
### Your contribution
Happy to look into submitting a PR if this is of interest to maintainers. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7527/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7527/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7526 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7526/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7526/comments | https://api.github.com/repos/huggingface/datasets/issues/7526/events | https://github.com/huggingface/datasets/issues/7526 | 3,005,107,536 | I_kwDODunzps6zHk1Q | 7,526 | Faster downloads/uploads with Xet storage | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-04-18T14:46:42Z | 2025-05-12T12:09:09Z | null | MEMBER | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | 
## Xet is out !
Over the past few weeks, Hugging Faceβs [Xet Team](https://huggingface.co/xet-team) took a major step forward by [migrating the first Model and Dataset repositories off LFS and to Xet storage](https://huggingface.co/posts/jsulz/911431940353906).
See more information on the HF blog: https://huggingface.co/blog/xet-on-the-hub
You can already enable Xet on Hugging Face account to benefit from faster downloads and uploads :)
We finalized an official integration with the `huggingface_hub` library that means you get the benefits of Xet without any significant changes to your current workflow.
## Previous versions of `datasets`
For older versions of `datasets` you might see this warning in `push_to_hub()`:
```
Uploading files as bytes or binary IO objects is not supported by Xet Storage.
```
This means the `huggingface_hub` + Xet integration isn't enabled for your version of `datasets`.
You can fix this by updating to `datasets>=3.6.0` and `huggingface_hub>=0.31.0`
```
pip install -U datasets huggingface_hub
```
## The future
Stay tuned for more Xet optimizations, especially on [Xet-optimized Parquet](https://huggingface.co/blog/improve_parquet_dedupe)
| null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 5,
"total_count": 5,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7526/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7526/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7525 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7525/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7525/comments | https://api.github.com/repos/huggingface/datasets/issues/7525/events | https://github.com/huggingface/datasets/pull/7525 | 3,003,032,248 | PR_kwDODunzps6TBOH1 | 7,525 | Fix indexing in split commit messages | {
"avatar_url": "https://avatars.githubusercontent.com/u/17013474?v=4",
"events_url": "https://api.github.com/users/klamike/events{/privacy}",
"followers_url": "https://api.github.com/users/klamike/followers",
"following_url": "https://api.github.com/users/klamike/following{/other_user}",
"gists_url": "https://api.github.com/users/klamike/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/klamike",
"id": 17013474,
"login": "klamike",
"node_id": "MDQ6VXNlcjE3MDEzNDc0",
"organizations_url": "https://api.github.com/users/klamike/orgs",
"received_events_url": "https://api.github.com/users/klamike/received_events",
"repos_url": "https://api.github.com/users/klamike/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/klamike/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/klamike/subscriptions",
"type": "User",
"url": "https://api.github.com/users/klamike",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! this is expected and is coherent with other naming conventions in `datasets` such as parquet shards naming"
] | 2025-04-17T17:06:26Z | 2025-04-28T14:26:27Z | 2025-04-28T14:26:27Z | NONE | null | null | null | When a large commit is split up, it seems the commit index in the message is zero-based while the total number is one-based. I came across this running `convert-to-parquet` and was wondering why there was no `6-of-6` commit. This PR fixes that by adding one to the commit index, so both are one-based.
Current behavior:
<img width="463" alt="Screenshot 2025-04-17 at 1 00 17β―PM" src="https://github.com/user-attachments/assets/7f3d389e-cb92-405d-a3c2-f2b1cdf0cb79" /> | {
"avatar_url": "https://avatars.githubusercontent.com/u/17013474?v=4",
"events_url": "https://api.github.com/users/klamike/events{/privacy}",
"followers_url": "https://api.github.com/users/klamike/followers",
"following_url": "https://api.github.com/users/klamike/following{/other_user}",
"gists_url": "https://api.github.com/users/klamike/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/klamike",
"id": 17013474,
"login": "klamike",
"node_id": "MDQ6VXNlcjE3MDEzNDc0",
"organizations_url": "https://api.github.com/users/klamike/orgs",
"received_events_url": "https://api.github.com/users/klamike/received_events",
"repos_url": "https://api.github.com/users/klamike/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/klamike/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/klamike/subscriptions",
"type": "User",
"url": "https://api.github.com/users/klamike",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7525/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7525/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7525.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7525",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7525.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7525"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7524 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7524/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7524/comments | https://api.github.com/repos/huggingface/datasets/issues/7524/events | https://github.com/huggingface/datasets/pull/7524 | 3,002,067,826 | PR_kwDODunzps6S99KB | 7,524 | correct use with polars example | {
"avatar_url": "https://avatars.githubusercontent.com/u/43832476?v=4",
"events_url": "https://api.github.com/users/SiQube/events{/privacy}",
"followers_url": "https://api.github.com/users/SiQube/followers",
"following_url": "https://api.github.com/users/SiQube/following{/other_user}",
"gists_url": "https://api.github.com/users/SiQube/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/SiQube",
"id": 43832476,
"login": "SiQube",
"node_id": "MDQ6VXNlcjQzODMyNDc2",
"organizations_url": "https://api.github.com/users/SiQube/orgs",
"received_events_url": "https://api.github.com/users/SiQube/received_events",
"repos_url": "https://api.github.com/users/SiQube/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/SiQube/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/SiQube/subscriptions",
"type": "User",
"url": "https://api.github.com/users/SiQube",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [] | 2025-04-17T10:19:19Z | 2025-04-28T13:48:34Z | 2025-04-28T13:48:33Z | CONTRIBUTOR | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7524/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7524/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7524.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7524",
"merged_at": "2025-04-28T13:48:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7524.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7524"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7523 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7523/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7523/comments | https://api.github.com/repos/huggingface/datasets/issues/7523/events | https://github.com/huggingface/datasets/pull/7523 | 2,999,616,692 | PR_kwDODunzps6S1r8w | 7,523 | mention av in video docs | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7523). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-04-16T13:11:12Z | 2025-04-16T13:13:45Z | 2025-04-16T13:11:42Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7523/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7523/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7523.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7523",
"merged_at": "2025-04-16T13:11:42Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7523.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7523"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7522 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7522/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7522/comments | https://api.github.com/repos/huggingface/datasets/issues/7522/events | https://github.com/huggingface/datasets/pull/7522 | 2,998,169,017 | PR_kwDODunzps6SwwXW | 7,522 | Preserve formatting in concatenated IterableDataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/5140987?v=4",
"events_url": "https://api.github.com/users/francescorubbo/events{/privacy}",
"followers_url": "https://api.github.com/users/francescorubbo/followers",
"following_url": "https://api.github.com/users/francescorubbo/following{/other_user}",
"gists_url": "https://api.github.com/users/francescorubbo/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/francescorubbo",
"id": 5140987,
"login": "francescorubbo",
"node_id": "MDQ6VXNlcjUxNDA5ODc=",
"organizations_url": "https://api.github.com/users/francescorubbo/orgs",
"received_events_url": "https://api.github.com/users/francescorubbo/received_events",
"repos_url": "https://api.github.com/users/francescorubbo/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/francescorubbo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/francescorubbo/subscriptions",
"type": "User",
"url": "https://api.github.com/users/francescorubbo",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7522). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-04-16T02:37:33Z | 2025-05-19T15:07:38Z | 2025-05-19T15:07:37Z | CONTRIBUTOR | null | null | null | Fixes #7515 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7522/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7522/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7522.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7522",
"merged_at": "2025-05-19T15:07:37Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7522.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7522"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7521 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7521/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7521/comments | https://api.github.com/repos/huggingface/datasets/issues/7521/events | https://github.com/huggingface/datasets/pull/7521 | 2,997,666,366 | PR_kwDODunzps6SvEZp | 7,521 | fix: Image Feature in Datasets Library Fails to Handle bytearray Objects from Spark DataFrames (#7517) | {
"avatar_url": "https://avatars.githubusercontent.com/u/73196164?v=4",
"events_url": "https://api.github.com/users/giraffacarp/events{/privacy}",
"followers_url": "https://api.github.com/users/giraffacarp/followers",
"following_url": "https://api.github.com/users/giraffacarp/following{/other_user}",
"gists_url": "https://api.github.com/users/giraffacarp/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/giraffacarp",
"id": 73196164,
"login": "giraffacarp",
"node_id": "MDQ6VXNlcjczMTk2MTY0",
"organizations_url": "https://api.github.com/users/giraffacarp/orgs",
"received_events_url": "https://api.github.com/users/giraffacarp/received_events",
"repos_url": "https://api.github.com/users/giraffacarp/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/giraffacarp/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/giraffacarp/subscriptions",
"type": "User",
"url": "https://api.github.com/users/giraffacarp",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"@lhoestq let me know if you prefer to change the spark iterator so it outputs `bytes`",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7521). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-04-15T21:23:58Z | 2025-05-07T14:17:29Z | 2025-05-07T14:17:29Z | CONTRIBUTOR | null | null | null | ## Task
Support bytes-like objects (bytes and bytearray) in Features classes
### Description
The `Features` classes only accept `bytes` objects for binary data, but not `bytearray`. This leads to errors when using `IterableDataset.from_spark()` with Spark DataFrames as they contain `bytearray` objects, even though both `bytes` and `bytearray` are valid [*bytes-like objects* in Python](https://docs.python.org/3/glossary.html#term-bytes-like-object).
### Changes
- Updated `Features` classes to accept both `bytes` and `bytearray` types for binary data fields.
### Reasoning
- `bytes` and `bytearray` serve the same purpose for binary data, with the only difference being mutability.
- Modifying the Spark iterator to convert `bytearray` to `bytes` would be a workaround, not a true fix. I think the correct solution is to accept all bytes-like objects as input.
- This approach is more robust and future-proof since Python 3.12+ provides a [standard way to check for buffer protocol](https://docs.python.org/3/c-api/buffer.html#bufferobjects).
### Testing
- Added tests to cover `bytearray` inputs for image features.
### Related Issues
- Fixes: #7517 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7521/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7521/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7521.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7521",
"merged_at": "2025-05-07T14:17:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7521.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7521"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7520 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7520/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7520/comments | https://api.github.com/repos/huggingface/datasets/issues/7520/events | https://github.com/huggingface/datasets/issues/7520 | 2,997,422,044 | I_kwDODunzps6yqQfc | 7,520 | Update items in the dataset without `map` | {
"avatar_url": "https://avatars.githubusercontent.com/u/122402293?v=4",
"events_url": "https://api.github.com/users/mashdragon/events{/privacy}",
"followers_url": "https://api.github.com/users/mashdragon/followers",
"following_url": "https://api.github.com/users/mashdragon/following{/other_user}",
"gists_url": "https://api.github.com/users/mashdragon/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mashdragon",
"id": 122402293,
"login": "mashdragon",
"node_id": "U_kgDOB0u19Q",
"organizations_url": "https://api.github.com/users/mashdragon/orgs",
"received_events_url": "https://api.github.com/users/mashdragon/received_events",
"repos_url": "https://api.github.com/users/mashdragon/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mashdragon/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mashdragon/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mashdragon",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | [] | null | [
"Hello!\n\nHave you looked at `Dataset.shard`? [Docs](https://huggingface.co/docs/datasets/en/process#shard)\n\nUsing this method you could break your dataset in N shards. Apply `map` on each shard and concatenate them back."
] | 2025-04-15T19:39:01Z | 2025-04-19T18:47:46Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
I would like to be able to update items in my dataset without affecting all rows. At least if there was a range option, I would be able to process those items, save the dataset, and then continue.
If I am supposed to split the dataset first, that is not clear, since the docs suggest that any of those functions returns a new object, so I don't think I can do that.
### Motivation
I am applying an extremely time-consuming function to each item in my `Dataset`. Unfortunately, datasets only supports updating values via `map`, so if my computer dies in the middle of this long-running process, I lose all progress. This is far from ideal. I would like to use `datasets` throughout this processing, but this limitation is now forcing me to write my own dataset format just to do this intermediary operation.
It would be less intuitive but I suppose I could split and then concatenate the dataset before saving? But this feels very inefficient.
### Your contribution
I can test the feature. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7520/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7520/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7519 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7519/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7519/comments | https://api.github.com/repos/huggingface/datasets/issues/7519/events | https://github.com/huggingface/datasets/pull/7519 | 2,996,458,961 | PR_kwDODunzps6Sq76Z | 7,519 | pdf docs fixes | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7519). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-04-15T13:35:56Z | 2025-04-15T13:38:31Z | 2025-04-15T13:36:03Z | MEMBER | null | null | null | close https://github.com/huggingface/datasets/issues/7494 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7519/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7519/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7519.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7519",
"merged_at": "2025-04-15T13:36:03Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7519.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7519"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7518 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7518/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7518/comments | https://api.github.com/repos/huggingface/datasets/issues/7518/events | https://github.com/huggingface/datasets/issues/7518 | 2,996,141,825 | I_kwDODunzps6ylX8B | 7,518 | num_proc parallelization works only for first ~10s. | {
"avatar_url": "https://avatars.githubusercontent.com/u/33901783?v=4",
"events_url": "https://api.github.com/users/pshishodiaa/events{/privacy}",
"followers_url": "https://api.github.com/users/pshishodiaa/followers",
"following_url": "https://api.github.com/users/pshishodiaa/following{/other_user}",
"gists_url": "https://api.github.com/users/pshishodiaa/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/pshishodiaa",
"id": 33901783,
"login": "pshishodiaa",
"node_id": "MDQ6VXNlcjMzOTAxNzgz",
"organizations_url": "https://api.github.com/users/pshishodiaa/orgs",
"received_events_url": "https://api.github.com/users/pshishodiaa/received_events",
"repos_url": "https://api.github.com/users/pshishodiaa/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/pshishodiaa/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pshishodiaa/subscriptions",
"type": "User",
"url": "https://api.github.com/users/pshishodiaa",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi, can you check if the processes are still alive ? It's a bit weird because `datasets` does check if processes crash and return an error in that case",
"Thank you for reverting quickly. I digged a bit, and realized my disk's IOPS is also limited - which is causing this. will check further and report if it's an issue of hf datasets' side or mine. "
] | 2025-04-15T11:44:03Z | 2025-04-15T13:12:13Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When I try to load an already downloaded dataset with num_proc=64, the speed is very high for the first 10-20 seconds acheiving 30-40K samples / s, and 100% utilization for all cores but it soon drops to <= 1000 with almost 0% utilization for most cores.
### Steps to reproduce the bug
```
// download dataset with cli
!huggingface-cli download --repo-type dataset timm/imagenet-1k-wds --max-workers 32
from datasets import load_dataset
ds = load_dataset("timm/imagenet-1k-wds", num_proc=64)
```
### Expected behavior
100% core utilization throughout.
### Environment info
Azure A100-80GB, 16 cores VM
 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7518/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7518/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7517 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7517/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7517/comments | https://api.github.com/repos/huggingface/datasets/issues/7517/events | https://github.com/huggingface/datasets/issues/7517 | 2,996,106,077 | I_kwDODunzps6ylPNd | 7,517 | Image Feature in Datasets Library Fails to Handle bytearray Objects from Spark DataFrames | {
"avatar_url": "https://avatars.githubusercontent.com/u/73196164?v=4",
"events_url": "https://api.github.com/users/giraffacarp/events{/privacy}",
"followers_url": "https://api.github.com/users/giraffacarp/followers",
"following_url": "https://api.github.com/users/giraffacarp/following{/other_user}",
"gists_url": "https://api.github.com/users/giraffacarp/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/giraffacarp",
"id": 73196164,
"login": "giraffacarp",
"node_id": "MDQ6VXNlcjczMTk2MTY0",
"organizations_url": "https://api.github.com/users/giraffacarp/orgs",
"received_events_url": "https://api.github.com/users/giraffacarp/received_events",
"repos_url": "https://api.github.com/users/giraffacarp/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/giraffacarp/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/giraffacarp/subscriptions",
"type": "User",
"url": "https://api.github.com/users/giraffacarp",
"user_view_type": "public"
} | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/73196164?v=4",
"events_url": "https://api.github.com/users/giraffacarp/events{/privacy}",
"followers_url": "https://api.github.com/users/giraffacarp/followers",
"following_url": "https://api.github.com/users/giraffacarp/following{/other_user}",
"gists_url": "https://api.github.com/users/giraffacarp/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/giraffacarp",
"id": 73196164,
"login": "giraffacarp",
"node_id": "MDQ6VXNlcjczMTk2MTY0",
"organizations_url": "https://api.github.com/users/giraffacarp/orgs",
"received_events_url": "https://api.github.com/users/giraffacarp/received_events",
"repos_url": "https://api.github.com/users/giraffacarp/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/giraffacarp/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/giraffacarp/subscriptions",
"type": "User",
"url": "https://api.github.com/users/giraffacarp",
"user_view_type": "public"
} | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/73196164?v=4",
"events_url": "https://api.github.com/users/giraffacarp/events{/privacy}",
"followers_url": "https://api.github.com/users/giraffacarp/followers",
"following_url": "https://api.github.com/users/giraffacarp/following{/other_user}",
"gists_url": "https://api.github.com/users/giraffacarp/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/giraffacarp",
"id": 73196164,
"login": "giraffacarp",
"node_id": "MDQ6VXNlcjczMTk2MTY0",
"organizations_url": "https://api.github.com/users/giraffacarp/orgs",
"received_events_url": "https://api.github.com/users/giraffacarp/received_events",
"repos_url": "https://api.github.com/users/giraffacarp/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/giraffacarp/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/giraffacarp/subscriptions",
"type": "User",
"url": "https://api.github.com/users/giraffacarp",
"user_view_type": "public"
}
] | null | [
"Hi ! The `Image()` type accepts either\n- a `bytes` object containing the image bytes\n- a `str` object containing the image path\n- a `PIL.Image` object\n\nbut it doesn't support `bytearray`, maybe you can convert to `bytes` beforehand ?",
"Hi @lhoestq, \nconverting to bytes is certainly possible and would work around the error. However, the core issue is that `Dataset` and `IterableDataset` behave differently with the features.\n\nIβd be happy to work on a fix for this issue.",
"I see, that's an issue indeed. Feel free to ping me if I can help with reviews or any guidance\n\nIf it can help, the code that takes a Spark DataFrame and iterates on the rows for `IterableDataset` is here: \n\nhttps://github.com/huggingface/datasets/blob/6a96bf313085d7538a999b929a550e14e1d406c9/src/datasets/packaged_modules/spark/spark.py#L49-L53",
"#self-assign"
] | 2025-04-15T11:29:17Z | 2025-05-07T14:17:30Z | 2025-05-07T14:17:30Z | CONTRIBUTOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When using `IterableDataset.from_spark()` with a Spark DataFrame containing image data, the `Image` feature class fails to properly process this data type, causing an `AttributeError: 'bytearray' object has no attribute 'get'`
### Steps to reproduce the bug
1. Create a Spark DataFrame with a column containing image data as bytearray objects
2. Define a Feature schema with an Image feature
3. Create an IterableDataset using `IterableDataset.from_spark()`
4. Attempt to iterate through the dataset
```
from pyspark.sql import SparkSession
from datasets import Dataset, IterableDataset, Features, Image, Value
# initialize spark
spark = SparkSession.builder.appName("MinimalRepro").getOrCreate()
# create spark dataframe
data = [(0, open("image.png", "rb").read())]
df = spark.createDataFrame(data, "idx: int, image: binary")
# convert to dataset
features = Features({"idx": Value("int64"), "image": Image()})
ds = Dataset.from_spark(df, features=features)
ds_iter = IterableDataset.from_spark(df, features=features)
# iterate
print(next(iter(ds)))
print(next(iter(ds_iter)))
```
### Expected behavior
The features should work on `IterableDataset` the same way they work on `Dataset`
### Environment info
- `datasets` version: 3.5.0
- Platform: macOS-15.3.2-arm64-arm-64bit
- Python version: 3.12.7
- `huggingface_hub` version: 0.30.2
- PyArrow version: 18.1.0
- Pandas version: 2.2.3
- `fsspec` version: 2024.12.0
| {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7517/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7517/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7516 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7516/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7516/comments | https://api.github.com/repos/huggingface/datasets/issues/7516/events | https://github.com/huggingface/datasets/issues/7516 | 2,995,780,283 | I_kwDODunzps6yj_q7 | 7,516 | unsloth/DeepSeek-R1-Distill-Qwen-32B server error | {
"avatar_url": "https://avatars.githubusercontent.com/u/164353862?v=4",
"events_url": "https://api.github.com/users/Editor-1/events{/privacy}",
"followers_url": "https://api.github.com/users/Editor-1/followers",
"following_url": "https://api.github.com/users/Editor-1/following{/other_user}",
"gists_url": "https://api.github.com/users/Editor-1/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Editor-1",
"id": 164353862,
"login": "Editor-1",
"node_id": "U_kgDOCcvXRg",
"organizations_url": "https://api.github.com/users/Editor-1/orgs",
"received_events_url": "https://api.github.com/users/Editor-1/received_events",
"repos_url": "https://api.github.com/users/Editor-1/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Editor-1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Editor-1/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Editor-1",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [] | 2025-04-15T09:26:53Z | 2025-04-15T09:57:26Z | 2025-04-15T09:57:26Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
hfhubhttperror: 500 server error: internal server error for url: https://huggingface.co/api/models/unsloth/deepseek-r1-distill-qwen-32b-bnb-4bit/commits/main (request id: root=1-67fe23fa-3a2150eb444c2a823c388579;de3aed68-c397-4da5-94d4-6565efd3b919) internal error - we're working hard to fix this as soon as possible!
### Steps to reproduce the bug
unsloth/DeepSeek-R1-Distill-Qwen-32B server error
### Expected behavior
Network repair
### Environment info
The web side is also unavailable | {
"avatar_url": "https://avatars.githubusercontent.com/u/164353862?v=4",
"events_url": "https://api.github.com/users/Editor-1/events{/privacy}",
"followers_url": "https://api.github.com/users/Editor-1/followers",
"following_url": "https://api.github.com/users/Editor-1/following{/other_user}",
"gists_url": "https://api.github.com/users/Editor-1/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Editor-1",
"id": 164353862,
"login": "Editor-1",
"node_id": "U_kgDOCcvXRg",
"organizations_url": "https://api.github.com/users/Editor-1/orgs",
"received_events_url": "https://api.github.com/users/Editor-1/received_events",
"repos_url": "https://api.github.com/users/Editor-1/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Editor-1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Editor-1/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Editor-1",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7516/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7516/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7515 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7515/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7515/comments | https://api.github.com/repos/huggingface/datasets/issues/7515/events | https://github.com/huggingface/datasets/issues/7515 | 2,995,082,418 | I_kwDODunzps6yhVSy | 7,515 | `concatenate_datasets` does not preserve Pytorch format for IterableDataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/5140987?v=4",
"events_url": "https://api.github.com/users/francescorubbo/events{/privacy}",
"followers_url": "https://api.github.com/users/francescorubbo/followers",
"following_url": "https://api.github.com/users/francescorubbo/following{/other_user}",
"gists_url": "https://api.github.com/users/francescorubbo/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/francescorubbo",
"id": 5140987,
"login": "francescorubbo",
"node_id": "MDQ6VXNlcjUxNDA5ODc=",
"organizations_url": "https://api.github.com/users/francescorubbo/orgs",
"received_events_url": "https://api.github.com/users/francescorubbo/received_events",
"repos_url": "https://api.github.com/users/francescorubbo/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/francescorubbo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/francescorubbo/subscriptions",
"type": "User",
"url": "https://api.github.com/users/francescorubbo",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! Oh indeed it would be cool to return the same format in that case. Would you like to submit a PR ? The function that does the concatenation is here:\n\nhttps://github.com/huggingface/datasets/blob/90e5bf8a8599b625d6103ee5ac83b98269991141/src/datasets/iterable_dataset.py#L3375-L3380",
"Thank you for the pointer, @lhoestq ! See #7522 "
] | 2025-04-15T04:36:34Z | 2025-05-19T15:07:38Z | 2025-05-19T15:07:38Z | CONTRIBUTOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When concatenating datasets with `concatenate_datasets`, I would expect the resulting combined dataset to be in the same format as the inputs (assuming it's consistent). This is indeed the behavior when combining `Dataset`, but not when combining `IterableDataset`. Specifically, when applying `concatenate_datasets` to a list of `IterableDataset` in Pytorch format (i.e. using `.with_format(Pytorch)`), the output `IterableDataset` is not in Pytorch format.
### Steps to reproduce the bug
```
import datasets
ds = datasets.Dataset.from_dict({"a": [1,2,3]})
iterable_ds = ds.to_iterable_dataset()
datasets.concatenate_datasets([ds.with_format("torch")]) # <- this preserves Pytorch format
datasets.concatenate_datasets([iterable_ds.with_format("torch")]) # <- this does NOT preserves Pytorch format
```
### Expected behavior
Pytorch format should be preserved when combining IterableDataset in Pytorch format.
### Environment info
datasets==3.5.0, Python 3.11.11, torch==2.2.2 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7515/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7515/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7514 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7514/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7514/comments | https://api.github.com/repos/huggingface/datasets/issues/7514/events | https://github.com/huggingface/datasets/pull/7514 | 2,994,714,923 | PR_kwDODunzps6Sk7Et | 7,514 | Do not hash `generator` in `BuilderConfig.create_config_id` | {
"avatar_url": "https://avatars.githubusercontent.com/u/43753582?v=4",
"events_url": "https://api.github.com/users/simonreise/events{/privacy}",
"followers_url": "https://api.github.com/users/simonreise/followers",
"following_url": "https://api.github.com/users/simonreise/following{/other_user}",
"gists_url": "https://api.github.com/users/simonreise/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/simonreise",
"id": 43753582,
"login": "simonreise",
"node_id": "MDQ6VXNlcjQzNzUzNTgy",
"organizations_url": "https://api.github.com/users/simonreise/orgs",
"received_events_url": "https://api.github.com/users/simonreise/received_events",
"repos_url": "https://api.github.com/users/simonreise/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/simonreise/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/simonreise/subscriptions",
"type": "User",
"url": "https://api.github.com/users/simonreise",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [] | 2025-04-15T01:26:43Z | 2025-04-23T11:55:55Z | 2025-04-15T16:27:51Z | NONE | null | null | null | `Dataset.from_generator` function passes all of its arguments to `BuilderConfig.create_config_id`, including generator function itself. `BuilderConfig.create_config_id` function tries to hash all the args, and hashing a `generator` can take a large amount of time or even cause MemoryError if the dataset processed in a generator function is large enough.
Maybe we should pop generator from `config_kwargs_to_add_to_suffix` before hashing to avoid it.
There is a more detailed description of the problem this PR solves in #7513 | {
"avatar_url": "https://avatars.githubusercontent.com/u/43753582?v=4",
"events_url": "https://api.github.com/users/simonreise/events{/privacy}",
"followers_url": "https://api.github.com/users/simonreise/followers",
"following_url": "https://api.github.com/users/simonreise/following{/other_user}",
"gists_url": "https://api.github.com/users/simonreise/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/simonreise",
"id": 43753582,
"login": "simonreise",
"node_id": "MDQ6VXNlcjQzNzUzNTgy",
"organizations_url": "https://api.github.com/users/simonreise/orgs",
"received_events_url": "https://api.github.com/users/simonreise/received_events",
"repos_url": "https://api.github.com/users/simonreise/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/simonreise/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/simonreise/subscriptions",
"type": "User",
"url": "https://api.github.com/users/simonreise",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7514/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7514/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7514.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7514",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7514.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7514"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7513 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7513/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7513/comments | https://api.github.com/repos/huggingface/datasets/issues/7513/events | https://github.com/huggingface/datasets/issues/7513 | 2,994,678,437 | I_kwDODunzps6yfyql | 7,513 | MemoryError while creating dataset from generator | {
"avatar_url": "https://avatars.githubusercontent.com/u/43753582?v=4",
"events_url": "https://api.github.com/users/simonreise/events{/privacy}",
"followers_url": "https://api.github.com/users/simonreise/followers",
"following_url": "https://api.github.com/users/simonreise/following{/other_user}",
"gists_url": "https://api.github.com/users/simonreise/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/simonreise",
"id": 43753582,
"login": "simonreise",
"node_id": "MDQ6VXNlcjQzNzUzNTgy",
"organizations_url": "https://api.github.com/users/simonreise/orgs",
"received_events_url": "https://api.github.com/users/simonreise/received_events",
"repos_url": "https://api.github.com/users/simonreise/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/simonreise/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/simonreise/subscriptions",
"type": "User",
"url": "https://api.github.com/users/simonreise",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Upd: created a PR that can probably solve the problem: #7514",
"Hi ! We need to take the generator into account for the cache. The generator is hashed to make the dataset fingerprint used by the cache. This way you can reload the Dataset from the cache without regenerating in subsequent `from_generator` calls.\n\nMaybe instead of removing generator from the hasher input, we can let users pass their own Dataset fingerprint to `from_generator`, and if it's specified we don't need to hash anything",
"Upd: I successfully generated a dataset from my large geospatial data with `generator` excluded from hashing and saved it to disk without running into memory errors. So, it looks like there are no other bottlenecks in dataset generation in my case\n\nMaybe letting users pass their own fingerprint to skip hashing can be a great solution to that issue!",
"@lhoestq I tried to implement user-defined dataset fingerprint in #7533 . Am I doing it right?"
] | 2025-04-15T01:02:02Z | 2025-04-23T19:37:08Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
# TL:DR
`Dataset.from_generator` function passes all of its arguments to `BuilderConfig.create_config_id`, including `generator` function itself. `BuilderConfig.create_config_id` function tries to hash all the args, which can take a large amount of time or even cause MemoryError if the dataset processed in a generator function is large enough.
Maybe we should pop `generator` from `config_kwargs_to_add_to_suffix` before hashing to avoid it.
# Full description
I have a pretty large spatial imagery dataset that is generated from two xbatcher.BatchGenerators via custom `dataset_generator` function that looks like this if simplified:
```
def dataset_generator():
for index in samples:
data_dict = {
"key": index,
"x": x_batches[index].data,
"y": y_batches[index].data,
}
yield data_dict
```
Then I use `datasets.Dataset.from_generator` to generate the dataset itself.
```
# Create dataset
ds = datasets.Dataset.from_generator(
dataset_generator,
features=feat,
cache_dir=(output / ".cache"),
)
```
It works nicely with pretty small data, but if the dataset is huge and barely fits in memory, it crashes with memory error:
<details>
<summary>Full stack trace</summary>
```
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\remote_sensing_processor\segmentation\semantic\tiles.py:248](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/remote_sensing_processor/segmentation/semantic/tiles.py#line=247), in generate_tiles(x, y, output, tile_size, shuffle, split, x_dtype, y_dtype, x_nodata, y_nodata)
245 yield data_dict
247 # Create dataset
--> 248 ds = datasets.Dataset.from_generator(
249 dataset_generator,
250 features=feat,
251 cache_dir=(output / ".cache"),
252 )
254 # Save dataset
255 ds.save_to_disk(output / name)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\arrow_dataset.py:1105](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/arrow_dataset.py#line=1104), in Dataset.from_generator(generator, features, cache_dir, keep_in_memory, gen_kwargs, num_proc, split, **kwargs)
1052 """Create a Dataset from a generator.
1053
1054 Args:
(...) 1101 ```
1102 """
1103 from .io.generator import GeneratorDatasetInputStream
-> 1105 return GeneratorDatasetInputStream(
1106 generator=generator,
1107 features=features,
1108 cache_dir=cache_dir,
1109 keep_in_memory=keep_in_memory,
1110 gen_kwargs=gen_kwargs,
1111 num_proc=num_proc,
1112 split=split,
1113 **kwargs,
1114 ).read()
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\io\generator.py:29](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/io/generator.py#line=28), in GeneratorDatasetInputStream.__init__(self, generator, features, cache_dir, keep_in_memory, streaming, gen_kwargs, num_proc, split, **kwargs)
9 def __init__(
10 self,
11 generator: Callable,
(...) 19 **kwargs,
20 ):
21 super().__init__(
22 features=features,
23 cache_dir=cache_dir,
(...) 27 **kwargs,
28 )
---> 29 self.builder = Generator(
30 cache_dir=cache_dir,
31 features=features,
32 generator=generator,
33 gen_kwargs=gen_kwargs,
34 split=split,
35 **kwargs,
36 )
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\builder.py:343](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/builder.py#line=342), in DatasetBuilder.__init__(self, cache_dir, dataset_name, config_name, hash, base_path, info, features, token, repo_id, data_files, data_dir, storage_options, writer_batch_size, **config_kwargs)
341 config_kwargs["data_dir"] = data_dir
342 self.config_kwargs = config_kwargs
--> 343 self.config, self.config_id = self._create_builder_config(
344 config_name=config_name,
345 custom_features=features,
346 **config_kwargs,
347 )
349 # prepare info: DatasetInfo are a standardized dataclass across all datasets
350 # Prefill datasetinfo
351 if info is None:
352 # TODO FOR PACKAGED MODULES IT IMPORTS DATA FROM src/packaged_modules which doesn't make sense
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\builder.py:604](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/builder.py#line=603), in DatasetBuilder._create_builder_config(self, config_name, custom_features, **config_kwargs)
598 builder_config._resolve_data_files(
599 base_path=self.base_path,
600 download_config=DownloadConfig(token=self.token, storage_options=self.storage_options),
601 )
603 # compute the config id that is going to be used for caching
--> 604 config_id = builder_config.create_config_id(
605 config_kwargs,
606 custom_features=custom_features,
607 )
608 is_custom = (config_id not in self.builder_configs) and config_id != "default"
609 if is_custom:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\builder.py:187](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/builder.py#line=186), in BuilderConfig.create_config_id(self, config_kwargs, custom_features)
185 suffix = Hasher.hash(config_kwargs_to_add_to_suffix)
186 else:
--> 187 suffix = Hasher.hash(config_kwargs_to_add_to_suffix)
189 if custom_features is not None:
190 m = Hasher()
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\fingerprint.py:188](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/fingerprint.py#line=187), in Hasher.hash(cls, value)
186 @classmethod
187 def hash(cls, value: Any) -> str:
--> 188 return cls.hash_bytes(dumps(value))
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:109](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=108), in dumps(obj)
107 """Pickle an object to a string."""
108 file = BytesIO()
--> 109 dump(obj, file)
110 return file.getvalue()
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:103](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=102), in dump(obj, file)
101 def dump(obj, file):
102 """Pickle an object to a file."""
--> 103 Pickler(file, recurse=True).dump(obj)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:420](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=419), in Pickler.dump(self, obj)
418 def dump(self, obj): #NOTE: if settings change, need to update attributes
419 logger.trace_setup(self)
--> 420 StockPickler.dump(self, obj)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:484](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=483), in _Pickler.dump(self, obj)
482 if self.proto >= 4:
483 self.framer.start_framing()
--> 484 self.save(obj)
485 self.write(STOP)
486 self.framer.end_framing()
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1217](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1216), in save_module_dict(pickler, obj)
1214 if is_dill(pickler, child=False) and pickler._session:
1215 # we only care about session the first pass thru
1216 pickler._first_pass = False
-> 1217 StockPickler.save_dict(pickler, obj)
1218 logger.trace(pickler, "# D2")
1219 return
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:990](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=989), in _Pickler.save_dict(self, obj)
987 self.write(MARK + DICT)
989 self.memoize(obj)
--> 990 self._batch_setitems(obj.items())
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:83](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=82), in Pickler._batch_setitems(self, items)
80 from datasets.fingerprint import Hasher
82 items = sorted(items, key=lambda x: Hasher.hash(x[0]))
---> 83 dill.Pickler._batch_setitems(self, items)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:1014](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=1013), in _Pickler._batch_setitems(self, items)
1012 for k, v in tmp:
1013 save(k)
-> 1014 save(v)
1015 write(SETITEMS)
1016 elif n:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1985](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1984), in save_function(pickler, obj)
1982 if state_dict:
1983 state = state, state_dict
-> 1985 _save_with_postproc(pickler, (_create_function, (
1986 obj.__code__, globs, obj.__name__, obj.__defaults__,
1987 closure
1988 ), state), obj=obj, postproc_list=postproc_list)
1990 # Lift closure cell update to earliest function (#458)
1991 if _postproc:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1117](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1116), in _save_with_postproc(pickler, reduction, is_pickler_dill, obj, postproc_list)
1115 continue
1116 else:
-> 1117 pickler.save_reduce(*reduction)
1118 # pop None created by calling preprocessing step off stack
1119 pickler.write(POP)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:690](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=689), in _Pickler.save_reduce(self, func, args, state, listitems, dictitems, state_setter, obj)
688 else:
689 save(func)
--> 690 save(args)
691 write(REDUCE)
693 if obj is not None:
694 # If the object is already in the memo, this means it is
695 # recursive. In this case, throw away everything we put on the
696 # stack, and fetch the object back from the memo.
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:905](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=904), in _Pickler.save_tuple(self, obj)
903 if n <= 3 and self.proto >= 2:
904 for element in obj:
--> 905 save(element)
906 # Subtle. Same as in the big comment below.
907 if id(obj) in memo:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:601](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=600), in _Pickler.save(self, obj, save_persistent_id)
597 raise PicklingError("Tuple returned by %s must have "
598 "two to six elements" % reduce)
600 # Save the reduce() output and finally memoize the object
--> 601 self.save_reduce(obj=obj, *rv)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:715](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=714), in _Pickler.save_reduce(self, func, args, state, listitems, dictitems, state_setter, obj)
713 if state is not None:
714 if state_setter is None:
--> 715 save(state)
716 write(BUILD)
717 else:
718 # If a state_setter is specified, call it instead of load_build
719 # to update obj's with its previous state.
720 # First, push state_setter and its tuple of expected arguments
721 # (obj, state) onto the stack.
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1217](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1216), in save_module_dict(pickler, obj)
1214 if is_dill(pickler, child=False) and pickler._session:
1215 # we only care about session the first pass thru
1216 pickler._first_pass = False
-> 1217 StockPickler.save_dict(pickler, obj)
1218 logger.trace(pickler, "# D2")
1219 return
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:990](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=989), in _Pickler.save_dict(self, obj)
987 self.write(MARK + DICT)
989 self.memoize(obj)
--> 990 self._batch_setitems(obj.items())
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:83](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=82), in Pickler._batch_setitems(self, items)
80 from datasets.fingerprint import Hasher
82 items = sorted(items, key=lambda x: Hasher.hash(x[0]))
---> 83 dill.Pickler._batch_setitems(self, items)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:1014](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=1013), in _Pickler._batch_setitems(self, items)
1012 for k, v in tmp:
1013 save(k)
-> 1014 save(v)
1015 write(SETITEMS)
1016 elif n:
[... skipping similar frames: Pickler.save at line 70 (1 times), Pickler.save at line 414 (1 times)]
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:601](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=600), in _Pickler.save(self, obj, save_persistent_id)
597 raise PicklingError("Tuple returned by %s must have "
598 "two to six elements" % reduce)
600 # Save the reduce() output and finally memoize the object
--> 601 self.save_reduce(obj=obj, *rv)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:715](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=714), in _Pickler.save_reduce(self, func, args, state, listitems, dictitems, state_setter, obj)
713 if state is not None:
714 if state_setter is None:
--> 715 save(state)
716 write(BUILD)
717 else:
718 # If a state_setter is specified, call it instead of load_build
719 # to update obj's with its previous state.
720 # First, push state_setter and its tuple of expected arguments
721 # (obj, state) onto the stack.
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:905](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=904), in _Pickler.save_tuple(self, obj)
903 if n <= 3 and self.proto >= 2:
904 for element in obj:
--> 905 save(element)
906 # Subtle. Same as in the big comment below.
907 if id(obj) in memo:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1217](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1216), in save_module_dict(pickler, obj)
1214 if is_dill(pickler, child=False) and pickler._session:
1215 # we only care about session the first pass thru
1216 pickler._first_pass = False
-> 1217 StockPickler.save_dict(pickler, obj)
1218 logger.trace(pickler, "# D2")
1219 return
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:990](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=989), in _Pickler.save_dict(self, obj)
987 self.write(MARK + DICT)
989 self.memoize(obj)
--> 990 self._batch_setitems(obj.items())
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:83](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=82), in Pickler._batch_setitems(self, items)
80 from datasets.fingerprint import Hasher
82 items = sorted(items, key=lambda x: Hasher.hash(x[0]))
---> 83 dill.Pickler._batch_setitems(self, items)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:1014](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=1013), in _Pickler._batch_setitems(self, items)
1012 for k, v in tmp:
1013 save(k)
-> 1014 save(v)
1015 write(SETITEMS)
1016 elif n:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:601](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=600), in _Pickler.save(self, obj, save_persistent_id)
597 raise PicklingError("Tuple returned by %s must have "
598 "two to six elements" % reduce)
600 # Save the reduce() output and finally memoize the object
--> 601 self.save_reduce(obj=obj, *rv)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:715](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=714), in _Pickler.save_reduce(self, func, args, state, listitems, dictitems, state_setter, obj)
713 if state is not None:
714 if state_setter is None:
--> 715 save(state)
716 write(BUILD)
717 else:
718 # If a state_setter is specified, call it instead of load_build
719 # to update obj's with its previous state.
720 # First, push state_setter and its tuple of expected arguments
721 # (obj, state) onto the stack.
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:905](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=904), in _Pickler.save_tuple(self, obj)
903 if n <= 3 and self.proto >= 2:
904 for element in obj:
--> 905 save(element)
906 # Subtle. Same as in the big comment below.
907 if id(obj) in memo:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1217](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1216), in save_module_dict(pickler, obj)
1214 if is_dill(pickler, child=False) and pickler._session:
1215 # we only care about session the first pass thru
1216 pickler._first_pass = False
-> 1217 StockPickler.save_dict(pickler, obj)
1218 logger.trace(pickler, "# D2")
1219 return
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:990](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=989), in _Pickler.save_dict(self, obj)
987 self.write(MARK + DICT)
989 self.memoize(obj)
--> 990 self._batch_setitems(obj.items())
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:83](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=82), in Pickler._batch_setitems(self, items)
80 from datasets.fingerprint import Hasher
82 items = sorted(items, key=lambda x: Hasher.hash(x[0]))
---> 83 dill.Pickler._batch_setitems(self, items)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:1014](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=1013), in _Pickler._batch_setitems(self, items)
1012 for k, v in tmp:
1013 save(k)
-> 1014 save(v)
1015 write(SETITEMS)
1016 elif n:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:601](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=600), in _Pickler.save(self, obj, save_persistent_id)
597 raise PicklingError("Tuple returned by %s must have "
598 "two to six elements" % reduce)
600 # Save the reduce() output and finally memoize the object
--> 601 self.save_reduce(obj=obj, *rv)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:690](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=689), in _Pickler.save_reduce(self, func, args, state, listitems, dictitems, state_setter, obj)
688 else:
689 save(func)
--> 690 save(args)
691 write(REDUCE)
693 if obj is not None:
694 # If the object is already in the memo, this means it is
695 # recursive. In this case, throw away everything we put on the
696 # stack, and fetch the object back from the memo.
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:920](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=919), in _Pickler.save_tuple(self, obj)
918 write(MARK)
919 for element in obj:
--> 920 save(element)
922 if id(obj) in memo:
923 # Subtle. d was not in memo when we entered save_tuple(), so
924 # the process of saving the tuple's elements must have saved
(...) 928 # could have been done in the "for element" loop instead, but
929 # recursive tuples are a rare thing.
930 get = self.get(memo[id(obj)][0])
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:601](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=600), in _Pickler.save(self, obj, save_persistent_id)
597 raise PicklingError("Tuple returned by %s must have "
598 "two to six elements" % reduce)
600 # Save the reduce() output and finally memoize the object
--> 601 self.save_reduce(obj=obj, *rv)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:715](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=714), in _Pickler.save_reduce(self, func, args, state, listitems, dictitems, state_setter, obj)
713 if state is not None:
714 if state_setter is None:
--> 715 save(state)
716 write(BUILD)
717 else:
718 # If a state_setter is specified, call it instead of load_build
719 # to update obj's with its previous state.
720 # First, push state_setter and its tuple of expected arguments
721 # (obj, state) onto the stack.
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1217](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1216), in save_module_dict(pickler, obj)
1214 if is_dill(pickler, child=False) and pickler._session:
1215 # we only care about session the first pass thru
1216 pickler._first_pass = False
-> 1217 StockPickler.save_dict(pickler, obj)
1218 logger.trace(pickler, "# D2")
1219 return
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:990](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=989), in _Pickler.save_dict(self, obj)
987 self.write(MARK + DICT)
989 self.memoize(obj)
--> 990 self._batch_setitems(obj.items())
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:83](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=82), in Pickler._batch_setitems(self, items)
80 from datasets.fingerprint import Hasher
82 items = sorted(items, key=lambda x: Hasher.hash(x[0]))
---> 83 dill.Pickler._batch_setitems(self, items)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:1014](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=1013), in _Pickler._batch_setitems(self, items)
1012 for k, v in tmp:
1013 save(k)
-> 1014 save(v)
1015 write(SETITEMS)
1016 elif n:
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1217](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1216), in save_module_dict(pickler, obj)
1214 if is_dill(pickler, child=False) and pickler._session:
1215 # we only care about session the first pass thru
1216 pickler._first_pass = False
-> 1217 StockPickler.save_dict(pickler, obj)
1218 logger.trace(pickler, "# D2")
1219 return
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:990](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=989), in _Pickler.save_dict(self, obj)
987 self.write(MARK + DICT)
989 self.memoize(obj)
--> 990 self._batch_setitems(obj.items())
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:83](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=82), in Pickler._batch_setitems(self, items)
80 from datasets.fingerprint import Hasher
82 items = sorted(items, key=lambda x: Hasher.hash(x[0]))
---> 83 dill.Pickler._batch_setitems(self, items)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:1019](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=1018), in _Pickler._batch_setitems(self, items)
1017 k, v = tmp[0]
1018 save(k)
-> 1019 save(v)
1020 write(SETITEM)
1021 # else tmp is empty, and we're done
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:601](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=600), in _Pickler.save(self, obj, save_persistent_id)
597 raise PicklingError("Tuple returned by %s must have "
598 "two to six elements" % reduce)
600 # Save the reduce() output and finally memoize the object
--> 601 self.save_reduce(obj=obj, *rv)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:715](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=714), in _Pickler.save_reduce(self, func, args, state, listitems, dictitems, state_setter, obj)
713 if state is not None:
714 if state_setter is None:
--> 715 save(state)
716 write(BUILD)
717 else:
718 # If a state_setter is specified, call it instead of load_build
719 # to update obj's with its previous state.
720 # First, push state_setter and its tuple of expected arguments
721 # (obj, state) onto the stack.
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1217](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1216), in save_module_dict(pickler, obj)
1214 if is_dill(pickler, child=False) and pickler._session:
1215 # we only care about session the first pass thru
1216 pickler._first_pass = False
-> 1217 StockPickler.save_dict(pickler, obj)
1218 logger.trace(pickler, "# D2")
1219 return
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:990](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=989), in _Pickler.save_dict(self, obj)
987 self.write(MARK + DICT)
989 self.memoize(obj)
--> 990 self._batch_setitems(obj.items())
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:83](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=82), in Pickler._batch_setitems(self, items)
80 from datasets.fingerprint import Hasher
82 items = sorted(items, key=lambda x: Hasher.hash(x[0]))
---> 83 dill.Pickler._batch_setitems(self, items)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:1014](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=1013), in _Pickler._batch_setitems(self, items)
1012 for k, v in tmp:
1013 save(k)
-> 1014 save(v)
1015 write(SETITEMS)
1016 elif n:
[... skipping similar frames: Pickler.save at line 70 (1 times), Pickler.save at line 414 (1 times)]
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:1217](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=1216), in save_module_dict(pickler, obj)
1214 if is_dill(pickler, child=False) and pickler._session:
1215 # we only care about session the first pass thru
1216 pickler._first_pass = False
-> 1217 StockPickler.save_dict(pickler, obj)
1218 logger.trace(pickler, "# D2")
1219 return
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:990](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=989), in _Pickler.save_dict(self, obj)
987 self.write(MARK + DICT)
989 self.memoize(obj)
--> 990 self._batch_setitems(obj.items())
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:83](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=82), in Pickler._batch_setitems(self, items)
80 from datasets.fingerprint import Hasher
82 items = sorted(items, key=lambda x: Hasher.hash(x[0]))
---> 83 dill.Pickler._batch_setitems(self, items)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:1014](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=1013), in _Pickler._batch_setitems(self, items)
1012 for k, v in tmp:
1013 save(k)
-> 1014 save(v)
1015 write(SETITEMS)
1016 elif n:
[... skipping similar frames: Pickler.save at line 70 (1 times), Pickler.save at line 414 (1 times)]
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:601](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=600), in _Pickler.save(self, obj, save_persistent_id)
597 raise PicklingError("Tuple returned by %s must have "
598 "two to six elements" % reduce)
600 # Save the reduce() output and finally memoize the object
--> 601 self.save_reduce(obj=obj, *rv)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:715](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=714), in _Pickler.save_reduce(self, func, args, state, listitems, dictitems, state_setter, obj)
713 if state is not None:
714 if state_setter is None:
--> 715 save(state)
716 write(BUILD)
717 else:
718 # If a state_setter is specified, call it instead of load_build
719 # to update obj's with its previous state.
720 # First, push state_setter and its tuple of expected arguments
721 # (obj, state) onto the stack.
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:920](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=919), in _Pickler.save_tuple(self, obj)
918 write(MARK)
919 for element in obj:
--> 920 save(element)
922 if id(obj) in memo:
923 # Subtle. d was not in memo when we entered save_tuple(), so
924 # the process of saving the tuple's elements must have saved
(...) 928 # could have been done in the "for element" loop instead, but
929 # recursive tuples are a rare thing.
930 get = self.get(memo[id(obj)][0])
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\datasets\utils\_dill.py:70](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/datasets/utils/_dill.py#line=69), in Pickler.save(self, obj, save_persistent_id)
68 if obj_type is FunctionType:
69 obj = getattr(obj, "_torchdynamo_orig_callable", obj)
---> 70 dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\site-packages\dill\_dill.py:414](file:///C:/ProgramData/miniforge3/envs/geo/Lib/site-packages/dill/_dill.py#line=413), in Pickler.save(self, obj, save_persistent_id)
412 msg = "Can't pickle %s: attribute lookup builtins.generator failed" % GeneratorType
413 raise PicklingError(msg)
--> 414 StockPickler.save(self, obj, save_persistent_id)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:558](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=557), in _Pickler.save(self, obj, save_persistent_id)
556 f = self.dispatch.get(t)
557 if f is not None:
--> 558 f(self, obj) # Call unbound method with explicit self
559 return
561 # Check private dispatch table if any, or else
562 # copyreg.dispatch_table
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:809](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=808), in _Pickler.save_bytes(self, obj)
806 self.save_reduce(codecs.encode,
807 (str(obj, 'latin1'), 'latin1'), obj=obj)
808 return
--> 809 self._save_bytes_no_memo(obj)
810 self.memoize(obj)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:797](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=796), in _Pickler._save_bytes_no_memo(self, obj)
795 self._write_large_bytes(BINBYTES8 + pack("<Q", n), obj)
796 elif n >= self.framer._FRAME_SIZE_TARGET:
--> 797 self._write_large_bytes(BINBYTES + pack("<I", n), obj)
798 else:
799 self.write(BINBYTES + pack("<I", n) + obj)
File [C:\ProgramData\miniforge3\envs\geo\Lib\pickle.py:254](file:///C:/ProgramData/miniforge3/envs/geo/Lib/pickle.py#line=253), in _Framer.write_large_bytes(self, header, payload)
247 # Perform direct write of the header and payload of the large binary
248 # object. Be careful not to concatenate the header and the payload
249 # prior to calling 'write' as we do not want to allocate a large
250 # temporary bytes object.
251 # We intentionally do not insert a protocol 4 frame opcode to make
252 # it possible to optimize file.read calls in the loader.
253 write(header)
--> 254 write(payload)
MemoryError:
```
</details>
Memory error is an expected type of error in such case, but when I started digging down, I found out that it occurs in a kinda unexpected place - in `create_config_id` function. It tries to hash `config_kwargs_to_add_to_suffix`, including generator function itself.
I modified the `BuilderConfig.create_config_id` code like this to check which values are hashed and how much time it takes to hash them and ran it on a toy dataset:
```
print(config_kwargs_to_add_to_suffix)
start_time = time.time()
if all(isinstance(v, (str, bool, int, float)) for v in config_kwargs_to_add_to_suffix.values()):
suffix = ",".join(
str(k) + "=" + urllib.parse.quote_plus(str(v)) for k, v in config_kwargs_to_add_to_suffix.items()
)
if len(suffix) > 32: # hash if too long
suffix = Hasher.hash(config_kwargs_to_add_to_suffix)
else:
suffix = Hasher.hash(config_kwargs_to_add_to_suffix)
end_time = time.time()
print(f"Execution time: {end_time - start_time:.4f} seconds")
print(suffix)
```
In my case the content of `config_kwargs_to_add_to_suffix` was like this:
```
{'features': {'key': Value(dtype='int64', id=None), 'x': Array3D(shape=(44, 128, 128), dtype='float32', id=None), 'y_class': Array2D(shape=(128, 128), dtype='int32', id=None)}, 'gen_kwargs': None, 'generator': <function generate_tiles.<locals>.dataset_generator at 0x00000139D10D7920>, 'split': NamedSplit('train')}
```
Also I noticed that hashing took a significant amount of time - 43.1482 seconds, while the overall function execution (with data loading, batching and saving dataset) took 2min 45s. The output of `create_config_id` is just a dataset id, so, it is inappropirately large amount of time.
But when I added `config_kwargs_to_add_to_suffix.pop("generator", None)`, the hashing took only 0.0060 seconds.
Maybe we shouldn't hash the generator function, as it can be really computationally and memory expensive.
### Steps to reproduce the bug
This is a simplified example of a workflow I used to generate dataset. But I think that you can use almost any workflow to reproduce that bug.
```
import pystac
import pystac_client
import planetary_computer
import numpy as np
import xarray as xr
import rioxarray as rxr
import dask
import xbatcher
import datasets
# Loading a dataset, in our case - single Landsat image
catalog = pystac_client.Client.open(
"https://planetarycomputer.microsoft.com/api/stac/v1",
modifier=planetary_computer.sign_inplace,
)
brazil = [-60.2, -3.31]
time_of_interest = "2021-06-01/2021-08-31"
search = catalog.search(collections=["landsat-c2-l2"], intersects={"type": "Point", "coordinates": brazil}, datetime=time_of_interest)
items = search.item_collection()
item = min(items, key=lambda item: pystac.extensions.eo.EOExtension.ext(item).cloud_cover)
# Getting x data
bands = []
for band in ["red", "green", "blue", "nir08", "coastal", "swir16", "swir22", "lwir11"]:
with rxr.open_rasterio(item.assets[band].href, chunks=True, lock=True) as raster:
raster = raster.to_dataset('band')
#print(raster)
raster = raster.rename({1: band})
bands.append(raster)
x = xr.merge(bands).squeeze().to_array("band").persist()
# Getting y data
with rxr.open_rasterio(item.assets['qa_pixel'].href, chunks=True, lock=True) as raster:
y = raster.squeeze().persist()
# Setting up batches generators
x_batches = xbatcher.BatchGenerator(ds=x, input_dims={"x": 256, "y": 256})
y_batches = xbatcher.BatchGenerator(ds=y, input_dims={"x": 256, "y": 256})
# Filtering samples that contain only nodata
samples = list(range(len(x_batches)))
samples_filtered = []
for i in samples:
if not np.array_equal(np.unique(x_batches[i]), np.array([0.])) and not np.array_equal(np.unique(y_batches[i]), np.array([0])):
samples_filtered.append(i)
samples = samples_filtered
np.random.shuffle(samples)
# Setting up features
feat = {
"key": datasets.Value(dtype="int64"),
"x": datasets.Array3D(dtype="float32", shape=(4, 256, 256)),
"y": datasets.Array2D(dtype="int32", shape=(256, 256))
}
feat = datasets.Features(feat)
# Setting up a generator
def dataset_generator():
for index in samples:
data_dict = {
"key": index,
"x": x_batches[index].data,
"y": y_batches[index].data,
}
yield data_dict
# Create dataset
ds = datasets.Dataset.from_generator(
dataset_generator,
features=feat,
cache_dir="temp/cache",
)
```
Please, try adding `config_kwargs_to_add_to_suffix.pop("generator", None)` to `BuilderConfig.create_config_id` and then measuring how much time it takes to run
```
if all(isinstance(v, (str, bool, int, float)) for v in config_kwargs_to_add_to_suffix.values()):
suffix = ",".join(
str(k) + "=" + urllib.parse.quote_plus(str(v)) for k, v in config_kwargs_to_add_to_suffix.items()
)
if len(suffix) > 32: # hash if too long
suffix = Hasher.hash(config_kwargs_to_add_to_suffix)
else:
suffix = Hasher.hash(config_kwargs_to_add_to_suffix)
```
code block with and without `config_kwargs_to_add_to_suffix.pop("generator", None)`
In my case the difference was 3.3828 seconds without popping generator function and 0.0010 seconds with popping.
### Expected behavior
Much faster hashing and no MemoryErrors
### Environment info
- `datasets` version: 3.5.0
- Platform: Windows-11-10.0.26100-SP0
- Python version: 3.12.9
- `huggingface_hub` version: 0.30.1
- PyArrow version: 17.0.0
- Pandas version: 2.2.2
- `fsspec` version: 2024.12.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7513/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7513/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7512 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7512/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7512/comments | https://api.github.com/repos/huggingface/datasets/issues/7512/events | https://github.com/huggingface/datasets/issues/7512 | 2,994,043,544 | I_kwDODunzps6ydXqY | 7,512 | .map() fails if function uses pyvista | {
"avatar_url": "https://avatars.githubusercontent.com/u/11832922?v=4",
"events_url": "https://api.github.com/users/el-hult/events{/privacy}",
"followers_url": "https://api.github.com/users/el-hult/followers",
"following_url": "https://api.github.com/users/el-hult/following{/other_user}",
"gists_url": "https://api.github.com/users/el-hult/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/el-hult",
"id": 11832922,
"login": "el-hult",
"node_id": "MDQ6VXNlcjExODMyOTIy",
"organizations_url": "https://api.github.com/users/el-hult/orgs",
"received_events_url": "https://api.github.com/users/el-hult/received_events",
"repos_url": "https://api.github.com/users/el-hult/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/el-hult/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/el-hult/subscriptions",
"type": "User",
"url": "https://api.github.com/users/el-hult",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"I found a similar (?) issue in https://github.com/huggingface/datasets/issues/6435, where someone had issues with forks and CUDA. According to https://huggingface.co/docs/datasets/main/en/process#multiprocessing we should do \n\n```\nfrom multiprocess import set_start_method\nset_start_method(\"spawn\")\n```\n\nto avoid the fork. The updated code\n\n```python\nimport numpy as np\nimport pyvista as pv\nimport datasets\nimport multiprocess\n\ndata = [{\"coords\": np.random.rand(5, 3)} for _ in range(3)]\n\ndef render_point(example):\n plotter = pv.Plotter(off_screen=True)\n cloud = pv.PolyData(example[\"coords\"])\n plotter.add_mesh(cloud)\n img = plotter.screenshot(return_img=True)\n return {\"image\": img}\n\n\n# breaks if num_proc>1\nmultiprocess.set_start_method(\"spawn\")\nds = datasets.Dataset.from_list(data).map(render_point, num_proc=2)\n```\n\ninstead fails with `TypeError: fork_exec() takes exactly 23 arguments (21 given)` which also seems like a bug to me."
] | 2025-04-14T19:43:02Z | 2025-04-14T20:01:53Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Using PyVista inside a .map() produces a crash with `objc[78796]: +[NSResponder initialize] may have been in progress in another thread when fork() was called. We cannot safely call it or ignore it in the fork() child process. Crashing instead. Set a breakpoint on objc_initializeAfterForkError to debug.`
### Steps to reproduce the bug
Run
```python
import numpy as np
import pyvista as pv
import datasets
data = [{"coords": np.random.rand(5, 3)} for _ in range(3)]
def render_point(example):
plotter = pv.Plotter(off_screen=True)
cloud = pv.PolyData(example["coords"])
plotter.add_mesh(cloud)
img = plotter.screenshot(return_img=True)
return {"image": img}
# breaks if num_proc>1
ds = datasets.Dataset.from_list(data).map(render_point, num_proc=2)
```
### Expected behavior
It should work. Just like when I use a process pool to make it work.
```python
import numpy as np
import pyvista as pv
import multiprocessing
data = [{"coords": np.random.rand(5, 3)} for _ in range(3)]
def render_point(example):
plotter = pv.Plotter(off_screen=True)
cloud = pv.PolyData(example["coords"])
plotter.add_mesh(cloud)
img = plotter.screenshot(return_img=True)
return {"image": img}
if __name__ == "__main__":
with multiprocessing.Pool(processes=2) as pool:
results = pool.map(render_point, data)
print(results[0]["image"].shape)
```
### Environment info
- `datasets` version: 3.3.2
- Platform: macOS-15.3.2-arm64-arm-64bit
- Python version: 3.11.10
- `huggingface_hub` version: 0.28.1
- PyArrow version: 18.1.0
- Pandas version: 2.2.3
- `fsspec` version: 2024.10.0
And then I suppose pyvista info is good to have.
```python
import pyvista as pv
print(pv.Report())
```
gives
--------------------------------------------------------------------------------
Date: Mon Apr 14 21:42:08 2025 CEST
OS : Darwin (macOS 15.3.2)
CPU(s) : 10
Machine : arm64
Architecture : 64bit
RAM : 32.0 GiB
Environment : IPython
File system : apfs
GPU Vendor : Apple
GPU Renderer : Apple M1 Max
GPU Version : 4.1 Metal - 89.3
MathText Support : True
Python 3.11.10 (main, Oct 7 2024, 23:25:27) [Clang 18.1.8 ]
pyvista : 0.44.2
vtk : 9.4.0
numpy : 2.2.2
matplotlib : 3.10.0
scooby : 0.10.0
pooch : 1.8.2
pillow : 11.1.0
imageio : 2.36.1
PyQt5 : 5.15.11
IPython : 8.30.0
scipy : 1.14.1
tqdm : 4.67.1
jupyterlab : 4.3.5
nest_asyncio : 1.6.0
-------------------------------------------------------------------------------- | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7512/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7512/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7510 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7510/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7510/comments | https://api.github.com/repos/huggingface/datasets/issues/7510/events | https://github.com/huggingface/datasets/issues/7510 | 2,992,131,117 | I_kwDODunzps6yWEwt | 7,510 | Incompatibile dill version (0.3.9) in datasets 2.18.0 - 3.5.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/98061329?v=4",
"events_url": "https://api.github.com/users/JGrel/events{/privacy}",
"followers_url": "https://api.github.com/users/JGrel/followers",
"following_url": "https://api.github.com/users/JGrel/following{/other_user}",
"gists_url": "https://api.github.com/users/JGrel/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/JGrel",
"id": 98061329,
"login": "JGrel",
"node_id": "U_kgDOBdhMEQ",
"organizations_url": "https://api.github.com/users/JGrel/orgs",
"received_events_url": "https://api.github.com/users/JGrel/received_events",
"repos_url": "https://api.github.com/users/JGrel/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/JGrel/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/JGrel/subscriptions",
"type": "User",
"url": "https://api.github.com/users/JGrel",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! We can bump `dill` to 0.3.9 if we make sure it's deterministic and doesn't break the caching mechanism in `datasets`.\n\nWould you be interested in opening a PR ? Then we can run the CI to see if it works",
"Hi!. Yeah I can do it. Should I make any changes besides dill versions?",
"There are probably some usage of internal functions from `dill` that we'll need to update in `datasets`\n\nIf you run `pytest tests/test_fingerprint.py` you should already have a good idea of what works and what doesn't.\nBut feel free to open a PR anyway, this way we can run the full CI and see the results\n",
"Hi, sorry for no response from my side. I will try to do it today.",
"Created pull request: [LINK](https://github.com/huggingface/datasets/pull/7535)\nTried to run tests by using command you have send and got few errors:\n\n",
"Thanks for running the test ! So it appears we have two issues to fix:\n1. 'log' is not defined: it seems an internal `dill` function has disappeared, so we should adapt the `datasets` code that was using it\n2. there are some hashes mismatches, which means `dill` doesn't seem to output the same dump when passed the same ipython function twice, or the same function but located at a different line in a python file"
] | 2025-04-14T07:22:44Z | 2025-05-19T14:54:04Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Datasets 2.18.0 - 3.5.0 has a dependency on dill < 0.3.9. This causes errors with dill >= 0.3.9.
Could you please take a look into it and make it compatible?
### Steps to reproduce the bug
1. Install setuptools >= 2.18.0
2. Install dill >=0.3.9
3. Run pip check
4. Output:
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
datasets 2.18.0 requires dill<0.3.9,>=0.3.0, but you have dill 0.3.9 which is incompatible.
### Expected behavior
Pip install both libraries without any errors
### Environment info
Datasets version: 2.18 - 3.5
Python: 3.11 | null | {
"+1": 2,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7510/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7510/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7509 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7509/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7509/comments | https://api.github.com/repos/huggingface/datasets/issues/7509/events | https://github.com/huggingface/datasets/issues/7509 | 2,991,484,542 | I_kwDODunzps6yTm5- | 7,509 | Dataset uses excessive memory when loading files | {
"avatar_url": "https://avatars.githubusercontent.com/u/36810152?v=4",
"events_url": "https://api.github.com/users/avishaiElmakies/events{/privacy}",
"followers_url": "https://api.github.com/users/avishaiElmakies/followers",
"following_url": "https://api.github.com/users/avishaiElmakies/following{/other_user}",
"gists_url": "https://api.github.com/users/avishaiElmakies/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/avishaiElmakies",
"id": 36810152,
"login": "avishaiElmakies",
"node_id": "MDQ6VXNlcjM2ODEwMTUy",
"organizations_url": "https://api.github.com/users/avishaiElmakies/orgs",
"received_events_url": "https://api.github.com/users/avishaiElmakies/received_events",
"repos_url": "https://api.github.com/users/avishaiElmakies/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/avishaiElmakies/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/avishaiElmakies/subscriptions",
"type": "User",
"url": "https://api.github.com/users/avishaiElmakies",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"small update: I converted the jsons to parquet and it now works well with 32 proc and the same node. \nI still think this needs to be understood, since json is a very popular and easy-to-use format. ",
"Hi ! The JSON loader loads full files in memory, unless they are JSON Lines. In this case it iterates on the JSON Lines in a memory efficient manner.\n\nI know there is an `ijson` package that works similarly but for general JSON files, maybe it can help and remove the need to load full JSON files in memory",
"Hi, i understand that json files are probably loaded into memory to read them but aren't they released when we write all the file content into arrow or something? ",
"Yes correct, the JSON data is only in memory during the conversion to Arrow. Then, the data is memory mapped from you disk",
"so the json files are all loaded into memory before converting to arrow? or do they convert 1 json at a time and then they are realeased?\nI don't understand how 200GB worth of jsons fill a 378GB node's memory.",
"Each process converts one JSON file at at time, So the total memory usage is num_proc * json_file_size * overhead, where overhead can be around 2 or 3 for the conversion.\n\nSo it's indeed surprising that you run out of memory. Is the dataset available somewhere ? or a subset maybe ?",
"This is a tokenized dataset I created for training a speech-language model with a few features (so it is not private but not easily available). I can send/upload a shard or two and you can copy them however many times you want so you can debug. this should give you something comparable to what I have, but will be easier than creating it yourself. so if you want that, let me know :)",
"Maybe you can measure the memory usage when loading 1 file with num_proc=1 ? This should already be helpful.\n\nMemory usage for tokenized data can be bigger than just text, for example the tokens type can be inferred as int64 and the lists offsets are int32",
"OK, I will try to do this in the near future. I am a little swamped at the moment. do you have a preferred tool?\n\nalso My data is just list of ints, there is no offsets",
"> so the json files are all loaded into memory before converting to arrow? or do they convert 1 json at a time and then they are realeased? I don't understand how 200GB worth of jsons fill a 378GB node's memory.\n\nHello! Is your query solved? I have the same confusion and would like to ask you for advice",
"no, the issue is still present. I converted the json files to parquet, but json seems to have a problem.\n\nUnfortunately i didn't have the time to try and profile the memory usage for 1 file. So if you want to do that, it will be great! ",
"My dataset is about image descriptions, stored as a 20MB JSON file on disk. However, I need to use the map function to preprocess the images, and after computation, the preprocessed dataset amounts to 70GB. My server has 122GB of RAM, but it still runs out of memory (OOM). This issue is very similar to yours.\n\nAfter some research during this period, I found that the map function does not perform disk mapping in memory while working. Using the command find /DataB/mjx -type f -mmin -10, I discovered that no temporary cache files were modified or created during program execution, meaning the data was continuously loaded into memory. After several attempts, I found that adding the parameter cache_file_name=\"your/path\" to the map function can enable memory-disk mapping. This is a strange setting, but after adding this parameter, the memory usage dropped to only 7GB, indicating that once the writer_batch_size worth of data is read into the disk cache, the corresponding data in memory is released. However, I don't think this is the intended behavior by the author, as memory-disk caching should have been enabled without needing this additional parameter.\n\nFinally, here is my map function call. I hope it helps you.\ntrain_data = train_data.map(process_fun, cache_file_name='./cache_file', remove_columns=['image_name', 'question_type', 'concern', 'question', 'candidate_answers', 'answer'])"
] | 2025-04-13T21:09:49Z | 2025-04-28T15:18:55Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Hi
I am having an issue when loading a dataset.
I have about 200 json files each about 1GB (total about 215GB). each row has a few features which are a list of ints.
I am trying to load the dataset using `load_dataset`.
The dataset is about 1.5M samples
I use `num_proc=32` and a node with 378GB of memory.
About a third of the way there I get an OOM.
I also saw an old bug with a similar issue, which says to set `writer_batch_size`. I tried to lower it to 10, but it still crashed.
I also tried to lower the `num_proc` to 16 and even 8, but still the same issue.
### Steps to reproduce the bug
`dataset = load_dataset("json", data_dir=data_config.train_path, num_proc=data_config.num_proc, writer_batch_size=50)["train"]`
### Expected behavior
Loading a dataset with more than 100GB to spare should not cause an OOM error.
maybe i am missing something but I would love some help.
### Environment info
- `datasets` version: 3.5.0
- Platform: Linux-6.6.20-aufs-1-x86_64-with-glibc2.36
- Python version: 3.11.2
- `huggingface_hub` version: 0.29.1
- PyArrow version: 19.0.1
- Pandas version: 2.2.3
- `fsspec` version: 2024.9.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7509/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7509/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7508 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7508/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7508/comments | https://api.github.com/repos/huggingface/datasets/issues/7508/events | https://github.com/huggingface/datasets/issues/7508 | 2,986,612,934 | I_kwDODunzps6yBBjG | 7,508 | Iterating over Image feature columns is extremely slow | {
"avatar_url": "https://avatars.githubusercontent.com/u/11831521?v=4",
"events_url": "https://api.github.com/users/sohamparikh/events{/privacy}",
"followers_url": "https://api.github.com/users/sohamparikh/followers",
"following_url": "https://api.github.com/users/sohamparikh/following{/other_user}",
"gists_url": "https://api.github.com/users/sohamparikh/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sohamparikh",
"id": 11831521,
"login": "sohamparikh",
"node_id": "MDQ6VXNlcjExODMxNTIx",
"organizations_url": "https://api.github.com/users/sohamparikh/orgs",
"received_events_url": "https://api.github.com/users/sohamparikh/received_events",
"repos_url": "https://api.github.com/users/sohamparikh/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sohamparikh/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sohamparikh/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sohamparikh",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! Could it be because the `Image()` type in dataset does `image = Image.open(image_path)` and also `image.load()` which actually loads the image data in memory ? This is needed to avoid too many open files issues, see https://github.com/huggingface/datasets/issues/3985",
"Yes, that seems to be it. For my purposes, I've cast the column to `Image(decode=False)`, and only load the images when necessary, which is much much faster"
] | 2025-04-10T19:00:54Z | 2025-04-15T17:57:08Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | We are trying to load datasets where the image column stores `PIL.PngImagePlugin.PngImageFile` images. However, iterating over these datasets is extremely slow.
What I have found:
1. It is the presence of the image column that causes the slowdown. Removing the column from the dataset results in blazingly fast (as expected) times
2. It is ~2x faster to iterate when the column contains a single image as opposed to a list of images i.e., the feature is a Sequence of Image objects. We often need multiple images per sample, so we need to work with a list of images
3. It is ~17x faster to store paths to PNG files and load them using `PIL.Image.open`, as opposed to iterating over a `Dataset` with an Image column, and ~30x faster compared to `Sequence` of `Image`s. See a simple script below with an openly available dataset.
It would be great to understand the standard practices for storing and loading multimodal datasets (image + text).
https://huggingface.co/docs/datasets/en/image_load seems a bit underdeveloped? (e.g., `dataset.decode` only works with `IterableDataset`, but it's not clear from the doc)
Thanks!
```python
from datasets import load_dataset, load_from_disk
from PIL import Image
from pathlib import Path
ds = load_dataset("getomni-ai/ocr-benchmark")
for idx, sample in enumerate(ds["test"]):
image = sample["image"]
image.save(f"/tmp/ds_files/images/image_{idx}.png")
ds.save_to_disk("/tmp/ds_columns")
# Remove the 'image' column
ds["test"] = ds["test"].remove_columns(["image"])
# Create image paths for each sample
image_paths = [f"images/image_{idx}.png" for idx in range(len(ds["test"]))]
# Add the 'image_path' column to the dataset
ds["test"] = ds["test"].add_column("image_path", image_paths)
# Save the updated dataset
ds.save_to_disk("/tmp/ds_files")
files_path = Path("/tmp/ds_files")
column_path = Path("/tmp/ds_columns")
# load and benchmark
ds_file = load_from_disk(files_path)
ds_column = load_from_disk(column_path)
import time
images_files = []
start = time.time()
for idx in range(len(ds_file["test"])):
image_path = files_path / ds_file["test"][idx]["image_path"]
image = Image.open(image_path)
images_files.append(image)
end = time.time()
print(f"Time taken to load images from files: {end - start} seconds")
# Time taken to load images from files: 1.2364635467529297 seconds
images_column = []
start = time.time()
for idx in range(len(ds_column["test"])):
images_column.append(ds_column["test"][idx]["image"])
end = time.time()
print(f"Time taken to load images from columns: {end - start} seconds")
# Time taken to load images from columns: 20.49347186088562 seconds
``` | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7508/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7508/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7507 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7507/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7507/comments | https://api.github.com/repos/huggingface/datasets/issues/7507/events | https://github.com/huggingface/datasets/issues/7507 | 2,984,309,806 | I_kwDODunzps6x4PQu | 7,507 | Front-end statistical data quantity deviation | {
"avatar_url": "https://avatars.githubusercontent.com/u/88258534?v=4",
"events_url": "https://api.github.com/users/rangehow/events{/privacy}",
"followers_url": "https://api.github.com/users/rangehow/followers",
"following_url": "https://api.github.com/users/rangehow/following{/other_user}",
"gists_url": "https://api.github.com/users/rangehow/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rangehow",
"id": 88258534,
"login": "rangehow",
"node_id": "MDQ6VXNlcjg4MjU4NTM0",
"organizations_url": "https://api.github.com/users/rangehow/orgs",
"received_events_url": "https://api.github.com/users/rangehow/received_events",
"repos_url": "https://api.github.com/users/rangehow/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rangehow/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rangehow/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rangehow",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! the format of this dataset is not supported by the Dataset Viewer. It looks like this dataset was saved using `save_to_disk()` which is meant for local storage / easy reload without compression, not for sharing online."
] | 2025-04-10T02:51:38Z | 2025-04-15T12:54:51Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
While browsing the dataset at https://huggingface.co/datasets/NeuML/wikipedia-20250123, I noticed that a dataset with nearly 7M entries was estimated to be only 4M in sizeβalmost half the actual amount. According to the post-download loading and the dataset_info (https://huggingface.co/datasets/NeuML/wikipedia-20250123/blob/main/train/dataset_info.json), the true data volume is indeed close to 7M. This significant discrepancy could mislead users when sorting datasets by row count. Why not directly retrieve this information from dataset_info?
Not sure if this is the right place to report this bug, but leaving it here for the team's awareness. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7507/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7507/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7506 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7506/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7506/comments | https://api.github.com/repos/huggingface/datasets/issues/7506/events | https://github.com/huggingface/datasets/issues/7506 | 2,981,687,450 | I_kwDODunzps6xuPCa | 7,506 | HfHubHTTPError: 429 Client Error: Too Many Requests for URL when trying to access Fineweb-10BT on 4A100 GPUs using SLURM | {
"avatar_url": "https://avatars.githubusercontent.com/u/66202555?v=4",
"events_url": "https://api.github.com/users/calvintanama/events{/privacy}",
"followers_url": "https://api.github.com/users/calvintanama/followers",
"following_url": "https://api.github.com/users/calvintanama/following{/other_user}",
"gists_url": "https://api.github.com/users/calvintanama/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/calvintanama",
"id": 66202555,
"login": "calvintanama",
"node_id": "MDQ6VXNlcjY2MjAyNTU1",
"organizations_url": "https://api.github.com/users/calvintanama/orgs",
"received_events_url": "https://api.github.com/users/calvintanama/received_events",
"repos_url": "https://api.github.com/users/calvintanama/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/calvintanama/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/calvintanama/subscriptions",
"type": "User",
"url": "https://api.github.com/users/calvintanama",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! make sure to be logged in with your HF account (e.g. using `huggingface-cli login` or passing `token=` to `load_dataset()`), otherwise you'll get rate limited at one point",
"Hey @calvintanama! Just building on what @lhoestq mentioned above β I ran into similar issues in multi-GPU SLURM setups and hereβs what worked for me...\n\nThis 429 Client Error: Too Many Requests comes from the Hugging Face Hubβs rate limiting, which restricts unauthenticated or high-volume access (especially in multi-GPU/distributed setups like SLURM).\n\nAs @lhoestq mentioned, the solution is to make sure you are authenticated with the Hugging Face Hub in every process (especially on each GPU/worker node). You can do this by:\n\nRunning huggingface-cli login (interactive)\n\nOr passing your token explicitly:\n\n```python\nload_dataset(\"HuggingFaceFW/fineweb\", token=\"hf_your_token_here\")\n# If youβre using a SLURM cluster, ensure every node/process receives access to the token via env var:\n```\n\n```bash\nexport HF_TOKEN=hf_your_token_here\n```\n\nand then in Python:\n```python\nfrom datasets import load_dataset\nload_dataset(\"HuggingFaceFW/fineweb\", token=os.environ[\"HF_TOKEN\"])\n```\nAlso consider downloading the dataset beforehand with load_dataset(..., streaming=False) and storing it locally if you're repeatedly training with it."
] | 2025-04-09T06:32:04Z | 2025-06-29T06:04:59Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I am trying to run some finetunings on 4 A100 GPUs using SLURM using axolotl training framework which in turn uses Huggingface's Trainer and Accelerate on [Fineweb-10BT](https://huggingface.co/datasets/HuggingFaceFW/fineweb), but I end up running into 429 Client Error: Too Many Requests for URL error when I call next(dataloader_iter). Funny is, that I can run some test fine tuning (for just 200 training steps) in 1 A100 GPU using SLURM. Is there any rate limiter set for querying dataset? I could run the fine tuning with the same settings (4 A100 GPUs in SLURM) last month.
### Steps to reproduce the bug
You would need a server installed with SLURM
1. Create conda environment
1.1 conda create -n example_env -c conda-forge gxx=11 python=3.10
1.2 conda activate example_env
1.3 pip install torch==2.5.1 torchvision==0.20.1 torchaudio==2.5.1 --index-url https://download.pytorch.org/whl/cu124
1.4 conda install nvidia/label/cuda-12.4.0::cuda-toolkit
1.5 Download flash_attn-2.7.4.post1+cu12torch2.5cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
1.6 pip3 install packaging
1.7 pip3 install ninja
1.8 pip3 install mlflow
1.9 Clone https://github.com/calvintanama/axolotl.git
1.10 `cd` to `axolotl`
1.11 pip3 install -e '.[deepspeed]'
2. Run the training
2.1. Create a folder called `config_run` in axolotl directory
2.2. Copy `config/phi3_pruned_extra_pretrain_22_29_bottleneck_residual_8_a100_4.yaml` to `config_run`
2.3. Change yaml file in the `config_run` accordingly
2.4. Change directory and conda environment name in `jobs/train_phi3_22_29_bottleneck_residual_8_a100_4_temp.sh`
2.5. `jobs/train_phi3_22_29_bottleneck_residual_8_a100_4_temp.sh`
### Expected behavior
This should not cause any error, but gotten
```
File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/accelerate/data_loader.py", line 552, in __iter__
[rank3]: current_batch = next(dataloader_iter)
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 701, in __next__
[rank3]: data = self._next_data()
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 757, in _next_data
[rank3]: data = self._dataset_fetcher.fetch(index) # may raise StopIteration
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/torch/utils/data/_utils/fetch.py", line 33, in fetch
[rank3]: data.append(next(self.dataset_iter))
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/accelerate/data_loader.py", line 338, in __iter__
[rank3]: for element in self.dataset:
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 2266, in __iter__
[rank3]: for key, example in ex_iterable:
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1866, in __iter__
[rank3]: for key, example in self.ex_iterable:
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1084, in __iter__
[rank3]: yield from self._iter()
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1263, in _iter
[rank3]: for key, transformed_example in outputs:
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1258, in <genexpr>
[rank3]: outputs = (
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1244, in iter_outputs
[rank3]: for i, key_example in inputs_iterator:
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1106, in iter_batched_inputs
[rank3]: for key, example in iterator:
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1866, in __iter__
[rank3]: for key, example in self.ex_iterable:
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1535, in __iter__
[rank3]: for x in self.ex_iterable:
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 374, in __iter__
[rank3]: for key, pa_table in self.generate_tables_fn(**gen_kwags):
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/packaged_modules/parquet/parquet.py", line 90, in _generate_tables
[rank3]: if parquet_fragment.row_groups:
[rank3]: File "pyarrow/_dataset_parquet.pyx", line 386, in pyarrow._dataset_parquet.ParquetFileFragment.row_groups.__get__
[rank3]: File "pyarrow/_dataset_parquet.pyx", line 393, in pyarrow._dataset_parquet.ParquetFileFragment.metadata.__get__
[rank3]: File "pyarrow/_dataset_parquet.pyx", line 382, in pyarrow._dataset_parquet.ParquetFileFragment.ensure_complete_metadata
[rank3]: File "pyarrow/error.pxi", line 89, in pyarrow.lib.check_status
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/datasets/utils/file_utils.py", line 827, in read_with_retries
[rank3]: out = read(*args, **kwargs)
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/huggingface_hub/hf_file_system.py", line 1013, in read
[rank3]: return super().read(length)
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/fsspec/spec.py", line 1941, in read
[rank3]: out = self.cache._fetch(self.loc, self.loc + length)
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/fsspec/caching.py", line 234, in _fetch
[rank3]: self.cache = self.fetcher(start, end) # new block replaces old
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/huggingface_hub/hf_file_system.py", line 976, in _fetch_range
[rank3]: hf_raise_for_status(r)
[rank3]: File "/home/hk-project-test-p0023745/cd7437/miniconda3/envs/llmpruning_train_temp/lib/python3.10/site-packages/huggingface_hub/utils/_http.py", line 482, in hf_raise_for_status
[rank3]: raise _format(HfHubHTTPError, str(e), response) from e
[rank3]: huggingface_hub.errors.HfHubHTTPError: 429 Client Error: Too Many Requests for url: https://huggingface.co/datasets/HuggingFaceFW/fineweb/resolve/0f039043b23fe1d4eed300b504aa4b4a68f1c7ba/sample/10BT/006_00000.parquet
```
### Environment info
- datasets 3.5.0
- torch 2.5.1
- transformers 4.46.2 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7506/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7506/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7505 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7505/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7505/comments | https://api.github.com/repos/huggingface/datasets/issues/7505/events | https://github.com/huggingface/datasets/issues/7505 | 2,979,926,156 | I_kwDODunzps6xnhCM | 7,505 | HfHubHTTPError: 403 Forbidden: None. Cannot access content at: https://hf.co/api/s3proxy | {
"avatar_url": "https://avatars.githubusercontent.com/u/1412262?v=4",
"events_url": "https://api.github.com/users/hissain/events{/privacy}",
"followers_url": "https://api.github.com/users/hissain/followers",
"following_url": "https://api.github.com/users/hissain/following{/other_user}",
"gists_url": "https://api.github.com/users/hissain/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/hissain",
"id": 1412262,
"login": "hissain",
"node_id": "MDQ6VXNlcjE0MTIyNjI=",
"organizations_url": "https://api.github.com/users/hissain/orgs",
"received_events_url": "https://api.github.com/users/hissain/received_events",
"repos_url": "https://api.github.com/users/hissain/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/hissain/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hissain/subscriptions",
"type": "User",
"url": "https://api.github.com/users/hissain",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-04-08T14:08:40Z | 2025-04-08T14:08:40Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | I have already logged in Huggingface using CLI with my valid token. Now trying to download the datasets using following code:
from transformers import WhisperProcessor, WhisperForConditionalGeneration, WhisperTokenizer, Trainer, TrainingArguments, DataCollatorForSeq2Seq
from datasets import load_dataset, DatasetDict, Audio
def load_and_preprocess_dataset():
dataset = load_dataset("mozilla-foundation/common_voice_17_0", "bn")
dataset = dataset.remove_columns(["accent", "age", "client_id", "down_votes", "gender", "locale", "segment", "up_votes"])
dataset = dataset.cast_column("audio", Audio(sampling_rate=16000))
dataset = dataset["train"].train_test_split(test_size=0.1)
dataset = DatasetDict({
"train": dataset["train"],
"test": dataset["test"]
})
return dataset
load_and_preprocess_dataset()
I am getting following error:
Downloadingβdata:β100%
β25/25β[00:01<00:00,β25.31files/s]
---------------------------------------------------------------------------
HTTPError Traceback (most recent call last)
File ~/github/bangla-asr/.venv/lib/python3.11/site-packages/huggingface_hub/utils/_http.py:409, in hf_raise_for_status(response, endpoint_name)
408 try:
--> 409 response.raise_for_status()
410 except HTTPError as e:
File ~/github/bangla-asr/.venv/lib/python3.11/site-packages/requests/models.py:1024, in Response.raise_for_status(self)
1023 if http_error_msg:
-> 1024 raise HTTPError(http_error_msg, response=self)
HTTPError: 403 Client Error: BlockSIEL for url: https://hf.co/api/s3proxy?GET=https%3A%2F%2Fhf-hub-lfs-us-east-1.s3.us-east-1.amazonaws.com%2Frepos%2Fa3%2F86%2Fa386bf65687d8a6928c1ea57c383aa3faade32f5171150e25af3fc1cfc273db8%2F67f1ac9cabd539bfbff3acbc549b60647833a250dc638866f22bf1b64e68806d%3FX-Amz-Algorithm%3DAWS4-HMAC-SHA256%26X-Amz-Content-Sha256%3DUNSIGNED-PAYLOAD%26X-Amz-Credential%3DAKIA2JU7TKAQLC2QXPN7%252F20250408%252Fus-east-1%252Fs3%252Faws4_request%26X-Amz-Date%3D20250408T134345Z%26X-Amz-Expires%3D3600%26X-Amz-Signature%3D621e731d4fd6d08afbf568379797746ab8e2b853b6728ff5e1122fef6e56880b%26X-Amz-SignedHeaders%3Dhost%26response-content-disposition%3Dinline%253B%2520filename%252A%253DUTF-8%2527%2527bn_validated_1.tar%253B%2520filename%253D%2522bn_validated_1.tar%2522%253B%26response-content-type%3Dapplication%252Fx-tar%26x-id%3DGetObject&HEAD=https%3A%2F%2Fhf-hub-lfs-us-east-1.s3.us-east-1.amazonaws.com%2Frepos%2Fa3%2F86%2Fa386bf65687d8a6928c1ea57c383aa3faade32f5171150e25af3fc1cfc273db8%2F67f1ac9cabd539bfbff3acbc549b60647833a250dc638866f22bf1b64e68806d%3FX-Amz-Algorithm%3DAWS4-HMAC-SHA256%26X-Amz-Content-Sha256%3DUNSIGNED-PAYLOAD%26X-Amz-Credential%3DAKIA2JU7TKAQLC2QXPN7%252F20250408%252Fus-east-1%252Fs3%252Faws4_request%26X-Amz-Date%3D20250408T134345Z%26X-Amz-Expires%3D3600%26X-Amz-Signature%3D15254fb79d30b0dc36b94a28138e675e0e00bb475b8a3ae774418500b095a661%26X-Amz-SignedHeaders%3Dhost&sign=eyJhbGciOiJIUzI1NiJ9.eyJyZWRpcmVjdF9kb21haW4iOiJoZi1odWItbGZzLXVzLWVhc3QtMS5zMy51cy1lYXN0LTEuYW1hem9uYXdzLmNvbSIsImlhdCI6MTc0NDExOTgyNSwiZXhwIjoxNzQ0MjA2MjI1LCJpc3MiOiJodHRwczovL2h1Z2dpbmdmYWNlLmNvIn0.5sJzudFDU3SmOdOLlwmQCOfQFf2r7y9590HoX8WBkRk
The above exception was the direct cause of the following exception:
HfHubHTTPError Traceback (most recent call last)
Cell In[16], line 15
9 dataset = DatasetDict({
10 "train": dataset["train"],
11 "test": dataset["test"]
12 })
13 return dataset
---> 15 load_and_preprocess_dataset()
17 # def setup_model():
18 # processor = WhisperProcessor.from_pretrained("openai/whisper-base")
...
475 range_header = response.request.headers.get("Range")
HfHubHTTPError: 403 Forbidden: None.
Cannot access content at: https://hf.co/api/s3proxy?GET=https%3A%2F%2Fhf-hub-lfs-us-east-1.s3.us-east-1.amazonaws.com%2Frepos%2Fa3%2F86%2Fa386bf6568724a6928c1ea57c383aa3faade32f5171150e25af3fc1cfc273db8%2F67f1ac9cabd539bfbff3acbc549b60647833a250dc638786f22bf1b64e68806d%3FX-Amz-Algorithm%3DAWS4-HMAC-SHA256%26X-Amz-Content-Sha256%3DUNSIGNED-PAYLOAD%26X-Amz-Credential%3DAKIA2JU7TKAQLC2QXPN7%252F20250408%252Fus-east-1%252Fs3%252Faws4_request%26X-Amz-Date%3D20250408T134345Z%26X-Amz-Expires%3D3600%26X-Amz-Signature%3D621e731d4fd6d08afbf568379797746ab394b853b6728ff5e1122fef6e56880b%26X-Amz-SignedHeaders%3Dhost%26response-content-disposition%3Dinline%253B%2520filename%252A%253DUTF-8%2527%2527bn_validated_1.tar%253B%2520filename%253D%2522bn_validated_1.tar%2522%253B%26response-content-type%3Dapplication%252Fx-tar%26x-id%3DGetObject&HEAD=https%3A%2F%2Fhf-hub-lfs-us-east-1.s3.us-east-1.amazonaws.com%2Frepos%2Fa3%2F86%2Fa386bf65687ab76928c1ea57c383aa3faade32f5171150e25af3fc1cfc273db8%2F67f1ac9cabd539bfbff3acbc549b60647833a250d2338866f222f1b64e68806d%3FX-Amz-Algorithm%3DAWS4-HMAC-SHA256%26X-Amz-Content-Sha256%3DUNSIGNED-PAYLOAD%26X-Amz-Credential%3DAKIA2JU7TKAQLC2QXPN7%252F20250408%252Fus-east-1%252Fs3%252Faws4_request%26X-Amz-Date%3D20250408T134345Z%26X-Amz-Expires%3D3600%26X-Amz-Signature%3D15254fb79d30b0dc36b94a28138e675e0e00bb475b8a3ae774418500b095a661%26X-Amz-SignedHeaders%3Dhost&sign=eyJhbGciOiJIUzI1NiJ9.eyJyZWRpcmVjds9kb21haW4iOiJoZi1odWItbGZzLXVzLWVhc3QtMS5zMy51cy1lYXN0LTEuYW1hem9uYXdzLmNvbSIsImlhdCI6MTc0NDExOT2yNSwiZXhwIjoxNzQ0MjA2MjI1LCJpc3MiOiJodHRwczovL2h1Z2dpbmdmYWNlLmNvIn0.5sJzudFDU3SmOdOLlwmQdOfQFf2r7y9590HoX8WBkRk.
Make sure your token has the correct permissions.
**What's wrong with the code?** Please note that the error is happening only when I am running from my office network due to probably proxy. Which URL, I need to take a proxy exception? | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7505/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7505/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7504 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7504/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7504/comments | https://api.github.com/repos/huggingface/datasets/issues/7504/events | https://github.com/huggingface/datasets/issues/7504 | 2,979,410,641 | I_kwDODunzps6xljLR | 7,504 | BuilderConfig ParquetConfig(...) doesn't have a 'use_auth_token' key. | {
"avatar_url": "https://avatars.githubusercontent.com/u/20015750?v=4",
"events_url": "https://api.github.com/users/tteguayco/events{/privacy}",
"followers_url": "https://api.github.com/users/tteguayco/followers",
"following_url": "https://api.github.com/users/tteguayco/following{/other_user}",
"gists_url": "https://api.github.com/users/tteguayco/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/tteguayco",
"id": 20015750,
"login": "tteguayco",
"node_id": "MDQ6VXNlcjIwMDE1NzUw",
"organizations_url": "https://api.github.com/users/tteguayco/orgs",
"received_events_url": "https://api.github.com/users/tteguayco/received_events",
"repos_url": "https://api.github.com/users/tteguayco/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/tteguayco/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/tteguayco/subscriptions",
"type": "User",
"url": "https://api.github.com/users/tteguayco",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"I encountered the same error, have you resolved it?",
"Hi ! `use_auth_token` has been deprecated and removed some time ago. You should use `token` instead in `load_dataset()`",
"Hi @lhoestq, I'd like to take this up.\n\nAs discussed in #7504, the issue arises when `use_auth_token` is passed to `load_dataset`, which forwards it to the config's `__init__`, where it's no longer a valid key.\n\nTo address this, Iβll intercept and strip `use_auth_token` inside `load_dataset()` (similar to how we handle `trust_remote_code`). A warning will be logged, and users will be encouraged to use `token` instead.\n\nThis avoids breaking older scripts that still use `use_auth_token`."
] | 2025-04-08T10:55:03Z | 2025-06-28T09:18:09Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Trying to run the following fine-tuning script (based on this page [here](https://github.com/huggingface/instruction-tuned-sd)):
```
! accelerate launch /content/instruction-tuned-sd/finetune_instruct_pix2pix.py \
--pretrained_model_name_or_path=${MODEL_ID} \
--dataset_name=${DATASET_NAME} \
--use_ema \
--enable_xformers_memory_efficient_attention \
--resolution=512 --random_flip \
--train_batch_size=2 --gradient_accumulation_steps=4 --gradient_checkpointing \
--max_train_steps=500 \
--checkpointing_steps=25 --checkpoints_total_limit=1 \
--learning_rate=5e-05 --max_grad_norm=1 --lr_warmup_steps=20 \
--conditioning_dropout_prob=0.1 \
--mixed_precision=fp16 \
--seed=42 \
--output_dir=${OUTPUT_DIR} \
--original_image_column=before \
--edit_prompt=prompt \
--edited_image=after
```
but I keep getting the following error:
```
Traceback (most recent call last):
File "/content/instruction-tuned-sd/finetune_instruct_pix2pix.py", line 1137, in <module>
main()
File "/content/instruction-tuned-sd/finetune_instruct_pix2pix.py", line 652, in main
dataset = load_dataset(
^^^^^^^^^^^^^
File "/usr/local/lib/python3.11/dist-packages/datasets/load.py", line 2129, in load_dataset
builder_instance = load_dataset_builder(
^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.11/dist-packages/datasets/load.py", line 1886, in load_dataset_builder
builder_instance: DatasetBuilder = builder_cls(
^^^^^^^^^^^^
File "/usr/local/lib/python3.11/dist-packages/datasets/builder.py", line 342, in __init__
self.config, self.config_id = self._create_builder_config(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.11/dist-packages/datasets/builder.py", line 590, in _create_builder_config
raise ValueError(f"BuilderConfig {builder_config} doesn't have a '{key}' key.")
ValueError: BuilderConfig ParquetConfig(name='default', version=0.0.0, data_dir=None, data_files={'train': ['data/train-*']}, description=None, batch_size=None, columns=None, features=None, filters=None) doesn't have a 'use_auth_token' key.
Traceback (most recent call last):
File "/usr/local/bin/accelerate", line 10, in <module>
sys.exit(main())
^^^^^^
```
Any ideas? `datasets` version should be `3.2.0`.
### Steps to reproduce the bug
Just running the script above.
### Expected behavior
No errors
### Environment info
Python 3.11.11
datasets==3.2.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7504/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7504/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7503 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7503/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7503/comments | https://api.github.com/repos/huggingface/datasets/issues/7503/events | https://github.com/huggingface/datasets/issues/7503 | 2,978,512,625 | I_kwDODunzps6xiH7x | 7,503 | Inconsistency between load_dataset and load_from_disk functionality | {
"avatar_url": "https://avatars.githubusercontent.com/u/60975422?v=4",
"events_url": "https://api.github.com/users/zzzzzec/events{/privacy}",
"followers_url": "https://api.github.com/users/zzzzzec/followers",
"following_url": "https://api.github.com/users/zzzzzec/following{/other_user}",
"gists_url": "https://api.github.com/users/zzzzzec/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zzzzzec",
"id": 60975422,
"login": "zzzzzec",
"node_id": "MDQ6VXNlcjYwOTc1NDIy",
"organizations_url": "https://api.github.com/users/zzzzzec/orgs",
"received_events_url": "https://api.github.com/users/zzzzzec/received_events",
"repos_url": "https://api.github.com/users/zzzzzec/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zzzzzec/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zzzzzec/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zzzzzec",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! you can find more info here: https://github.com/huggingface/datasets/issues/5044#issuecomment-1263714347\n\n> What's the recommended approach for this use case? Should I manually process my gsm8k-new dataset to make it compatible with load_dataset? Is there a standard way to convert between these formats?\n\nYou can use push_to_hub() or to_parquet() for example",
"Hi @zzzzzec & @lhoestq π\n\nThanks for raising and discussing this β I've submitted a patch that improves this exact scenario."
] | 2025-04-08T03:46:22Z | 2025-06-28T08:51:16Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ## Issue Description
I've encountered confusion when using `load_dataset` and `load_from_disk` in the datasets library. Specifically, when working offline with the gsm8k dataset, I can load it using a local path:
```python
import datasets
ds = datasets.load_dataset('/root/xxx/datasets/gsm8k', 'main')
```
output:
```text
DatasetDict({
train: Dataset({
features: ['question', 'answer'],
num_rows: 7473
})
test: Dataset({
features: ['question', 'answer'],
num_rows: 1319
})
})
```
This works as expected. However, after processing the dataset (converting answer format from #### to \boxed{})
```python
import datasets
ds = datasets.load_dataset('/root/xxx/datasets/gsm8k', 'main')
ds_train = ds['train']
ds_test = ds['test']
import re
def convert(sample):
solution = sample['answer']
solution = re.sub(r'####\s*(\S+)', r'\\boxed{\1}', solution)
sample = {
'problem': sample['question'],
'solution': solution
}
return sample
ds_train = ds_train.map(convert, remove_columns=['question', 'answer'])
ds_test = ds_test.map(convert,remove_columns=['question', 'answer'])
```
I saved it using save_to_disk:
```python
from datasets.dataset_dict import DatasetDict
data_dict = DatasetDict({
'train': ds_train,
'test': ds_test
})
data_dict.save_to_disk('/root/xxx/datasets/gsm8k-new')
```
But now I can only load it using load_from_disk:
```python
new_ds = load_from_disk('/root/xxx/datasets/gsm8k-new')
```
output:
```text
DatasetDict({
train: Dataset({
features: ['problem', 'solution'],
num_rows: 7473
})
test: Dataset({
features: ['problem', 'solution'],
num_rows: 1319
})
})
```
Attempting to use load_dataset produces unexpected results:
```python
new_ds = load_dataset('/root/xxx/datasets/gsm8k-new')
```
output:
```text
DatasetDict({
train: Dataset({
features: ['_data_files', '_fingerprint', '_format_columns', '_format_kwargs', '_format_type', '_output_all_columns', '_split'],
num_rows: 1
})
test: Dataset({
features: ['_data_files', '_fingerprint', '_format_columns', '_format_kwargs', '_format_type', '_output_all_columns', '_split'],
num_rows: 1
})
})
```
Questions
1. Why is it designed such that after using `save_to_disk`, the dataset cannot be loaded with `load_dataset`? For small projects with limited code, it might be relatively easy to change all instances of `load_dataset` to `load_from_disk`. However, for complex frameworks like TRL or lighteval, diving into the framework code to change `load_dataset` to `load_from_disk` is extremely tedious and error-prone.
Additionally, `load_from_disk` cannot load datasets directly downloaded from the hub, which means that if you need to modify a dataset, you have to choose between using `load_from_disk` or `load_dataset`. This creates an unnecessary dichotomy in the API and complicates workflow when working with modified datasets.
2. What's the recommended approach for this use case? Should I manually process my gsm8k-new dataset to make it compatible with load_dataset? Is there a standard way to convert between these formats?
thanks~ | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7503/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7503/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7502 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7502/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7502/comments | https://api.github.com/repos/huggingface/datasets/issues/7502/events | https://github.com/huggingface/datasets/issues/7502 | 2,977,453,814 | I_kwDODunzps6xeFb2 | 7,502 | `load_dataset` of size 40GB creates a cache of >720GB | {
"avatar_url": "https://avatars.githubusercontent.com/u/61748653?v=4",
"events_url": "https://api.github.com/users/pietrolesci/events{/privacy}",
"followers_url": "https://api.github.com/users/pietrolesci/followers",
"following_url": "https://api.github.com/users/pietrolesci/following{/other_user}",
"gists_url": "https://api.github.com/users/pietrolesci/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/pietrolesci",
"id": 61748653,
"login": "pietrolesci",
"node_id": "MDQ6VXNlcjYxNzQ4NjUz",
"organizations_url": "https://api.github.com/users/pietrolesci/orgs",
"received_events_url": "https://api.github.com/users/pietrolesci/received_events",
"repos_url": "https://api.github.com/users/pietrolesci/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/pietrolesci/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pietrolesci/subscriptions",
"type": "User",
"url": "https://api.github.com/users/pietrolesci",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! Parquet is a compressed format. When you load a dataset, it uncompresses the Parquet data into Arrow data on your disk. That's why you can indeed end up with 720GB of uncompressed data on disk. The uncompression is needed to enable performant dataset objects (especially for random access).\n\nTo save some storage you can instead load the dataset with `streaming=True`. This way you get an `IterableDataset` that reads the Parquet data iteratively without ever writing to disk.\n\nPS: `ReadInstruction` might not be implemented for `streaming=True`, if it's the case you can use `ds.take()` and `ds.skip()` instead",
"Hi @lhoestq, thanks a lot for your answer. This makes perfect sense. I will try using the streaming mode. Closing the issue."
] | 2025-04-07T16:52:34Z | 2025-04-15T15:22:12Z | 2025-04-15T15:22:11Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | Hi there,
I am trying to load a dataset from the Hugging Face Hub and split it into train and validation splits. Somehow, when I try to do it with `load_dataset`, it exhausts my disk quota. So, I tried manually downloading the parquet files from the hub and loading them as follows:
```python
ds = DatasetDict(
{
"train": load_dataset(
"parquet",
data_dir=f"{local_dir}/{tok}",
cache_dir=cache_dir,
num_proc=min(12, os.cpu_count()), # type: ignore
split=ReadInstruction("train", from_=0, to=NUM_TRAIN, unit="abs"), # type: ignore
),
"validation": load_dataset(
"parquet",
data_dir=f"{local_dir}/{tok}",
cache_dir=cache_dir,
num_proc=min(12, os.cpu_count()), # type: ignore
split=ReadInstruction("train", from_=NUM_TRAIN, unit="abs"), # type: ignore
)
}
)
```
which still strangely creates 720GB of cache. In addition, if I remove the raw parquet file folder (`f"{local_dir}/{tok}"` in this example), I am not able to load anything. So, I am left wondering what this cache is doing. Am I missing something? Is there a solution to this problem?
Thanks a lot in advance for your help!
A related issue: https://github.com/huggingface/transformers/issues/10204#issue-809007443.
---
Python: 3.11.11
datasets: 3.5.0
| {
"avatar_url": "https://avatars.githubusercontent.com/u/61748653?v=4",
"events_url": "https://api.github.com/users/pietrolesci/events{/privacy}",
"followers_url": "https://api.github.com/users/pietrolesci/followers",
"following_url": "https://api.github.com/users/pietrolesci/following{/other_user}",
"gists_url": "https://api.github.com/users/pietrolesci/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/pietrolesci",
"id": 61748653,
"login": "pietrolesci",
"node_id": "MDQ6VXNlcjYxNzQ4NjUz",
"organizations_url": "https://api.github.com/users/pietrolesci/orgs",
"received_events_url": "https://api.github.com/users/pietrolesci/received_events",
"repos_url": "https://api.github.com/users/pietrolesci/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/pietrolesci/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pietrolesci/subscriptions",
"type": "User",
"url": "https://api.github.com/users/pietrolesci",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7502/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7502/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7501 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7501/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7501/comments | https://api.github.com/repos/huggingface/datasets/issues/7501/events | https://github.com/huggingface/datasets/issues/7501 | 2,976,721,014 | I_kwDODunzps6xbSh2 | 7,501 | Nested Feature raises ArrowNotImplementedError: Unsupported cast using function cast_struct | {
"avatar_url": "https://avatars.githubusercontent.com/u/26623948?v=4",
"events_url": "https://api.github.com/users/yaner-here/events{/privacy}",
"followers_url": "https://api.github.com/users/yaner-here/followers",
"following_url": "https://api.github.com/users/yaner-here/following{/other_user}",
"gists_url": "https://api.github.com/users/yaner-here/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yaner-here",
"id": 26623948,
"login": "yaner-here",
"node_id": "MDQ6VXNlcjI2NjIzOTQ4",
"organizations_url": "https://api.github.com/users/yaner-here/orgs",
"received_events_url": "https://api.github.com/users/yaner-here/received_events",
"repos_url": "https://api.github.com/users/yaner-here/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yaner-here/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yaner-here/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yaner-here",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Solved by the default `load_dataset(features)` parameters. Do not use `Sequence` for the `list` in `list[any]` json schema, just simply use `[]`. For example, `\"b\": Sequence({...})` fails but `\"b\": [{...}]` works fine."
] | 2025-04-07T12:35:39Z | 2025-04-07T12:43:04Z | 2025-04-07T12:43:03Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
`datasets.Features` seems to be unable to handle json file that contains fields of `list[dict]`.
### Steps to reproduce the bug
```json
// test.json
{"a": 1, "b": [{"c": 2, "d": 3}, {"c": 4, "d": 5}]}
{"a": 5, "b": [{"c": 7, "d": 8}, {"c": 9, "d": 10}]}
```
```python
import json
from datasets import Dataset, Features, Value, Sequence, load_dataset
annotation_feature = Features({
"a": Value("int32"),
"b": Sequence({
"c": Value("int32"),
"d": Value("int32"),
}),
})
annotation_dataset = load_dataset(
"json",
data_files="test.json",
features=annotation_feature
)
```
```
ArrowNotImplementedError: Unsupported cast from list<item: struct<c: int32, d: int32>> to struct using function cast_struct
The above exception was the direct cause of the following exception:
DatasetGenerationError Traceback (most recent call last)
Cell In[46], line 11
2 from datasets import Dataset, Features, Value, Sequence, load_dataset
4 annotation_feature = Features({
5 "a": Value("int32"),
6 "b": Sequence({
(...) 9 }),
10 })
---> 11 annotation_dataset = load_dataset(
12 "json",
13 data_files="test.json",
14 features=annotation_feature
15 )
```
### Expected behavior
A `datasets.Datasets` instance should be initialized.
### Environment info
- `datasets` version: 3.5.0
- Platform: Linux-6.11.0-21-generic-x86_64-with-glibc2.39
- Python version: 3.11.11
- `huggingface_hub` version: 0.30.1
- PyArrow version: 19.0.1
- Pandas version: 2.2.3
- `fsspec` version: 2024.12.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/26623948?v=4",
"events_url": "https://api.github.com/users/yaner-here/events{/privacy}",
"followers_url": "https://api.github.com/users/yaner-here/followers",
"following_url": "https://api.github.com/users/yaner-here/following{/other_user}",
"gists_url": "https://api.github.com/users/yaner-here/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yaner-here",
"id": 26623948,
"login": "yaner-here",
"node_id": "MDQ6VXNlcjI2NjIzOTQ4",
"organizations_url": "https://api.github.com/users/yaner-here/orgs",
"received_events_url": "https://api.github.com/users/yaner-here/received_events",
"repos_url": "https://api.github.com/users/yaner-here/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yaner-here/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yaner-here/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yaner-here",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7501/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7501/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7500 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7500/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7500/comments | https://api.github.com/repos/huggingface/datasets/issues/7500/events | https://github.com/huggingface/datasets/issues/7500 | 2,974,841,921 | I_kwDODunzps6xUHxB | 7,500 | Make `with_format` correctly indicate that a `Dataset` is compatible with PyTorch's `Dataset` class | {
"avatar_url": "https://avatars.githubusercontent.com/u/3817460?v=4",
"events_url": "https://api.github.com/users/benglewis/events{/privacy}",
"followers_url": "https://api.github.com/users/benglewis/followers",
"following_url": "https://api.github.com/users/benglewis/following{/other_user}",
"gists_url": "https://api.github.com/users/benglewis/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/benglewis",
"id": 3817460,
"login": "benglewis",
"node_id": "MDQ6VXNlcjM4MTc0NjA=",
"organizations_url": "https://api.github.com/users/benglewis/orgs",
"received_events_url": "https://api.github.com/users/benglewis/received_events",
"repos_url": "https://api.github.com/users/benglewis/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/benglewis/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/benglewis/subscriptions",
"type": "User",
"url": "https://api.github.com/users/benglewis",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | [] | null | [
"Does the torch `DataLoader` really require the dataset to be a subclass of `torch.utils.data.Dataset` ? Or is there a simpler type we could use ?\n\nPS: also note that a dataset without `with_format()` can also be used in a torch `DataLoader` . Calling `with_format(\"torch\")` simply makes the output of the dataset torch Tensors in an efficient way."
] | 2025-04-06T09:56:09Z | 2025-04-15T12:57:39Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
Currently `datasets` does not correctly indicate to the Python type-checker (e.g. `pyright` / `Pylance`) that the output of `with_format` is compatible with PyTorch's `Dataloader` since it does not indicate that the HuggingFace `Dataset` is compatible with the PyTorch `Dataset` class. It would be great if we could get the typing to work nicely.
### Motivation
To avoid casting types in our Python code.
### Your contribution
I would be happy to contribute a PR if this is something that may be accepted and could work with the current approach.
This doesn't have to be for just PyTorch, but I imagine that the same thing would be useful for `tensorflow` and such, but we only have a need for PyTorch at this stage. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7500/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7500/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7499 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7499/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7499/comments | https://api.github.com/repos/huggingface/datasets/issues/7499/events | https://github.com/huggingface/datasets/pull/7499 | 2,973,489,126 | PR_kwDODunzps6Rd4Zp | 7,499 | Added cache dirs to load and file_utils | {
"avatar_url": "https://avatars.githubusercontent.com/u/43501738?v=4",
"events_url": "https://api.github.com/users/gmongaras/events{/privacy}",
"followers_url": "https://api.github.com/users/gmongaras/followers",
"following_url": "https://api.github.com/users/gmongaras/following{/other_user}",
"gists_url": "https://api.github.com/users/gmongaras/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/gmongaras",
"id": 43501738,
"login": "gmongaras",
"node_id": "MDQ6VXNlcjQzNTAxNzM4",
"organizations_url": "https://api.github.com/users/gmongaras/orgs",
"received_events_url": "https://api.github.com/users/gmongaras/received_events",
"repos_url": "https://api.github.com/users/gmongaras/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/gmongaras/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gmongaras/subscriptions",
"type": "User",
"url": "https://api.github.com/users/gmongaras",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"hi ! the `hf_hub_download` cache_dir is a different cache directory than the one for `datasets`.\r\n\r\n`hf_hub_download` uses the `huggingface_hub` cache which is located in by default in `~/.cache/huggingface/hub`, while `datasets` uses a different cache for Arrow files and map() results `~/.cache/huggingface/datasets`",
"Is there a way to change the default cache directory for both of these on calling load_dataset? Currently, cache_dir makes dealing with where I want files to go a bit confusing as the documentation doesn't mention it only relocates.../datasets and not .../hub.",
"You can set `HF_HOME` which is the common parent directory for those two caches. Or individually `HF_DATASETS_CACHE` and `HF_HUB_CACHE`",
"Got it. Can this be added to the documentation for load_dataset and related functions to avoid confusion with cache_dir?",
"done in https://github.com/huggingface/datasets/pull/7532 :)"
] | 2025-04-04T22:36:04Z | 2025-05-07T14:07:34Z | 2025-05-07T14:07:34Z | NONE | null | null | null | When adding "cache_dir" to datasets.load_dataset, the cache_dir gets lost in the function calls, changing the cache dir to the default path. This fixes a few of these instances. | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7499/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7499/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7499.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7499",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7499.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7499"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7498 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7498/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7498/comments | https://api.github.com/repos/huggingface/datasets/issues/7498/events | https://github.com/huggingface/datasets/issues/7498 | 2,969,218,273 | I_kwDODunzps6w-qzh | 7,498 | Extreme memory bandwidth. | {
"avatar_url": "https://avatars.githubusercontent.com/u/185079645?v=4",
"events_url": "https://api.github.com/users/J0SZ/events{/privacy}",
"followers_url": "https://api.github.com/users/J0SZ/followers",
"following_url": "https://api.github.com/users/J0SZ/following{/other_user}",
"gists_url": "https://api.github.com/users/J0SZ/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/J0SZ",
"id": 185079645,
"login": "J0SZ",
"node_id": "U_kgDOCwgXXQ",
"organizations_url": "https://api.github.com/users/J0SZ/orgs",
"received_events_url": "https://api.github.com/users/J0SZ/received_events",
"repos_url": "https://api.github.com/users/J0SZ/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/J0SZ/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/J0SZ/subscriptions",
"type": "User",
"url": "https://api.github.com/users/J0SZ",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-04-03T11:09:08Z | 2025-04-03T11:11:22Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When I use hf datasets on 4 GPU with 40 workers I get some extreme memory bandwidth of constant ~3GB/s.
However, if I wrap the dataset in `IterableDataset`, this issue is gone and the data also loads way faster (4x faster training on 1 worker).
It seems like the workers don't share memory and basically duplicate the data 4x40.
### Steps to reproduce the bug
Trainer arguments:
```
dataloader_pin_memory=True,
dataloader_num_workers=40,
dataloader_prefetch_factor=2,
dataloader_persistent_workers=True,
```
Call trainer:
```
trainer = Trainer(
model=model,
args=train_args,
train_dataset=load_from_disk('..').with_fromat('torch'),
)
```
The dataset has 600GB and consists of 1225 files.
### Expected behavior
The optimal bandwidth should be 100MB/s to keep up with GPU.
### Environment info
Linux
Python 3.11
datasets==3.2.0
| null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7498/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7498/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7497 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7497/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7497/comments | https://api.github.com/repos/huggingface/datasets/issues/7497/events | https://github.com/huggingface/datasets/issues/7497 | 2,968,553,693 | I_kwDODunzps6w8Ijd | 7,497 | How to convert videos to images? | {
"avatar_url": "https://avatars.githubusercontent.com/u/171649931?v=4",
"events_url": "https://api.github.com/users/Loki-Lu/events{/privacy}",
"followers_url": "https://api.github.com/users/Loki-Lu/followers",
"following_url": "https://api.github.com/users/Loki-Lu/following{/other_user}",
"gists_url": "https://api.github.com/users/Loki-Lu/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Loki-Lu",
"id": 171649931,
"login": "Loki-Lu",
"node_id": "U_kgDOCjsriw",
"organizations_url": "https://api.github.com/users/Loki-Lu/orgs",
"received_events_url": "https://api.github.com/users/Loki-Lu/received_events",
"repos_url": "https://api.github.com/users/Loki-Lu/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Loki-Lu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Loki-Lu/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Loki-Lu",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | [] | null | [
"Hi ! there is some documentation here on how to read video frames: https://huggingface.co/docs/datasets/video_load"
] | 2025-04-03T07:08:39Z | 2025-04-15T12:35:15Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
Does someone know how to return the images from videos?
### Motivation
I am trying to use openpi(https://github.com/Physical-Intelligence/openpi) to finetune my Lerobot dataset(V2.0 and V2.1). I find that although the codedaset is v2.0, they are different. It seems like Lerobot V2.0 has two version, one is data include images infos and another one is separate to data and videos.
Does someone know how to return the images from videos?
| null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7497/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7497/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7496 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7496/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7496/comments | https://api.github.com/repos/huggingface/datasets/issues/7496/events | https://github.com/huggingface/datasets/issues/7496 | 2,967,345,522 | I_kwDODunzps6w3hly | 7,496 | Json builder: Allow features to override problematic Arrow types | {
"avatar_url": "https://avatars.githubusercontent.com/u/1017189?v=4",
"events_url": "https://api.github.com/users/edmcman/events{/privacy}",
"followers_url": "https://api.github.com/users/edmcman/followers",
"following_url": "https://api.github.com/users/edmcman/following{/other_user}",
"gists_url": "https://api.github.com/users/edmcman/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/edmcman",
"id": 1017189,
"login": "edmcman",
"node_id": "MDQ6VXNlcjEwMTcxODk=",
"organizations_url": "https://api.github.com/users/edmcman/orgs",
"received_events_url": "https://api.github.com/users/edmcman/received_events",
"repos_url": "https://api.github.com/users/edmcman/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/edmcman/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/edmcman/subscriptions",
"type": "User",
"url": "https://api.github.com/users/edmcman",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | [] | null | [
"Hi ! It would be cool indeed, currently the JSON data are generally loaded here: \n\nhttps://github.com/huggingface/datasets/blob/90e5bf8a8599b625d6103ee5ac83b98269991141/src/datasets/packaged_modules/json/json.py#L137-L140\n\nMaybe we can pass a Arrow `schema` to avoid errors ?"
] | 2025-04-02T19:27:16Z | 2025-04-15T13:06:09Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
In the JSON builder, use explicitly requested feature types before or while converting to Arrow.
### Motivation
Working with JSON datasets is really hard because of Arrow. At the very least, it seems like it should be possible to work-around these problems by explicitly setting problematic columns's types. But it seems like this is not possible because the features are only used *after* converting to arrow.
Here's a simple example where the Arrow error could potentially be avoided by converting the column to a string: https://colab.research.google.com/drive/16QHRdbUwKSrpwVfGwu8V8AHr8v2dv0dt?usp=sharing
### Your contribution
Maybe with some guidance. I'm not very familiar with arrow or pandas. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7496/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7496/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7495 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7495/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7495/comments | https://api.github.com/repos/huggingface/datasets/issues/7495/events | https://github.com/huggingface/datasets/issues/7495 | 2,967,034,060 | I_kwDODunzps6w2VjM | 7,495 | Columns in the dataset obtained though load_dataset do not correspond to the one in the dataset viewer since 3.4.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4",
"events_url": "https://api.github.com/users/bruno-hays/events{/privacy}",
"followers_url": "https://api.github.com/users/bruno-hays/followers",
"following_url": "https://api.github.com/users/bruno-hays/following{/other_user}",
"gists_url": "https://api.github.com/users/bruno-hays/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/bruno-hays",
"id": 48770768,
"login": "bruno-hays",
"node_id": "MDQ6VXNlcjQ4NzcwNzY4",
"organizations_url": "https://api.github.com/users/bruno-hays/orgs",
"received_events_url": "https://api.github.com/users/bruno-hays/received_events",
"repos_url": "https://api.github.com/users/bruno-hays/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/bruno-hays/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bruno-hays/subscriptions",
"type": "User",
"url": "https://api.github.com/users/bruno-hays",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi, the dataset viewer shows all the possible columns and their types, but `load_dataset()` iterates through all the columns that you defined. It seems that you only have one column (βaudioβ) defined in your dataset because when I ran `print(ds.column_names)`, the only name I got was βaudioβ. You need to clearly define all the other features of the dataset as columns to enable your original code to work. Furthermore, you can run this code to print out all the features of your dataset: \n```python\nfrom datasets import load_dataset_builder\nds_builder = load_dataset_builder(\"BrunoHays/Accueil_UBS\")\nprint(ds_builder.info.features)\n```\n",
"@phoebecd \nGood catch, even in datasets<3.4.0, the only feature is \"audio\".\nThis datasets follows the [audio folder](https://huggingface.co/docs/datasets/en/audio_dataset#audiofolder) structure with metadata.csv.\nMaybe I missed something or there is a bug when having and audio_folder with a metadata file\n\nWhat do you think @lhoestq ?",
"I opened a PR to fix the issue :) https://huggingface.co/datasets/BrunoHays/Accueil_UBS/discussions/2\n\nWe expect the metadata file to be in the <split>/ folder now to allow one CSV metadata file per split. But in the PR I just added a manual configuration instead of moving the file and updating all the relative paths it contains."
] | 2025-04-02T17:01:11Z | 2025-07-02T23:24:57Z | 2025-07-02T23:24:57Z | CONTRIBUTOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I have noticed that on my dataset named [BrunoHays/Accueil_UBS](https://huggingface.co/datasets/BrunoHays/Accueil_UBS), since the version 3.4.0, every column except audio is missing when I load the dataset.
Interestingly, the dataset viewer still shows the correct columns
### Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_dataset("BrunoHays/Accueil_UBS", streaming=True)
print(next(iter(ds["test"])).keys())
```
With datasets >= 3.4.0:
-> dict_keys(['audio'])
With datasets == 3.3.2:
-> dict_keys(['audio', 'id', 'speaker', 'sentence', 'raw_sentence', 'start_timestamp', 'end_timestamp', 'overlap'])
### Expected behavior
All the columns should be present
### Environment info
- `datasets` version: 3.3.2
- Platform: macOS-14.6.1-x86_64-i386-64bit
- Python version: 3.10.15
- `huggingface_hub` version: 0.30.1
- PyArrow version: 16.1.0
- Pandas version: 1.5.3
- `fsspec` version: 2023.10.0
| {
"avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4",
"events_url": "https://api.github.com/users/bruno-hays/events{/privacy}",
"followers_url": "https://api.github.com/users/bruno-hays/followers",
"following_url": "https://api.github.com/users/bruno-hays/following{/other_user}",
"gists_url": "https://api.github.com/users/bruno-hays/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/bruno-hays",
"id": 48770768,
"login": "bruno-hays",
"node_id": "MDQ6VXNlcjQ4NzcwNzY4",
"organizations_url": "https://api.github.com/users/bruno-hays/orgs",
"received_events_url": "https://api.github.com/users/bruno-hays/received_events",
"repos_url": "https://api.github.com/users/bruno-hays/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/bruno-hays/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bruno-hays/subscriptions",
"type": "User",
"url": "https://api.github.com/users/bruno-hays",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7495/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7495/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7494 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7494/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7494/comments | https://api.github.com/repos/huggingface/datasets/issues/7494/events | https://github.com/huggingface/datasets/issues/7494 | 2,965,347,685 | I_kwDODunzps6wv51l | 7,494 | Broken links in pdf loading documentation | {
"avatar_url": "https://avatars.githubusercontent.com/u/75789232?v=4",
"events_url": "https://api.github.com/users/VyoJ/events{/privacy}",
"followers_url": "https://api.github.com/users/VyoJ/followers",
"following_url": "https://api.github.com/users/VyoJ/following{/other_user}",
"gists_url": "https://api.github.com/users/VyoJ/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/VyoJ",
"id": 75789232,
"login": "VyoJ",
"node_id": "MDQ6VXNlcjc1Nzg5MjMy",
"organizations_url": "https://api.github.com/users/VyoJ/orgs",
"received_events_url": "https://api.github.com/users/VyoJ/received_events",
"repos_url": "https://api.github.com/users/VyoJ/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/VyoJ/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/VyoJ/subscriptions",
"type": "User",
"url": "https://api.github.com/users/VyoJ",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"thanks for reporting ! I fixed the links, the docs will be updated in the next release"
] | 2025-04-02T06:45:22Z | 2025-04-15T13:36:25Z | 2025-04-15T13:36:04Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Hi, just a couple of small issues I ran into while reading the docs for [loading pdf data](https://huggingface.co/docs/datasets/main/en/document_load):
1. The link for the [`Create a pdf dataset`](https://huggingface.co/docs/datasets/main/en/document_load#pdffolder) points to https://huggingface.co/docs/datasets/main/en/pdf_dataset instead of https://huggingface.co/docs/datasets/main/en/document_dataset and hence gives a 404 error.
2. At the top of the page, it's mentioned that to work with pdf datasets we need to have the `pdfplumber` package installed but the link to its installation guide points to `pytorch/vision` [installation instructions](https://github.com/pytorch/vision#installation) instead of `pdfplumber`'s [guide](https://github.com/jsvine/pdfplumber#installation)
I love the work on enabling pdf dataset support and these small tweaks would help everyone navigate the docs better. Thanks!
### Steps to reproduce the bug
The issue is on the [Load Document Data](https://huggingface.co/docs/datasets/main/en/document_load) page of the datasets docs.
### Expected behavior
1. For solving the first issue, I went through the [source .mdx code](https://github.com/huggingface/datasets/blob/main/docs/source/document_load.mdx?plain=1#L188) of the datasets docs and found that the link is pointing to `./pdf_dataset` instead of `./document_dataset`
2. For the second issue, I went through the [source .mdx code](https://github.com/huggingface/datasets/blob/main/docs/source/document_load.mdx?plain=1#L13) of the datasets docs and found that the link is `pytorch/vision` [installation instructions](https://github.com/pytorch/vision#installation) instead of `pdfplumber`'s [guide](https://github.com/jsvine/pdfplumber#installation)
Just replacing these two links should fix the bugs
### Environment info
datasets v3.5.0 (main at the time of writing) | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7494/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7494/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7493 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7493/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7493/comments | https://api.github.com/repos/huggingface/datasets/issues/7493/events | https://github.com/huggingface/datasets/issues/7493 | 2,964,025,179 | I_kwDODunzps6wq29b | 7,493 | push_to_hub does not upload videos | {
"avatar_url": "https://avatars.githubusercontent.com/u/9339403?v=4",
"events_url": "https://api.github.com/users/DominikVincent/events{/privacy}",
"followers_url": "https://api.github.com/users/DominikVincent/followers",
"following_url": "https://api.github.com/users/DominikVincent/following{/other_user}",
"gists_url": "https://api.github.com/users/DominikVincent/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/DominikVincent",
"id": 9339403,
"login": "DominikVincent",
"node_id": "MDQ6VXNlcjkzMzk0MDM=",
"organizations_url": "https://api.github.com/users/DominikVincent/orgs",
"received_events_url": "https://api.github.com/users/DominikVincent/received_events",
"repos_url": "https://api.github.com/users/DominikVincent/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/DominikVincent/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/DominikVincent/subscriptions",
"type": "User",
"url": "https://api.github.com/users/DominikVincent",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! the `Video` type is still experimental, and in particular `push_to_hub` doesn't upload videos at the moment (only the paths).\n\nThere is an open question to either upload the videos inside the Parquet files, or rather have them as separate files (which is great to enable remote seeking/streaming)"
] | 2025-04-01T17:00:20Z | 2025-04-15T12:34:23Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Hello,
I would like to upload a video dataset (some .mp4 files and some segments within them), i.e. rows correspond to subsequences from videos. Videos might be referenced by several rows.
I created a dataset locally and it references the videos and the video readers can read them correctly. I use push_to_hub() to upload the dataset to the hub.
Expectation: A user uses `load_dataset` and can load the videos.
However, the videos seem to be just referenced via paths on the computer and not uploaded to the hub. Therefore a target user cannot load the videos in the dataset.
### Steps to reproduce the bug
1. create a video dataset with paths e.g. { ["videos"]: [path1, path2, ...]}
2. dataset.push_to_hub
3. on a different computer (or same pc if relative paths are used in a different folder):
```
dataset = load_dataset("siplab/egosim", split="train")
video = dataset[0]["video_head"]
```
3. will fail
### Expected behavior
Expectation: A user uses `load_dataset` and can load the videos.
### Environment info
datasets 3.1.0
Python 3.8.18 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7493/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7493/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7492 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7492/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7492/comments | https://api.github.com/repos/huggingface/datasets/issues/7492/events | https://github.com/huggingface/datasets/pull/7492 | 2,959,088,568 | PR_kwDODunzps6QtCQM | 7,492 | Closes #7457 | {
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"This PR fixes issue #7457"
] | 2025-03-30T20:41:20Z | 2025-04-13T22:05:07Z | 2025-04-13T22:05:07Z | CONTRIBUTOR | null | null | null | This PR updates the documentation to include the HF_DATASETS_CACHE environment variable, which allows users to customize the cache location for datasetsβsimilar to HF_HUB_CACHE for models. | {
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7492/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7492/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7492.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7492",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7492.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7492"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7491 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7491/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7491/comments | https://api.github.com/repos/huggingface/datasets/issues/7491/events | https://github.com/huggingface/datasets/pull/7491 | 2,959,085,647 | PR_kwDODunzps6QtBsD | 7,491 | docs: update cache.mdx to include HF_DATASETS_CACHE documentation | {
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Already included HF_DATASETS_CACHE"
] | 2025-03-30T20:35:03Z | 2025-03-30T20:36:40Z | 2025-03-30T20:36:40Z | CONTRIBUTOR | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7491/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7491/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7491.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7491",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7491.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7491"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7490 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7490/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7490/comments | https://api.github.com/repos/huggingface/datasets/issues/7490/events | https://github.com/huggingface/datasets/pull/7490 | 2,958,826,222 | PR_kwDODunzps6QsPUI | 7,490 | (refactor) remove redundant logic in _check_valid_index_key | {
"avatar_url": "https://avatars.githubusercontent.com/u/2980993?v=4",
"events_url": "https://api.github.com/users/suzyahyah/events{/privacy}",
"followers_url": "https://api.github.com/users/suzyahyah/followers",
"following_url": "https://api.github.com/users/suzyahyah/following{/other_user}",
"gists_url": "https://api.github.com/users/suzyahyah/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/suzyahyah",
"id": 2980993,
"login": "suzyahyah",
"node_id": "MDQ6VXNlcjI5ODA5OTM=",
"organizations_url": "https://api.github.com/users/suzyahyah/orgs",
"received_events_url": "https://api.github.com/users/suzyahyah/received_events",
"repos_url": "https://api.github.com/users/suzyahyah/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/suzyahyah/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/suzyahyah/subscriptions",
"type": "User",
"url": "https://api.github.com/users/suzyahyah",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-03-30T11:45:42Z | 2025-03-30T11:50:22Z | null | NONE | null | null | null | This PR contributes a minor refactor, in a small function in `src/datasets/formatting/formatting.py`. No change in logic.
In the original code, there are separate if-conditionals for `isinstance(key, range)` and `isinstance(key, Iterable)`, with essentially the same logic.
This PR combines these two using a single if statement.
**Considerations**
1. Although range in python is guaranteed to have integers, internally calling `int()` on an object that is already an int is negligible. (In python it returns the original object. It doesn't create a new integer object or perform any actual conversion)
2. Technically a range is already an Iterable, and we could just do `isinstance(key, Iterable)` but I explicitly did `isinstance(key, (range, Iterable))` just to be super obvious and consistent that both cases are handled because I see `slice, range, Iterable` everywhere in this `formatting.py`
3. This PR removes the `if len(key)>0` conditional. I think it is cleaner to have it this way for three reasons.
- There was originally no else statement and the code would have failed silently anyway.
- The if len(key)>0 should be caught much earlier, rather than in `formatting.py`.
- There are actually multiple cases where this would fail, if len(key)>0, if key is non numeric or float, or if key is a list of lists. It's clunky to state all this and the error be thrown during max or indexing.
**Previous PR and Issues Checks**
1. No known PR or Issues (both closed or open) in hf datasets repository
**Tests**
1. Tested using Dataset (load_dataset("wikitext", "wikitext-103-raw-v1")), Pytorch DataLoader, with a Pytorch BatchSampler (list of indexes returned instead of single index).
| null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7490/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7490/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7490.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7490",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7490.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7490"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7489 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7489/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7489/comments | https://api.github.com/repos/huggingface/datasets/issues/7489/events | https://github.com/huggingface/datasets/pull/7489 | 2,958,204,763 | PR_kwDODunzps6QqSRD | 7,489 | fix: loading of datasets from Disk(#7373) | {
"avatar_url": "https://avatars.githubusercontent.com/u/40773225?v=4",
"events_url": "https://api.github.com/users/sam-hey/events{/privacy}",
"followers_url": "https://api.github.com/users/sam-hey/followers",
"following_url": "https://api.github.com/users/sam-hey/following{/other_user}",
"gists_url": "https://api.github.com/users/sam-hey/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sam-hey",
"id": 40773225,
"login": "sam-hey",
"node_id": "MDQ6VXNlcjQwNzczMjI1",
"organizations_url": "https://api.github.com/users/sam-hey/orgs",
"received_events_url": "https://api.github.com/users/sam-hey/received_events",
"repos_url": "https://api.github.com/users/sam-hey/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sam-hey/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sam-hey/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sam-hey",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"@nepfaff Could you confirm if this fixes the issue for you? I checked Memray, and everything looked good on my end.\r\n\r\nInstall: `pip install git+https://github.com/sam-hey/datasets.git@fix/concatenate_datasets`\r\n",
"Will aim to get to this soon. I don't have a rapid testing pipeline setup but need to wait for some AWS nodes to become free",
"I now set up a small experiment:\r\n\r\n```python\r\n# Log initial RAM usage\r\n process = psutil.Process(os.getpid())\r\n initial_ram = process.memory_info().rss / (1024 * 1024) # Convert to MB\r\n logging.info(f\"Initial RAM usage: {initial_ram:.2f} MB\")\r\n\r\n chunk_datasets = [\r\n Dataset.load_from_disk(dataset_path, keep_in_memory=False) for _ in range(N)\r\n ]\r\n combined_dataset = concatenate_datasets(chunk_datasets)\r\n\r\n # Log final RAM usage\r\n final_ram = process.memory_info().rss / (1024 * 1024) # Convert to MB\r\n ram_diff = final_ram - initial_ram\r\n logging.info(f\"Final RAM usage: {final_ram:.2f} MB\")\r\n logging.info(f\"RAM usage increase: {ram_diff:.2f} MB\")\r\n```\r\n\r\nThe RAM usage is linearly correlated with `N` on datasets master!\r\n\r\nFor my test dataset:\r\n- N=5 => RAM usage increase: 26302.91 MB\r\n- N=10 => RAM usage increase: 52315.18 MB\r\n- N=20 => RAM usage increase: 104510.65 MB\r\n- N=40 => RAM usage increase: 209166.30 MB\r\n\r\nUnfortunately, your patch doesn't seem to change this:\r\n```bash\r\npip install git+https://github.com/sam-hey/datasets.git@fix/concatenate_datasets\r\npip list | grep datasets\r\ndatasets 3.5.1.dev0\r\n```\r\nGives exactly the same RAM statistics.\r\n\r\n**Edit:** The results are a bit flawed as the memory increase all seems to come from `Dataset.load_from_disk(dataset_path, keep_in_memory=False)` here (which I don't think should happen either?) and not from `concatenate_datasets`. This seems different from my large-scale setup that runs out of memory during `concatenate_datasets` but I don't seem to be able to replicate this here...",
"Thanks a lot, @nepfaff, for taking a look at this! It seems that `concatenate_datasets()` is fixed with this PR. I can also confirm that loading a large number of files requires significant memory. However, as I understand it, this is expected/a bug since the memory consumption stems from `pa.memory_map()`, which returns a memory-mapped file.\r\n\r\nThis behavior might be related to this bug: https://github.com/apache/arrow/issues/34423 \r\n\r\n<img width=\"1728\" alt=\"Screenshot 2025-04-03 at 16 01 11\" src=\"https://github.com/user-attachments/assets/475691d8-3aba-4d7e-b8ef-5e7552c70b14\" />\r\n",
"Great ! have you tested that it also fixes the memory issue in your case @iamollas ?\r\n\r\nHappy to know that it works for you @sam-hey ! Looking forward to merging this",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7489). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-03-29T16:22:58Z | 2025-04-24T16:36:36Z | null | NONE | null | null | null | Fixes dataset loading from disk by ensuring that memory maps and streams are properly closed.
For more details, see https://github.com/huggingface/datasets/issues/7373. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7489/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7489/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7489.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7489",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7489.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7489"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7488 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7488/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7488/comments | https://api.github.com/repos/huggingface/datasets/issues/7488/events | https://github.com/huggingface/datasets/pull/7488 | 2,956,559,358 | PR_kwDODunzps6QlLmn | 7,488 | Support underscore int read instruction | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7488). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"you rock, Quentin - thank you!"
] | 2025-03-28T16:01:15Z | 2025-03-28T16:20:44Z | 2025-03-28T16:20:43Z | MEMBER | null | null | null | close https://github.com/huggingface/datasets/issues/7481 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7488/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7488/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7488.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7488",
"merged_at": "2025-03-28T16:20:43Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7488.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7488"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7487 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7487/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7487/comments | https://api.github.com/repos/huggingface/datasets/issues/7487/events | https://github.com/huggingface/datasets/pull/7487 | 2,956,533,448 | PR_kwDODunzps6QlF8N | 7,487 | Write pdf in map | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7487). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-03-28T15:49:25Z | 2025-03-28T17:09:53Z | 2025-03-28T17:09:51Z | MEMBER | null | null | null | Fix this error when mapping a PDF dataset
```
pyarrow.lib.ArrowInvalid: Could not convert <pdfplumber.pdf.PDF object at 0x13498ee40> with type PDF: did not recognize Python value type when inferring an Arrow data type
```
and also let map() outputs be lists of images or pdfs | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7487/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7487/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7487.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7487",
"merged_at": "2025-03-28T17:09:51Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7487.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7487"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7486 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7486/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7486/comments | https://api.github.com/repos/huggingface/datasets/issues/7486/events | https://github.com/huggingface/datasets/issues/7486 | 2,954,042,179 | I_kwDODunzps6wExtD | 7,486 | `shared_datadir` fixture is missing | {
"avatar_url": "https://avatars.githubusercontent.com/u/1289205?v=4",
"events_url": "https://api.github.com/users/lahwaacz/events{/privacy}",
"followers_url": "https://api.github.com/users/lahwaacz/followers",
"following_url": "https://api.github.com/users/lahwaacz/following{/other_user}",
"gists_url": "https://api.github.com/users/lahwaacz/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lahwaacz",
"id": 1289205,
"login": "lahwaacz",
"node_id": "MDQ6VXNlcjEyODkyMDU=",
"organizations_url": "https://api.github.com/users/lahwaacz/orgs",
"received_events_url": "https://api.github.com/users/lahwaacz/received_events",
"repos_url": "https://api.github.com/users/lahwaacz/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lahwaacz/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lahwaacz/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lahwaacz",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"OK I was missing the `pytest-datadir` package. Sorry for the noise!"
] | 2025-03-27T18:17:12Z | 2025-03-27T19:49:11Z | 2025-03-27T19:49:10Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Running the tests for the latest release fails due to missing `shared_datadir` fixture.
### Steps to reproduce the bug
Running `pytest` while building a package for Arch Linux leads to these errors:
```
==================================== ERRORS ====================================
_________ ERROR at setup of test_pdf_feature_encode_example[<lambda>1] _________
[gw44] linux -- Python 3.13.2 /build/python-datasets/src/datasets-3.5.0/test-env/bin/python
file /build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py, line 8
@require_pdfplumber
@pytest.mark.parametrize(
"build_example",
[
lambda pdf_path: pdf_path,
lambda pdf_path: open(pdf_path, "rb").read(),
lambda pdf_path: {"path": pdf_path},
lambda pdf_path: {"path": pdf_path, "bytes": None},
lambda pdf_path: {"path": pdf_path, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"path": None, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"bytes": open(pdf_path, "rb").read()},
],
)
def test_pdf_feature_encode_example(shared_datadir, build_example):
E fixture 'shared_datadir' not found
> available fixtures: _hf_gated_dataset_repo_txt_data, arrow_file, arrow_path, audio_file, bz2_csv_path, bz2_file, cache, capfd, capfdbinary, caplog, capsys, capsysbinary, ci_hfh_hf_hub_url, ci_hub_config, cleanup_repo, csv2_path, csv_path, data_dir_with_hidden_files, dataset, dataset_dict, disable_implicit_token, disable_tqdm_output, doctest_namespace, geoparquet_path, gz_file, hf_api, hf_gated_dataset_repo_txt_data, hf_private_dataset_repo_txt_data, hf_private_dataset_repo_txt_data_, hf_private_dataset_repo_zipped_img_data, hf_private_dataset_repo_zipped_img_data_, hf_private_dataset_repo_zipped_txt_data, hf_private_dataset_repo_zipped_txt_data_, hf_token, image_file, json_dict_of_lists_path, json_list_of_dicts_path, jsonl2_path, jsonl_312_path, jsonl_gz_path, jsonl_path, jsonl_str_path, lz4_file, mock_fsspec, mockfs, monkeypatch, parquet_path, pytestconfig, record_property, record_testsuite_property, record_xml_attribute, recwarn, set_ci_hub_access_token, set_sqlalchemy_silence_uber_warning, set_test_cache_config, set_update_download_counts_to_false, seven_zip_file, sqlite_path, tar_file, tar_jsonl_path, tar_nested_jsonl_path, temporary_repo, tensor_file, testrun_uid, text2_path, text_dir, text_dir_with_unsupported_extension, text_file, text_file_content, text_gz_path, text_path, text_path_with_unicode_new_lines, tmp_path, tmp_path_factory, tmpdir, tmpdir_factory, tmpfs, worker_id, xml_file, xz_file, zero_time_out_for_remote_code, zip_csv_path, zip_csv_with_dir_path, zip_file, zip_image_path, zip_jsonl_path, zip_jsonl_with_dir_path, zip_nested_jsonl_path, zip_text_path, zip_text_with_dir_path, zip_unsupported_ext_path, zip_uppercase_csv_path, zstd_file
> use 'pytest --fixtures [testpath]' for help on them.
/build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py:8
_________ ERROR at setup of test_pdf_feature_encode_example[<lambda>2] _________
[gw44] linux -- Python 3.13.2 /build/python-datasets/src/datasets-3.5.0/test-env/bin/python
file /build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py, line 8
@require_pdfplumber
@pytest.mark.parametrize(
"build_example",
[
lambda pdf_path: pdf_path,
lambda pdf_path: open(pdf_path, "rb").read(),
lambda pdf_path: {"path": pdf_path},
lambda pdf_path: {"path": pdf_path, "bytes": None},
lambda pdf_path: {"path": pdf_path, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"path": None, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"bytes": open(pdf_path, "rb").read()},
],
)
def test_pdf_feature_encode_example(shared_datadir, build_example):
E fixture 'shared_datadir' not found
> available fixtures: _hf_gated_dataset_repo_txt_data, arrow_file, arrow_path, audio_file, bz2_csv_path, bz2_file, cache, capfd, capfdbinary, caplog, capsys, capsysbinary, ci_hfh_hf_hub_url, ci_hub_config, cleanup_repo, csv2_path, csv_path, data_dir_with_hidden_files, dataset, dataset_dict, disable_implicit_token, disable_tqdm_output, doctest_namespace, geoparquet_path, gz_file, hf_api, hf_gated_dataset_repo_txt_data, hf_private_dataset_repo_txt_data, hf_private_dataset_repo_txt_data_, hf_private_dataset_repo_zipped_img_data, hf_private_dataset_repo_zipped_img_data_, hf_private_dataset_repo_zipped_txt_data, hf_private_dataset_repo_zipped_txt_data_, hf_token, image_file, json_dict_of_lists_path, json_list_of_dicts_path, jsonl2_path, jsonl_312_path, jsonl_gz_path, jsonl_path, jsonl_str_path, lz4_file, mock_fsspec, mockfs, monkeypatch, parquet_path, pytestconfig, record_property, record_testsuite_property, record_xml_attribute, recwarn, set_ci_hub_access_token, set_sqlalchemy_silence_uber_warning, set_test_cache_config, set_update_download_counts_to_false, seven_zip_file, sqlite_path, tar_file, tar_jsonl_path, tar_nested_jsonl_path, temporary_repo, tensor_file, testrun_uid, text2_path, text_dir, text_dir_with_unsupported_extension, text_file, text_file_content, text_gz_path, text_path, text_path_with_unicode_new_lines, tmp_path, tmp_path_factory, tmpdir, tmpdir_factory, tmpfs, worker_id, xml_file, xz_file, zero_time_out_for_remote_code, zip_csv_path, zip_csv_with_dir_path, zip_file, zip_image_path, zip_jsonl_path, zip_jsonl_with_dir_path, zip_nested_jsonl_path, zip_text_path, zip_text_with_dir_path, zip_unsupported_ext_path, zip_uppercase_csv_path, zstd_file
> use 'pytest --fixtures [testpath]' for help on them.
/build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py:8
_________ ERROR at setup of test_pdf_feature_encode_example[<lambda>3] _________
[gw44] linux -- Python 3.13.2 /build/python-datasets/src/datasets-3.5.0/test-env/bin/python
file /build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py, line 8
@require_pdfplumber
@pytest.mark.parametrize(
"build_example",
[
lambda pdf_path: pdf_path,
lambda pdf_path: open(pdf_path, "rb").read(),
lambda pdf_path: {"path": pdf_path},
lambda pdf_path: {"path": pdf_path, "bytes": None},
lambda pdf_path: {"path": pdf_path, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"path": None, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"bytes": open(pdf_path, "rb").read()},
],
)
def test_pdf_feature_encode_example(shared_datadir, build_example):
E fixture 'shared_datadir' not found
> available fixtures: _hf_gated_dataset_repo_txt_data, arrow_file, arrow_path, audio_file, bz2_csv_path, bz2_file, cache, capfd, capfdbinary, caplog, capsys, capsysbinary, ci_hfh_hf_hub_url, ci_hub_config, cleanup_repo, csv2_path, csv_path, data_dir_with_hidden_files, dataset, dataset_dict, disable_implicit_token, disable_tqdm_output, doctest_namespace, geoparquet_path, gz_file, hf_api, hf_gated_dataset_repo_txt_data, hf_private_dataset_repo_txt_data, hf_private_dataset_repo_txt_data_, hf_private_dataset_repo_zipped_img_data, hf_private_dataset_repo_zipped_img_data_, hf_private_dataset_repo_zipped_txt_data, hf_private_dataset_repo_zipped_txt_data_, hf_token, image_file, json_dict_of_lists_path, json_list_of_dicts_path, jsonl2_path, jsonl_312_path, jsonl_gz_path, jsonl_path, jsonl_str_path, lz4_file, mock_fsspec, mockfs, monkeypatch, parquet_path, pytestconfig, record_property, record_testsuite_property, record_xml_attribute, recwarn, set_ci_hub_access_token, set_sqlalchemy_silence_uber_warning, set_test_cache_config, set_update_download_counts_to_false, seven_zip_file, sqlite_path, tar_file, tar_jsonl_path, tar_nested_jsonl_path, temporary_repo, tensor_file, testrun_uid, text2_path, text_dir, text_dir_with_unsupported_extension, text_file, text_file_content, text_gz_path, text_path, text_path_with_unicode_new_lines, tmp_path, tmp_path_factory, tmpdir, tmpdir_factory, tmpfs, worker_id, xml_file, xz_file, zero_time_out_for_remote_code, zip_csv_path, zip_csv_with_dir_path, zip_file, zip_image_path, zip_jsonl_path, zip_jsonl_with_dir_path, zip_nested_jsonl_path, zip_text_path, zip_text_with_dir_path, zip_unsupported_ext_path, zip_uppercase_csv_path, zstd_file
> use 'pytest --fixtures [testpath]' for help on them.
/build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py:8
_________ ERROR at setup of test_pdf_feature_encode_example[<lambda>4] _________
[gw44] linux -- Python 3.13.2 /build/python-datasets/src/datasets-3.5.0/test-env/bin/python
file /build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py, line 8
@require_pdfplumber
@pytest.mark.parametrize(
"build_example",
[
lambda pdf_path: pdf_path,
lambda pdf_path: open(pdf_path, "rb").read(),
lambda pdf_path: {"path": pdf_path},
lambda pdf_path: {"path": pdf_path, "bytes": None},
lambda pdf_path: {"path": pdf_path, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"path": None, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"bytes": open(pdf_path, "rb").read()},
],
)
def test_pdf_feature_encode_example(shared_datadir, build_example):
E fixture 'shared_datadir' not found
> available fixtures: _hf_gated_dataset_repo_txt_data, arrow_file, arrow_path, audio_file, bz2_csv_path, bz2_file, cache, capfd, capfdbinary, caplog, capsys, capsysbinary, ci_hfh_hf_hub_url, ci_hub_config, cleanup_repo, csv2_path, csv_path, data_dir_with_hidden_files, dataset, dataset_dict, disable_implicit_token, disable_tqdm_output, doctest_namespace, geoparquet_path, gz_file, hf_api, hf_gated_dataset_repo_txt_data, hf_private_dataset_repo_txt_data, hf_private_dataset_repo_txt_data_, hf_private_dataset_repo_zipped_img_data, hf_private_dataset_repo_zipped_img_data_, hf_private_dataset_repo_zipped_txt_data, hf_private_dataset_repo_zipped_txt_data_, hf_token, image_file, json_dict_of_lists_path, json_list_of_dicts_path, jsonl2_path, jsonl_312_path, jsonl_gz_path, jsonl_path, jsonl_str_path, lz4_file, mock_fsspec, mockfs, monkeypatch, parquet_path, pytestconfig, record_property, record_testsuite_property, record_xml_attribute, recwarn, set_ci_hub_access_token, set_sqlalchemy_silence_uber_warning, set_test_cache_config, set_update_download_counts_to_false, seven_zip_file, sqlite_path, tar_file, tar_jsonl_path, tar_nested_jsonl_path, temporary_repo, tensor_file, testrun_uid, text2_path, text_dir, text_dir_with_unsupported_extension, text_file, text_file_content, text_gz_path, text_path, text_path_with_unicode_new_lines, tmp_path, tmp_path_factory, tmpdir, tmpdir_factory, tmpfs, worker_id, xml_file, xz_file, zero_time_out_for_remote_code, zip_csv_path, zip_csv_with_dir_path, zip_file, zip_image_path, zip_jsonl_path, zip_jsonl_with_dir_path, zip_nested_jsonl_path, zip_text_path, zip_text_with_dir_path, zip_unsupported_ext_path, zip_uppercase_csv_path, zstd_file
> use 'pytest --fixtures [testpath]' for help on them.
/build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py:8
_________ ERROR at setup of test_pdf_feature_encode_example[<lambda>5] _________
[gw44] linux -- Python 3.13.2 /build/python-datasets/src/datasets-3.5.0/test-env/bin/python
file /build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py, line 8
@require_pdfplumber
@pytest.mark.parametrize(
"build_example",
[
lambda pdf_path: pdf_path,
lambda pdf_path: open(pdf_path, "rb").read(),
lambda pdf_path: {"path": pdf_path},
lambda pdf_path: {"path": pdf_path, "bytes": None},
lambda pdf_path: {"path": pdf_path, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"path": None, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"bytes": open(pdf_path, "rb").read()},
],
)
def test_pdf_feature_encode_example(shared_datadir, build_example):
E fixture 'shared_datadir' not found
> available fixtures: _hf_gated_dataset_repo_txt_data, arrow_file, arrow_path, audio_file, bz2_csv_path, bz2_file, cache, capfd, capfdbinary, caplog, capsys, capsysbinary, ci_hfh_hf_hub_url, ci_hub_config, cleanup_repo, csv2_path, csv_path, data_dir_with_hidden_files, dataset, dataset_dict, disable_implicit_token, disable_tqdm_output, doctest_namespace, geoparquet_path, gz_file, hf_api, hf_gated_dataset_repo_txt_data, hf_private_dataset_repo_txt_data, hf_private_dataset_repo_txt_data_, hf_private_dataset_repo_zipped_img_data, hf_private_dataset_repo_zipped_img_data_, hf_private_dataset_repo_zipped_txt_data, hf_private_dataset_repo_zipped_txt_data_, hf_token, image_file, json_dict_of_lists_path, json_list_of_dicts_path, jsonl2_path, jsonl_312_path, jsonl_gz_path, jsonl_path, jsonl_str_path, lz4_file, mock_fsspec, mockfs, monkeypatch, parquet_path, pytestconfig, record_property, record_testsuite_property, record_xml_attribute, recwarn, set_ci_hub_access_token, set_sqlalchemy_silence_uber_warning, set_test_cache_config, set_update_download_counts_to_false, seven_zip_file, sqlite_path, tar_file, tar_jsonl_path, tar_nested_jsonl_path, temporary_repo, tensor_file, testrun_uid, text2_path, text_dir, text_dir_with_unsupported_extension, text_file, text_file_content, text_gz_path, text_path, text_path_with_unicode_new_lines, tmp_path, tmp_path_factory, tmpdir, tmpdir_factory, tmpfs, worker_id, xml_file, xz_file, zero_time_out_for_remote_code, zip_csv_path, zip_csv_with_dir_path, zip_file, zip_image_path, zip_jsonl_path, zip_jsonl_with_dir_path, zip_nested_jsonl_path, zip_text_path, zip_text_with_dir_path, zip_unsupported_ext_path, zip_uppercase_csv_path, zstd_file
> use 'pytest --fixtures [testpath]' for help on them.
/build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py:8
_________ ERROR at setup of test_pdf_feature_encode_example[<lambda>6] _________
[gw44] linux -- Python 3.13.2 /build/python-datasets/src/datasets-3.5.0/test-env/bin/python
file /build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py, line 8
@require_pdfplumber
@pytest.mark.parametrize(
"build_example",
[
lambda pdf_path: pdf_path,
lambda pdf_path: open(pdf_path, "rb").read(),
lambda pdf_path: {"path": pdf_path},
lambda pdf_path: {"path": pdf_path, "bytes": None},
lambda pdf_path: {"path": pdf_path, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"path": None, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"bytes": open(pdf_path, "rb").read()},
],
)
def test_pdf_feature_encode_example(shared_datadir, build_example):
E fixture 'shared_datadir' not found
> available fixtures: _hf_gated_dataset_repo_txt_data, arrow_file, arrow_path, audio_file, bz2_csv_path, bz2_file, cache, capfd, capfdbinary, caplog, capsys, capsysbinary, ci_hfh_hf_hub_url, ci_hub_config, cleanup_repo, csv2_path, csv_path, data_dir_with_hidden_files, dataset, dataset_dict, disable_implicit_token, disable_tqdm_output, doctest_namespace, geoparquet_path, gz_file, hf_api, hf_gated_dataset_repo_txt_data, hf_private_dataset_repo_txt_data, hf_private_dataset_repo_txt_data_, hf_private_dataset_repo_zipped_img_data, hf_private_dataset_repo_zipped_img_data_, hf_private_dataset_repo_zipped_txt_data, hf_private_dataset_repo_zipped_txt_data_, hf_token, image_file, json_dict_of_lists_path, json_list_of_dicts_path, jsonl2_path, jsonl_312_path, jsonl_gz_path, jsonl_path, jsonl_str_path, lz4_file, mock_fsspec, mockfs, monkeypatch, parquet_path, pytestconfig, record_property, record_testsuite_property, record_xml_attribute, recwarn, set_ci_hub_access_token, set_sqlalchemy_silence_uber_warning, set_test_cache_config, set_update_download_counts_to_false, seven_zip_file, sqlite_path, tar_file, tar_jsonl_path, tar_nested_jsonl_path, temporary_repo, tensor_file, testrun_uid, text2_path, text_dir, text_dir_with_unsupported_extension, text_file, text_file_content, text_gz_path, text_path, text_path_with_unicode_new_lines, tmp_path, tmp_path_factory, tmpdir, tmpdir_factory, tmpfs, worker_id, xml_file, xz_file, zero_time_out_for_remote_code, zip_csv_path, zip_csv_with_dir_path, zip_file, zip_image_path, zip_jsonl_path, zip_jsonl_with_dir_path, zip_nested_jsonl_path, zip_text_path, zip_text_with_dir_path, zip_unsupported_ext_path, zip_uppercase_csv_path, zstd_file
> use 'pytest --fixtures [testpath]' for help on them.
/build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py:8
_______________ ERROR at setup of test_dataset_with_pdf_feature ________________
[gw44] linux -- Python 3.13.2 /build/python-datasets/src/datasets-3.5.0/test-env/bin/python
file /build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py, line 34
@require_pdfplumber
def test_dataset_with_pdf_feature(shared_datadir):
E fixture 'shared_datadir' not found
> available fixtures: _hf_gated_dataset_repo_txt_data, arrow_file, arrow_path, audio_file, bz2_csv_path, bz2_file, cache, capfd, capfdbinary, caplog, capsys, capsysbinary, ci_hfh_hf_hub_url, ci_hub_config, cleanup_repo, csv2_path, csv_path, data_dir_with_hidden_files, dataset, dataset_dict, disable_implicit_token, disable_tqdm_output, doctest_namespace, geoparquet_path, gz_file, hf_api, hf_gated_dataset_repo_txt_data, hf_private_dataset_repo_txt_data, hf_private_dataset_repo_txt_data_, hf_private_dataset_repo_zipped_img_data, hf_private_dataset_repo_zipped_img_data_, hf_private_dataset_repo_zipped_txt_data, hf_private_dataset_repo_zipped_txt_data_, hf_token, image_file, json_dict_of_lists_path, json_list_of_dicts_path, jsonl2_path, jsonl_312_path, jsonl_gz_path, jsonl_path, jsonl_str_path, lz4_file, mock_fsspec, mockfs, monkeypatch, parquet_path, pytestconfig, record_property, record_testsuite_property, record_xml_attribute, recwarn, set_ci_hub_access_token, set_sqlalchemy_silence_uber_warning, set_test_cache_config, set_update_download_counts_to_false, seven_zip_file, sqlite_path, tar_file, tar_jsonl_path, tar_nested_jsonl_path, temporary_repo, tensor_file, testrun_uid, text2_path, text_dir, text_dir_with_unsupported_extension, text_file, text_file_content, text_gz_path, text_path, text_path_with_unicode_new_lines, tmp_path, tmp_path_factory, tmpdir, tmpdir_factory, tmpfs, worker_id, xml_file, xz_file, zero_time_out_for_remote_code, zip_csv_path, zip_csv_with_dir_path, zip_file, zip_image_path, zip_jsonl_path, zip_jsonl_with_dir_path, zip_nested_jsonl_path, zip_text_path, zip_text_with_dir_path, zip_unsupported_ext_path, zip_uppercase_csv_path, zstd_file
> use 'pytest --fixtures [testpath]' for help on them.
/build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py:34
_________ ERROR at setup of test_pdf_feature_encode_example[<lambda>0] _________
[gw46] linux -- Python 3.13.2 /build/python-datasets/src/datasets-3.5.0/test-env/bin/python
file /build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py, line 8
@require_pdfplumber
@pytest.mark.parametrize(
"build_example",
[
lambda pdf_path: pdf_path,
lambda pdf_path: open(pdf_path, "rb").read(),
lambda pdf_path: {"path": pdf_path},
lambda pdf_path: {"path": pdf_path, "bytes": None},
lambda pdf_path: {"path": pdf_path, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"path": None, "bytes": open(pdf_path, "rb").read()},
lambda pdf_path: {"bytes": open(pdf_path, "rb").read()},
],
)
def test_pdf_feature_encode_example(shared_datadir, build_example):
E fixture 'shared_datadir' not found
> available fixtures: _hf_gated_dataset_repo_txt_data, arrow_file, arrow_path, audio_file, bz2_csv_path, bz2_file, cache, capfd, capfdbinary, caplog, capsys, capsysbinary, ci_hfh_hf_hub_url, ci_hub_config, cleanup_repo, csv2_path, csv_path, data_dir_with_hidden_files, dataset, dataset_dict, disable_implicit_token, disable_tqdm_output, doctest_namespace, geoparquet_path, gz_file, hf_api, hf_gated_dataset_repo_txt_data, hf_private_dataset_repo_txt_data, hf_private_dataset_repo_txt_data_, hf_private_dataset_repo_zipped_img_data, hf_private_dataset_repo_zipped_img_data_, hf_private_dataset_repo_zipped_txt_data, hf_private_dataset_repo_zipped_txt_data_, hf_token, image_file, json_dict_of_lists_path, json_list_of_dicts_path, jsonl2_path, jsonl_312_path, jsonl_gz_path, jsonl_path, jsonl_str_path, lz4_file, mock_fsspec, mockfs, monkeypatch, parquet_path, pytestconfig, record_property, record_testsuite_property, record_xml_attribute, recwarn, set_ci_hub_access_token, set_sqlalchemy_silence_uber_warning, set_test_cache_config, set_update_download_counts_to_false, seven_zip_file, sqlite_path, tar_file, tar_jsonl_path, tar_nested_jsonl_path, temporary_repo, tensor_file, testrun_uid, text2_path, text_dir, text_dir_with_unsupported_extension, text_file, text_file_content, text_gz_path, text_path, text_path_with_unicode_new_lines, tmp_path, tmp_path_factory, tmpdir, tmpdir_factory, tmpfs, worker_id, xml_file, xz_file, zero_time_out_for_remote_code, zip_csv_path, zip_csv_with_dir_path, zip_file, zip_image_path, zip_jsonl_path, zip_jsonl_with_dir_path, zip_nested_jsonl_path, zip_text_path, zip_text_with_dir_path, zip_unsupported_ext_path, zip_uppercase_csv_path, zstd_file
> use 'pytest --fixtures [testpath]' for help on them.
/build/python-datasets/src/datasets-3.5.0/tests/features/test_pdf.py:8
```
### Expected behavior
All fixtures used in tests should be available.
### Environment info
Arch Linux build system, building the [python-datasets](https://gitlab.archlinux.org/archlinux/packaging/packages/python-datasets) package.
There are actually [many deselected tests](https://gitlab.archlinux.org/archlinux/packaging/packages/python-datasets/-/blob/6f97957f0c326cc7b3da6b7f12326305bcaef374/PKGBUILD#L66-148) which were failing on previous releases, but these errors popped up in 3.5.0. | {
"avatar_url": "https://avatars.githubusercontent.com/u/1289205?v=4",
"events_url": "https://api.github.com/users/lahwaacz/events{/privacy}",
"followers_url": "https://api.github.com/users/lahwaacz/followers",
"following_url": "https://api.github.com/users/lahwaacz/following{/other_user}",
"gists_url": "https://api.github.com/users/lahwaacz/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lahwaacz",
"id": 1289205,
"login": "lahwaacz",
"node_id": "MDQ6VXNlcjEyODkyMDU=",
"organizations_url": "https://api.github.com/users/lahwaacz/orgs",
"received_events_url": "https://api.github.com/users/lahwaacz/received_events",
"repos_url": "https://api.github.com/users/lahwaacz/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lahwaacz/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lahwaacz/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lahwaacz",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7486/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7486/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7485 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7485/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7485/comments | https://api.github.com/repos/huggingface/datasets/issues/7485/events | https://github.com/huggingface/datasets/pull/7485 | 2,953,696,519 | PR_kwDODunzps6QbjFJ | 7,485 | set dev version | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7485). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-03-27T16:39:34Z | 2025-03-27T16:41:59Z | 2025-03-27T16:39:42Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7485/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7485/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7485.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7485",
"merged_at": "2025-03-27T16:39:42Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7485.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7485"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7484 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7484/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7484/comments | https://api.github.com/repos/huggingface/datasets/issues/7484/events | https://github.com/huggingface/datasets/pull/7484 | 2,953,677,168 | PR_kwDODunzps6Qbevn | 7,484 | release: 3.5.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7484). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-03-27T16:33:27Z | 2025-03-27T16:35:44Z | 2025-03-27T16:34:22Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7484/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7484/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7484.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7484",
"merged_at": "2025-03-27T16:34:22Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7484.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7484"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7483 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7483/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7483/comments | https://api.github.com/repos/huggingface/datasets/issues/7483/events | https://github.com/huggingface/datasets/pull/7483 | 2,951,856,468 | PR_kwDODunzps6QVInB | 7,483 | Support skip_trying_type | {
"avatar_url": "https://avatars.githubusercontent.com/u/11156001?v=4",
"events_url": "https://api.github.com/users/yoshitomo-matsubara/events{/privacy}",
"followers_url": "https://api.github.com/users/yoshitomo-matsubara/followers",
"following_url": "https://api.github.com/users/yoshitomo-matsubara/following{/other_user}",
"gists_url": "https://api.github.com/users/yoshitomo-matsubara/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yoshitomo-matsubara",
"id": 11156001,
"login": "yoshitomo-matsubara",
"node_id": "MDQ6VXNlcjExMTU2MDAx",
"organizations_url": "https://api.github.com/users/yoshitomo-matsubara/orgs",
"received_events_url": "https://api.github.com/users/yoshitomo-matsubara/received_events",
"repos_url": "https://api.github.com/users/yoshitomo-matsubara/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yoshitomo-matsubara/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yoshitomo-matsubara/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yoshitomo-matsubara",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7483). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Cool ! Can you run `make style` to fix code formatting ?\r\n\r\nI was also thinking of naming the argument `try_original_type` and have it `True` by default",
"@lhoestq \r\n\r\nThank you for the suggestion! I renamed the argument with `True` by default and ran `make style`\r\nDoes it look good?",
"Thanks @lhoestq !\r\n\r\nLet me know if there are anything that I can do for this PR. Otherwise, looking forward to seeing this update in the package soon!",
"CI failures are unrelated, merging :)",
"Great, thanks for your support!\r\nI can't wait for the next release :)"
] | 2025-03-27T07:07:20Z | 2025-04-29T04:14:57Z | 2025-04-09T09:53:10Z | CONTRIBUTOR | null | null | null | This PR addresses Issue #7472
cc: @lhoestq | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7483/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7483/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7483.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7483",
"merged_at": "2025-04-09T09:53:10Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7483.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7483"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7482 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7482/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7482/comments | https://api.github.com/repos/huggingface/datasets/issues/7482/events | https://github.com/huggingface/datasets/pull/7482 | 2,950,890,368 | PR_kwDODunzps6QRyY6 | 7,482 | Implement capability to restore non-nullability in Features | {
"avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4",
"events_url": "https://api.github.com/users/BramVanroy/events{/privacy}",
"followers_url": "https://api.github.com/users/BramVanroy/followers",
"following_url": "https://api.github.com/users/BramVanroy/following{/other_user}",
"gists_url": "https://api.github.com/users/BramVanroy/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/BramVanroy",
"id": 2779410,
"login": "BramVanroy",
"node_id": "MDQ6VXNlcjI3Nzk0MTA=",
"organizations_url": "https://api.github.com/users/BramVanroy/orgs",
"received_events_url": "https://api.github.com/users/BramVanroy/received_events",
"repos_url": "https://api.github.com/users/BramVanroy/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/BramVanroy/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BramVanroy/subscriptions",
"type": "User",
"url": "https://api.github.com/users/BramVanroy",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Interestingly, this does not close #7479. The Features are not correctly maintained when calling `from_dict` with the custom Features.",
"Unfortunately this PR does not fix the reported issue. After more digging:\r\n\r\n- when the dataset is created, nullability information is lost in Features;\r\n- even with this PR, it will get lost eventually because of internal copying/recreation of the Features object without accounting for the nullable fields;\r\n- even if that is also fixed, and Features.arrow_schema correctly holds the nullability info, [casting the arrow Table](https://github.com/huggingface/datasets/blob/5f8d2ad9a1b0bccfd962d998987228addfd5be9f/src/datasets/arrow_dataset.py#L677) with a less strict schema to a more strict one (with nullability) will fail (only on deeper structs, not on flat fields). \r\n\r\nInterestingly, passing custom Features does not immediately load the underlying data with the right arrow_schema. Instead, the workflow is like this:\r\n\r\n- load pyarrow table with any of the methods (from_dict, from_pandas, etc.), which will always AUTO INFER rather than use a provided schema\r\n- the loaded table with auto-schema will be used to initialize the `Dataset` class, and only during construction will [CAST](https://github.com/huggingface/datasets/blob/5f8d2ad9a1b0bccfd962d998987228addfd5be9f/src/datasets/arrow_dataset.py#L677) the table to the user-provided schema if needed, if it differs from the auto-inferred one.\r\n\r\nSo I figured, since many/all of the pyarrow [`Table.from_*`](https://arrow.apache.org/docs/python/generated/pyarrow.Table.html) methods have a `schema=` argument, we should already load the Table with the correct schema to begin with. As an example, I tried changing this line:\r\n\r\nhttps://github.com/huggingface/datasets/blob/5f8d2ad9a1b0bccfd962d998987228addfd5be9f/src/datasets/arrow_dataset.py#L940\r\n\r\nto include the arrow_schema, if provided:\r\n\r\n```python\r\npa_table = InMemoryTable.from_pydict(mapping=mapping, schema=features.arrow_schema if features is not None else None)\r\n```\r\n\r\nBut that leads to:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"/home/ampere/vanroy/datasets/scratch.py\", line 33, in <module>\r\n ds = Dataset.from_dict(\r\n ^^^^^^^^^^^^^^^^^^\r\n File \"/home/local/vanroy/datasets/src/datasets/arrow_dataset.py\", line 957, in from_dict\r\n pa_table = InMemoryTable.from_pydict(mapping=mapping, schema=features.arrow_schema if features is not None else None)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"/home/local/vanroy/datasets/src/datasets/table.py\", line 758, in from_pydict\r\n return cls(pa.Table.from_pydict(*args, **kwargs))\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"pyarrow/table.pxi\", line 1968, in pyarrow.lib._Tabular.from_pydict\r\n File \"pyarrow/table.pxi\", line 6354, in pyarrow.lib._from_pydict\r\n File \"pyarrow/array.pxi\", line 402, in pyarrow.lib.asarray\r\n File \"pyarrow/array.pxi\", line 252, in pyarrow.lib.array\r\n File \"pyarrow/array.pxi\", line 114, in pyarrow.lib._handle_arrow_array_protocol\r\n File \"/home/local/vanroy/datasets/src/datasets/arrow_writer.py\", line 201, in __arrow_array__\r\n raise ValueError(\"TypedSequence is supposed to be used with pa.array(typed_sequence, type=None)\")\r\nValueError: TypedSequence is supposed to be used with pa.array(typed_sequence, type=None)\r\n```\r\n\r\nand I am not too familiar with pyarrow to solve this.\r\n\r\nSo ultimately I'm a bit at a loss here. I *think*, if we'd want to do this right, the automatic casting in init should be removed in favor of handling the logic inside `Dataset.from_*`, by passing the schema explicitly to `pa.Table.from_*(..., schema=schema)`. But I lack the knowledge of pyarrow to go further than what I've written about above.\r\n",
"It's indeed a bit more work to support nullable since in addition to your comments, there are unclear behavior when it comes to concatenating nullable with non-nullable, and maybe how to handle non-nullable lists and nested data.\r\n\r\nBut yup I agree having the `Dataset.from_*` function pass the `schema` to the `pa.Table.from*` would be the way.\r\n\r\nJust one comment about this error: \r\n\r\n```\r\nValueError: TypedSequence is supposed to be used with pa.array(typed_sequence, type=None)\r\n```\r\n\r\nThis happens because `Dataset.from_dict` uses `OptimizedTypedSequence` by default, which should only be used if the user doesn't specify a schema"
] | 2025-03-26T22:16:09Z | 2025-05-15T15:00:59Z | 2025-05-15T15:00:59Z | CONTRIBUTOR | null | null | null | This PR attempts to keep track of non_nullable pyarrow fields when converting a `pa.Schema` to `Features`. At the same time, when outputting the `arrow_schema`, the original non-nullable fields are restored. This allows for more consistent behavior and avoids breaking behavior as illustrated in #7479.
I am by no means a pyarrow expert so some logic in `find_non_nullable_fields` may not perfect. Not sure if more logic (type checks) are needed for deep-checking a given schema. Maybe there are other pyarrow structures that need to be covered?
Tests are added, but again, these may not have sufficient coverage in terms of pyarrow structure types.
closes #7479 | {
"avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4",
"events_url": "https://api.github.com/users/BramVanroy/events{/privacy}",
"followers_url": "https://api.github.com/users/BramVanroy/followers",
"following_url": "https://api.github.com/users/BramVanroy/following{/other_user}",
"gists_url": "https://api.github.com/users/BramVanroy/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/BramVanroy",
"id": 2779410,
"login": "BramVanroy",
"node_id": "MDQ6VXNlcjI3Nzk0MTA=",
"organizations_url": "https://api.github.com/users/BramVanroy/orgs",
"received_events_url": "https://api.github.com/users/BramVanroy/received_events",
"repos_url": "https://api.github.com/users/BramVanroy/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/BramVanroy/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BramVanroy/subscriptions",
"type": "User",
"url": "https://api.github.com/users/BramVanroy",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7482/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7482/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7482.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7482",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7482.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7482"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7481 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7481/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7481/comments | https://api.github.com/repos/huggingface/datasets/issues/7481/events | https://github.com/huggingface/datasets/issues/7481 | 2,950,692,971 | I_kwDODunzps6v4ABr | 7,481 | deal with python `10_000` legal number in slice syntax | {
"avatar_url": "https://avatars.githubusercontent.com/u/196988264?v=4",
"events_url": "https://api.github.com/users/sfc-gh-sbekman/events{/privacy}",
"followers_url": "https://api.github.com/users/sfc-gh-sbekman/followers",
"following_url": "https://api.github.com/users/sfc-gh-sbekman/following{/other_user}",
"gists_url": "https://api.github.com/users/sfc-gh-sbekman/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sfc-gh-sbekman",
"id": 196988264,
"login": "sfc-gh-sbekman",
"node_id": "U_kgDOC73NaA",
"organizations_url": "https://api.github.com/users/sfc-gh-sbekman/orgs",
"received_events_url": "https://api.github.com/users/sfc-gh-sbekman/received_events",
"repos_url": "https://api.github.com/users/sfc-gh-sbekman/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sfc-gh-sbekman/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sfc-gh-sbekman/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sfc-gh-sbekman",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | null | [
"should be an easy fix, I opened a PR"
] | 2025-03-26T20:10:54Z | 2025-03-28T16:20:44Z | 2025-03-28T16:20:44Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
```
In [6]: ds = datasets.load_dataset("HuggingFaceH4/ultrachat_200k", split="train_sft[:1000]")
In [7]: ds = datasets.load_dataset("HuggingFaceH4/ultrachat_200k", split="train_sft[:1_000]")
[dozens of frames skipped]
File /usr/local/lib/python3.10/dist-packages/datasets/arrow_reader.py:444, in _str_to_read_instruction(spec)
442 res = _SUB_SPEC_RE.match(spec)
443 if not res:
--> 444 raise ValueError(f"Unrecognized instruction format: {spec}")
ValueError: Unrecognized instruction format: train_sft[:1_000]
```
It took me a while to understand what the problem was. But apparently `pyarrow` doesn't allow python numbers that may include `_` as in `1_000`. The `_` aids readability since `10_000_000` vs `10000000` is obviously easier to grasp of what the actual number is.
Feature request:
ideally `datasets` being a python module will do the right thing and convert python numbers into whatever pyarrow supports - in this case stripping `_`s.
Second best it'd err and tell the user that using numbers with `_` in split slices is not acceptible, so that the user won't have to deal with a huge pyarrow assert they know nothing about.
Thank you! | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7481/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7481/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7480 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7480/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7480/comments | https://api.github.com/repos/huggingface/datasets/issues/7480/events | https://github.com/huggingface/datasets/issues/7480 | 2,950,315,214 | I_kwDODunzps6v2jzO | 7,480 | HF_DATASETS_CACHE ignored? | {
"avatar_url": "https://avatars.githubusercontent.com/u/31896?v=4",
"events_url": "https://api.github.com/users/stephenroller/events{/privacy}",
"followers_url": "https://api.github.com/users/stephenroller/followers",
"following_url": "https://api.github.com/users/stephenroller/following{/other_user}",
"gists_url": "https://api.github.com/users/stephenroller/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/stephenroller",
"id": 31896,
"login": "stephenroller",
"node_id": "MDQ6VXNlcjMxODk2",
"organizations_url": "https://api.github.com/users/stephenroller/orgs",
"received_events_url": "https://api.github.com/users/stephenroller/received_events",
"repos_url": "https://api.github.com/users/stephenroller/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/stephenroller/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/stephenroller/subscriptions",
"type": "User",
"url": "https://api.github.com/users/stephenroller",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"FWIW, it does eventually write to /tmp/roller/datasets when generating the final version.",
"Hey, Iβd love to work on this issue but I am a beginner, can I work it with you?",
"Hi @lhoestq,\nI'd like to look into this issue but I'm still learning. Could you share any quick pointers on the HF_DATASETS_CACHE behavior here? Thanks!",
"Hi ! `HF_DATASETS_CACHE` is only for the cache files of the `datasets` library, not for the `huggingface_hub` cache for files downloaded from the Hugging Face Hub.\n\nYou should either specify `HF_HOME` (parent cache path for everything HF) or both `HF_DATASETS_CACHE` and `HF_HUB_CACHE`",
"\n\nThanks for clarifying, @lhoestq! To make sure Iβve got it right:\n\n1. **HF_DATASETS_CACHE** only controls where the **datasets** library writes its own cache files (e.g. processed shards, Arrow files, etc.).\n2. Anything downloaded via **huggingface_hub** (models, tokenizers, raw files) still goes into the Hub cache (by default `~/.cache/huggingface/hub`), unless you set **HF_HUB_CACHE** or the parent **HF_HOME**.\n\nSo if you want everything off NFS and onto local disk you have two options:\n\n- **Set both** \n ```bash\n export HF_DATASETS_CACHE=/tmp/roller/datasets \n export HF_HUB_CACHE=/tmp/roller/hub\n ```\n- **Or set** \n ```bash\n export HF_HOME=/tmp/roller\n ```\n which will apply to both subdirectories.\n\nIs that correct? And would it make sense to add a note to the docs clarifying the distinction (or even support S3 for the Hub cache in the future)? Iβm happy to draft a small docs PR if that would help.",
"Hi, yes that's correct, thanks for the clarification ! A note in the docs would be welcome, thanks"
] | 2025-03-26T17:19:34Z | 2025-04-28T10:16:16Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I'm struggling to get things to respect HF_DATASETS_CACHE.
Rationale: I'm on a system that uses NFS for homedir, so downloading to NFS is expensive, slow, and wastes valuable quota compared to local disk. Instead, it seems to rely mostly on HF_HUB_CACHE.
Current version: 3.2.1dev. In the process of testing 3.4.0
### Steps to reproduce the bug
[Currently writing using datasets 3.2.1dev. Will follow up with 3.4.0 results]
dump.py:
```python
from datasets import load_dataset
dataset = load_dataset("HuggingFaceFW/fineweb", name="sample-100BT", split="train")
```
Repro steps
```bash
# ensure no cache
$ mv ~/.cache/huggingface ~/.cache/huggingface.bak
$ export HF_DATASETS_CACHE=/tmp/roller/datasets
$ rm -rf ${HF_DATASETS_CACHE}
$ env | grep HF | grep -v TOKEN
HF_DATASETS_CACHE=/tmp/roller/datasets
$ python dump.py
# (omitted for brevity)
# (while downloading)
$ du -hcs ~/.cache/huggingface/hub
18G hub
18G total
# (after downloading)
$ du -hcs ~/.cache/huggingface/hub
```
It's a shame because datasets supports s3 (which I could really use right now) but hub does not.
### Expected behavior
* ~/.cache/huggingface/hub stays empty
* /tmp/roller/datasets becomes full of stuff
### Environment info
[Currently writing using datasets 3.2.1dev. Will follow up with 3.4.0 results] | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7480/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7480/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7479 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7479/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7479/comments | https://api.github.com/repos/huggingface/datasets/issues/7479/events | https://github.com/huggingface/datasets/issues/7479 | 2,950,235,396 | I_kwDODunzps6v2QUE | 7,479 | Features.from_arrow_schema is destructive | {
"avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4",
"events_url": "https://api.github.com/users/BramVanroy/events{/privacy}",
"followers_url": "https://api.github.com/users/BramVanroy/followers",
"following_url": "https://api.github.com/users/BramVanroy/following{/other_user}",
"gists_url": "https://api.github.com/users/BramVanroy/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/BramVanroy",
"id": 2779410,
"login": "BramVanroy",
"node_id": "MDQ6VXNlcjI3Nzk0MTA=",
"organizations_url": "https://api.github.com/users/BramVanroy/orgs",
"received_events_url": "https://api.github.com/users/BramVanroy/received_events",
"repos_url": "https://api.github.com/users/BramVanroy/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/BramVanroy/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BramVanroy/subscriptions",
"type": "User",
"url": "https://api.github.com/users/BramVanroy",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-03-26T16:46:43Z | 2025-03-26T16:46:58Z | null | CONTRIBUTOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I came across this, perhaps niche, bug where `Features` does not/cannot account for pyarrow's `nullable=False` option in Fields. Interestingly, I found that in regular "flat" fields this does not necessarily lead to conflicts, but when a non-nullable field is in a struct, an incompatibility arises.
It's not easy to explain in words, so the minimal example below should help I hope.
Note that I suggest a solution in the comments in the code, simply allowing `Dataset.to_parquet` to allow for a `schema` argument which, when provided, will override the default ds.features.arrow_schema.
### Steps to reproduce the bug
```python
import os
from datasets import Dataset, Features
import pyarrow as pa
import pyarrow.parquet as pq
# HF datasets is destructive when you call Features.from_arrow_schema(schema) on a schema
# because it will not account for nullable and non-nullable fields in structs (it will always allow nullable)
# Reloading the same dataset with the original schema will raise an error because the schema is not the same anymore
non_nullable_schema = pa.schema(
[
pa.field("text", pa.string(), nullable=False),
pa.field("meta",
pa.struct(
[
pa.field("date", pa.list_(pa.string()), nullable=False),
],
),
),
]
)
print("ORIGINAL SCHEMA")
print(non_nullable_schema)
print()
feats = Features.from_arrow_schema(non_nullable_schema)
print("FEATUR-IZED SCHEMA (nullable-restrictions are gone)")
print(feats.arrow_schema)
print()
ds = Dataset.from_dict(
{
"text": ["a", "b", "c"],
"meta": [{"date": ["2021-01-01"]}, {"date": ["2021-01-02"]}, {"date": ["2021-01-03"]}],
},
features=feats,
)
fname = "tmp.parquet"
# This is not possible: TypeError: pyarrow.parquet.core.ParquetWriter() got multiple values for keyword argument 'schema'
# Though I believe this would be the easiest fix: allow schema to be passed to to_parquet and overwrite the schema in the dataset
# ds.to_parquet(fname, schema=non_nullable_schema)
ds.to_parquet(fname)
try:
_ = pq.read_table(fname, schema=non_nullable_schema)
finally:
os.unlink(fname)
```
### Expected behavior
- Non-destructive behavior when converting an arrow schema to Features; or
- the ability to override the default arrow schema with a custom one
### Environment info
- `datasets` version: 3.2.0
- Platform: Linux-5.14.0-427.20.1.el9_4.x86_64-x86_64-with-glibc2.34
- Python version: 3.11.10
- `huggingface_hub` version: 0.27.1
- PyArrow version: 18.1.0
- Pandas version: 2.2.3
- `fsspec` version: 2024.9.0 | null | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7479/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7479/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7478 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7478/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7478/comments | https://api.github.com/repos/huggingface/datasets/issues/7478/events | https://github.com/huggingface/datasets/pull/7478 | 2,948,993,461 | PR_kwDODunzps6QLPe3 | 7,478 | update fsspec 2025.3.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/487783?v=4",
"events_url": "https://api.github.com/users/peteski22/events{/privacy}",
"followers_url": "https://api.github.com/users/peteski22/followers",
"following_url": "https://api.github.com/users/peteski22/following{/other_user}",
"gists_url": "https://api.github.com/users/peteski22/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/peteski22",
"id": 487783,
"login": "peteski22",
"node_id": "MDQ6VXNlcjQ4Nzc4Mw==",
"organizations_url": "https://api.github.com/users/peteski22/orgs",
"received_events_url": "https://api.github.com/users/peteski22/received_events",
"repos_url": "https://api.github.com/users/peteski22/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/peteski22/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/peteski22/subscriptions",
"type": "User",
"url": "https://api.github.com/users/peteski22",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Sorry for tagging you @lhoestq but since you merged the linked PR, I wondered if you might be able to help me get this triaged so it can be reviewed/rejected etc. ππΌ ",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7478). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-03-26T09:53:05Z | 2025-03-28T19:15:54Z | 2025-03-28T15:51:55Z | CONTRIBUTOR | null | null | null | It appears there have been two releases of fsspec since this dependency was last updated, it would be great if Datasets could be updated so that it didn't hold back the usage of newer fsspec versions in consuming projects.
PR based on https://github.com/huggingface/datasets/pull/7352 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7478/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7478/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7478.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7478",
"merged_at": "2025-03-28T15:51:54Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7478.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7478"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7477 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7477/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7477/comments | https://api.github.com/repos/huggingface/datasets/issues/7477/events | https://github.com/huggingface/datasets/issues/7477 | 2,947,169,460 | I_kwDODunzps6vqjy0 | 7,477 | What is the canonical way to compress a Dataset? | {
"avatar_url": "https://avatars.githubusercontent.com/u/6130352?v=4",
"events_url": "https://api.github.com/users/eric-czech/events{/privacy}",
"followers_url": "https://api.github.com/users/eric-czech/followers",
"following_url": "https://api.github.com/users/eric-czech/following{/other_user}",
"gists_url": "https://api.github.com/users/eric-czech/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/eric-czech",
"id": 6130352,
"login": "eric-czech",
"node_id": "MDQ6VXNlcjYxMzAzNTI=",
"organizations_url": "https://api.github.com/users/eric-czech/orgs",
"received_events_url": "https://api.github.com/users/eric-czech/received_events",
"repos_url": "https://api.github.com/users/eric-czech/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/eric-czech/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/eric-czech/subscriptions",
"type": "User",
"url": "https://api.github.com/users/eric-czech",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"I saw this post by @lhoestq: https://discuss.huggingface.co/t/increased-arrow-table-size-by-factor-of-2/26561/4 suggesting that there is at least some internal code for writing sharded parquet datasets non-concurrently. This appears to be that code: https://github.com/huggingface/datasets/blob/94ccd1b4fada8a92cea96dc8df4e915041d695b6/src/datasets/arrow_dataset.py#L5380-L5397\n\nIs there any fundamental reason (e.g. race conditions) that this kind of operation couldn't exist as a utility or method on a `Dataset` with a `num_proc` argument? I am not seeing any other issues explicitly for that ask. \n",
"We simply haven't implemented a method to save as sharded parquet locally yet ^^'\n\nRight now the only sharded parquet export method is `push_to_hub()` which writes to HF. But we can have a local one as well. \n\nIn the meantime the easiest way to export as sharded parquet locally is to `.shard()` and `.to_parquet()` (see code from my comment [here](https://github.com/huggingface/datasets/issues/7047#issuecomment-2233163406))",
"> In the meantime the easiest way to export as sharded parquet locally is to .shard() and .to_parquet()\n\nMakes sense, BUT how can it be done concurrently? I could of course use multiprocessing myself or a dozen other libraries for parallelizing single-node/local operations like that.\n\nWhat I'm asking though is, what is the way to do this that is most canonical for `datasets` specifically? I.e. what is least likely to causing pickling or other issues because it is used frequently internally by `datasets` and already likely tests for a lot of library-native edge-cases?",
"Everything in `datasets` is picklable :) and even better: since the data are memory mapped from disk, pickling in one process and unpickling in another doesn't do any copy - it instantaneously reloads the memory map.\n\nSo feel free to use the library you prefer to parallelize your operations.\n\n(it's another story in distributed setups though, because in that case you either need to copy and send the data or setup a distributed filesystem)"
] | 2025-03-25T16:47:51Z | 2025-04-03T09:13:11Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | Given that Arrow is the preferred backend for a Dataset, what is a user supposed to do if they want concurrent reads, concurrent writes AND on-disk compression for a larger dataset?
Parquet would be the obvious answer except that there is no native support for writing sharded, parquet datasets concurrently [[1](https://github.com/huggingface/datasets/issues/7047)].
Am I missing something?
And if so, why is this not the standard/default way that `Dataset`'s work as they do in Xarray, Ray Data, Composer, etc.? | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7477/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7477/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7476 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7476/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7476/comments | https://api.github.com/repos/huggingface/datasets/issues/7476/events | https://github.com/huggingface/datasets/pull/7476 | 2,946,997,924 | PR_kwDODunzps6QEbmO | 7,476 | Priotitize json | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7476). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-03-25T15:44:31Z | 2025-03-25T15:47:00Z | 2025-03-25T15:45:00Z | MEMBER | null | null | null | `datasets` should load the JSON data in https://huggingface.co/datasets/facebook/natural_reasoning, not the PDF | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7476/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7476/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7476.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7476",
"merged_at": "2025-03-25T15:45:00Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7476.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7476"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7475 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7475/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7475/comments | https://api.github.com/repos/huggingface/datasets/issues/7475/events | https://github.com/huggingface/datasets/issues/7475 | 2,946,640,570 | I_kwDODunzps6voiq6 | 7,475 | IterableDataset's state_dict shard_example_idx is always equal to the number of samples in a shard | {
"avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4",
"events_url": "https://api.github.com/users/bruno-hays/events{/privacy}",
"followers_url": "https://api.github.com/users/bruno-hays/followers",
"following_url": "https://api.github.com/users/bruno-hays/following{/other_user}",
"gists_url": "https://api.github.com/users/bruno-hays/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/bruno-hays",
"id": 48770768,
"login": "bruno-hays",
"node_id": "MDQ6VXNlcjQ4NzcwNzY4",
"organizations_url": "https://api.github.com/users/bruno-hays/orgs",
"received_events_url": "https://api.github.com/users/bruno-hays/received_events",
"repos_url": "https://api.github.com/users/bruno-hays/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/bruno-hays/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bruno-hays/subscriptions",
"type": "User",
"url": "https://api.github.com/users/bruno-hays",
"user_view_type": "public"
} | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
} | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/129883215?v=4",
"events_url": "https://api.github.com/users/Harry-Yang0518/events{/privacy}",
"followers_url": "https://api.github.com/users/Harry-Yang0518/followers",
"following_url": "https://api.github.com/users/Harry-Yang0518/following{/other_user}",
"gists_url": "https://api.github.com/users/Harry-Yang0518/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Harry-Yang0518",
"id": 129883215,
"login": "Harry-Yang0518",
"node_id": "U_kgDOB73cTw",
"organizations_url": "https://api.github.com/users/Harry-Yang0518/orgs",
"received_events_url": "https://api.github.com/users/Harry-Yang0518/received_events",
"repos_url": "https://api.github.com/users/Harry-Yang0518/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Harry-Yang0518/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Harry-Yang0518/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Harry-Yang0518",
"user_view_type": "public"
}
] | null | [
"Hey, Iβd love to work on this issue but I am a beginner, can I work it with you?",
"Hello. I'm sorry but I don't have much time to get in the details for now.\nHave you managed to reproduce the issue with the code provided ?\nIf you want to work on it, you can self-assign and ask @lhoestq for directions",
"Hi Bruno, I am trying to reproduce it this later in this week and let you know what I found.",
"#self-assign",
"Good catch, I tried and if the dataset is bigger (e.g. `range(9999)`) it returns `\"shard_example_idx\": 1000` with is the `config.DEFAULT_MAX_BATCH_SIZE`\n\nhttps://github.com/huggingface/datasets/blob/94ccd1b4fada8a92cea96dc8df4e915041d695b6/src/datasets/arrow_dataset.py#L5313-L5317\n\nIt looks like the state_dict is incorrect in that case, it should account for this and use the `RebatchedArrowExamplesIterable` which buffers the batch of 1000 rows and counts the iteration within the batch in the state_dict",
"\nHello @lhoestq,\n\nIβve been debugging the `IterableDataset.state_dict()` behavior and applied a patch to `ArrowExamplesIterable._iter_arrow()` in an attempt to fix the issue described in #7475βspecifically, that `shard_example_idx` always equals the number of samples in the shard, even if only a few examples have been consumed.\n\n### What I Tried\n\nI updated `_iter_arrow` to slice off already-consumed rows and increment the state only by the number of actual examples yielded, like this:\n\n```python\nclass ArrowExamplesIterable(_BaseExamplesIterable):\n # ... __init__ and _init_state_dict as before ...\n\n def _iter_arrow(self):\n shard_idx_start = self._state_dict[\"shard_idx\"] if self._state_dict else 0\n\n for gen_kwargs in islice(\n _split_gen_kwargs(self.kwargs, max_num_jobs=self.num_shards),\n shard_idx_start, None\n ):\n shard_example_idx_start = self._state_dict[\"shard_example_idx\"] if self._state_dict else 0\n shard_example_idx = 0\n\n for key, pa_table in self.generate_tables_fn(**gen_kwargs):\n num_rows = len(pa_table)\n next_idx = shard_example_idx + num_rows\n\n if next_idx <= shard_example_idx_start:\n shard_example_idx = next_idx\n continue\n\n offset = max(0, shard_example_idx_start - shard_example_idx)\n sliced_table = pa_table.slice(offset)\n\n if self._state_dict:\n self._state_dict[\"shard_example_idx\"] += len(sliced_table)\n\n yield key, sliced_table\n shard_example_idx = next_idx\n\n if self._state_dict:\n self._state_dict[\"shard_idx\"] += 1\n self._state_dict[\"shard_example_idx\"] = 0\n```\n\nI verified that the updated code was being used, and I added debug prints to confirm the table slicing and counter updates.\n\n### The Issue Still Exists\n\nDespite the changes, the behavior remains the same. Running this minimal repro:\n\n```python\nds = Dataset.from_dict({\"a\": range(6)}).to_iterable_dataset(num_shards=1)\nfor idx, example in enumerate(ds):\n print(example)\n if idx == 2:\n print(\"checkpoint\")\n print(ds.state_dict())\n break\n```\n\nStill outputs:\n\n```bash\n{'a': 0}\n{'a': 1}\n{'a': 2}\ncheckpoint\n{'examples_iterable': {'shard_idx': 0, 'shard_example_idx': 6, 'type': 'ArrowExamplesIterable'}, 'epoch': 0}\n```\n\nEven though only 3 examples were consumed, `shard_example_idx` jumps to 6.\n\n### Questions\n\n- Could there be another place (e.g., in `__iter__`, `RebatchedArrowExamplesIterable`, or the `IterableDataset` wrapper) that's still using the old logic and overriding the state?\n- Is there a better location to intercept and count yielded examples?\n- Would you recommend tracking a new `true_example_idx` to avoid modifying existing behavior?\n\nLet me know your thoughtsβhappy to iterate further and submit a PR once we align on the right approach. Thanks again for your help and feedback!",
"I found a fix using RebatchedArrowExamplesIterable, let me know if it's all good for you now",
"Hi @lhoestq, thanks for the quick fix and for referencing RebatchedArrowExamplesIterable! π\n\nI just tested your patch locally and can confirm that shard_example_idx is now tracking correctly when only a subset of examples is consumed. This resolves the issue I was seeing in #7475.\n\nReally appreciate the guidance earlier on where to lookβit was a great learning opportunity. If there are other parts of the IterableDataset internals that could use cleanup or testing, Iβd be happy to help."
] | 2025-03-25T13:58:07Z | 2025-05-06T14:22:19Z | 2025-05-06T14:05:07Z | CONTRIBUTOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I've noticed a strange behaviour with Iterable state_dict: the value of shard_example_idx is always equal to the amount of samples in a shard.
### Steps to reproduce the bug
I am reusing the example from the doc
```python
from datasets import Dataset
ds = Dataset.from_dict({"a": range(6)}).to_iterable_dataset(num_shards=1)
state_dict = None
# Iterate through the dataset and print examples
for idx, example in enumerate(ds):
print(example)
if idx == 2:
state_dict = ds.state_dict()
print("checkpoint")
break
print(state_dict)
```
Returns:
```
{'a': 0}
{'a': 1}
checkpoint
{'examples_iterable': {'shard_idx': 0, 'shard_example_idx': 6, 'type': 'ArrowExamplesIterable'}, 'epoch': 0}
```
### Expected behavior
shard_example_idx should be 2 instead of 6
If we run with num_shards=2, then shard_example_idx is 3 instead of 2 and so on.
### Environment info
- `datasets` version: 3.4.1
- Platform: macOS-14.6.1-arm64-arm-64bit
- Python version: 3.12.9
- `huggingface_hub` version: 0.29.3
- PyArrow version: 19.0.1
- Pandas version: 2.2.3
- `fsspec` version: 2024.12.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7475/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7475/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7474 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7474/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7474/comments | https://api.github.com/repos/huggingface/datasets/issues/7474/events | https://github.com/huggingface/datasets/pull/7474 | 2,945,066,258 | PR_kwDODunzps6P91lM | 7,474 | Remove conditions for Python < 3.9 | {
"avatar_url": "https://avatars.githubusercontent.com/u/17618148?v=4",
"events_url": "https://api.github.com/users/cyyever/events{/privacy}",
"followers_url": "https://api.github.com/users/cyyever/followers",
"following_url": "https://api.github.com/users/cyyever/following{/other_user}",
"gists_url": "https://api.github.com/users/cyyever/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/cyyever",
"id": 17618148,
"login": "cyyever",
"node_id": "MDQ6VXNlcjE3NjE4MTQ4",
"organizations_url": "https://api.github.com/users/cyyever/orgs",
"received_events_url": "https://api.github.com/users/cyyever/received_events",
"repos_url": "https://api.github.com/users/cyyever/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/cyyever/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/cyyever/subscriptions",
"type": "User",
"url": "https://api.github.com/users/cyyever",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7474). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Thanks ! can you run `make style` to fix code formatting ? then we can merge",
"@lhoestq Done"
] | 2025-03-25T03:08:04Z | 2025-04-16T00:11:06Z | 2025-04-15T16:07:55Z | CONTRIBUTOR | null | null | null | This PR remove conditions for Python < 3.9. | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7474/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7474/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7474.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7474",
"merged_at": "2025-04-15T16:07:54Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7474.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7474"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7473 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7473/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7473/comments | https://api.github.com/repos/huggingface/datasets/issues/7473/events | https://github.com/huggingface/datasets/issues/7473 | 2,939,034,643 | I_kwDODunzps6vLhwT | 7,473 | Webdataset data format problem | {
"avatar_url": "https://avatars.githubusercontent.com/u/1017189?v=4",
"events_url": "https://api.github.com/users/edmcman/events{/privacy}",
"followers_url": "https://api.github.com/users/edmcman/followers",
"following_url": "https://api.github.com/users/edmcman/following{/other_user}",
"gists_url": "https://api.github.com/users/edmcman/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/edmcman",
"id": 1017189,
"login": "edmcman",
"node_id": "MDQ6VXNlcjEwMTcxODk=",
"organizations_url": "https://api.github.com/users/edmcman/orgs",
"received_events_url": "https://api.github.com/users/edmcman/received_events",
"repos_url": "https://api.github.com/users/edmcman/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/edmcman/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/edmcman/subscriptions",
"type": "User",
"url": "https://api.github.com/users/edmcman",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"I was able to work around it"
] | 2025-03-21T17:23:52Z | 2025-03-21T19:19:58Z | 2025-03-21T19:19:58Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Please see https://huggingface.co/datasets/ejschwartz/idioms/discussions/1
Error code: FileFormatMismatchBetweenSplitsError
All three splits, train, test, and validation, use webdataset. But only the train split has more than one file. How can I force the other two splits to also be interpreted as being the webdataset format? (I don't think there is currently a way, but happy to be told that I am wrong.)
### Steps to reproduce the bug
```
import datasets
datasets.load_dataset("ejschwartz/idioms")
### Expected behavior
The dataset loads. Alternatively, there is a YAML syntax for manually specifying the format.
### Environment info
- `datasets` version: 3.2.0
- Platform: Linux-6.8.0-52-generic-x86_64-with-glibc2.35
- Python version: 3.10.12
- `huggingface_hub` version: 0.28.1
- PyArrow version: 19.0.0
- Pandas version: 2.2.3
- `fsspec` version: 2024.9.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/1017189?v=4",
"events_url": "https://api.github.com/users/edmcman/events{/privacy}",
"followers_url": "https://api.github.com/users/edmcman/followers",
"following_url": "https://api.github.com/users/edmcman/following{/other_user}",
"gists_url": "https://api.github.com/users/edmcman/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/edmcman",
"id": 1017189,
"login": "edmcman",
"node_id": "MDQ6VXNlcjEwMTcxODk=",
"organizations_url": "https://api.github.com/users/edmcman/orgs",
"received_events_url": "https://api.github.com/users/edmcman/received_events",
"repos_url": "https://api.github.com/users/edmcman/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/edmcman/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/edmcman/subscriptions",
"type": "User",
"url": "https://api.github.com/users/edmcman",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7473/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7473/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7472 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7472/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7472/comments | https://api.github.com/repos/huggingface/datasets/issues/7472/events | https://github.com/huggingface/datasets/issues/7472 | 2,937,607,272 | I_kwDODunzps6vGFRo | 7,472 | Label casting during `map` process is canceled after the `map` process | {
"avatar_url": "https://avatars.githubusercontent.com/u/11156001?v=4",
"events_url": "https://api.github.com/users/yoshitomo-matsubara/events{/privacy}",
"followers_url": "https://api.github.com/users/yoshitomo-matsubara/followers",
"following_url": "https://api.github.com/users/yoshitomo-matsubara/following{/other_user}",
"gists_url": "https://api.github.com/users/yoshitomo-matsubara/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yoshitomo-matsubara",
"id": 11156001,
"login": "yoshitomo-matsubara",
"node_id": "MDQ6VXNlcjExMTU2MDAx",
"organizations_url": "https://api.github.com/users/yoshitomo-matsubara/orgs",
"received_events_url": "https://api.github.com/users/yoshitomo-matsubara/received_events",
"repos_url": "https://api.github.com/users/yoshitomo-matsubara/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yoshitomo-matsubara/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yoshitomo-matsubara/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yoshitomo-matsubara",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! By default `map()` tries to keep the types of each column of the dataset, so here it reuses the int type since all your float values can be converted to integers. But I agree it would be nice to store float values as float values and don't try to reuse the same type in this case.\n\nIn the meantime, you can either store the float values in a new column, or pass the output `features=` manually to `map()`",
"Hi @lhoestq \n\nThank you for the answer & suggestion!\n\nCan we add some flag to `map()` function like `reuses_original_type=True` and skip reusing the original type when it's False?\n\nLet me know if it sounds like a reasonable solution. I am happy to submit a PR for this.",
"In general we try to avoid adding new parameters when it's already possible to achieve the same results with existing parameters (here `features=`). But since it's not always convenient to know in advance the `features=` I'm open to contributions to adding this parameter yes",
"Thank you for sharing the context. Good to know that. \n\nI submitted a PR #7483. Could you review the PR?",
"Hi @lhoestq \n\nLet me know if there is something that I should add to [the PR](https://github.com/huggingface/datasets/pull/7483)!",
"Closing this issue as the PR #7483 was merged"
] | 2025-03-21T07:56:22Z | 2025-04-10T05:11:15Z | 2025-04-10T05:11:14Z | CONTRIBUTOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When preprocessing a multi-label dataset, I introduced a step to convert int labels to float labels as [BCEWithLogitsLoss](https://pytorch.org/docs/stable/generated/torch.nn.BCEWithLogitsLoss.html) expects float labels and forward function of models in transformers package internally use `BCEWithLogitsLoss`
However, the casting was canceled after `.map` process and the label values still use int values, which leads to an error
```
File "/home/yoshitomo/anaconda3/envs/torchdistill/lib/python3.10/site-packages/transformers/models/bert/modeling_bert.py", line 1711, in forward
loss = loss_fct(logits, labels)
File "/home/yoshitomo/anaconda3/envs/torchdistill/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/yoshitomo/anaconda3/envs/torchdistill/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/home/yoshitomo/anaconda3/envs/torchdistill/lib/python3.10/site-packages/torch/nn/modules/loss.py", line 819, in forward
return F.binary_cross_entropy_with_logits(
File "/home/yoshitomo/anaconda3/envs/torchdistill/lib/python3.10/site-packages/torch/nn/functional.py", line 3628, in binary_cross_entropy_with_logits
return torch.binary_cross_entropy_with_logits(
RuntimeError: result type Float can't be cast to the desired output type Long
```
This seems like happening only when the original labels are int values (see examples below)
### Steps to reproduce the bug
If the original dataset uses a list of int labels, it will cancel the int->float casting
```python
from datasets import Dataset
data = {
'text': ['text1', 'text2', 'text3', 'text4'],
'labels': [[0, 1, 2], [3], [3, 4], [3]]
}
dataset = Dataset.from_dict(data)
label_set = set([label for labels in data['labels'] for label in labels])
label2idx = {label: idx for idx, label in enumerate(sorted(label_set))}
def multi_labels_to_ids(labels):
ids = [0.0] * len(label2idx)
for label in labels:
ids[label2idx[label]] = 1.0
return ids
def preprocess(examples):
result = {'sentence': [[0, 3, 4] for _ in range(len(examples['labels']))]}
print('"labels" are int', examples['labels'])
result['labels'] = [multi_labels_to_ids(l) for l in examples['labels']]
print('"labels" were converted to multi-label format with float values', result['labels'])
return result
preprocessed_dataset = dataset.map(preprocess, batched=True, remove_columns=['labels', 'text'])
print(preprocessed_dataset[0]['labels'])
# Output: "[1, 1, 1, 0, 0]"
# Expected: "[1.0, 1.0, 1.0, 0.0, 0.0]"
```
If the original dataset uses non-int labels, it works as expected.
```python
from datasets import Dataset
data = {
'text': ['text1', 'text2', 'text3', 'text4'],
'labels': [['label1', 'label2', 'label3'], ['label4'], ['label4', 'label5'], ['label4']]
}
dataset = Dataset.from_dict(data)
label_set = set([label for labels in data['labels'] for label in labels])
label2idx = {label: idx for idx, label in enumerate(sorted(label_set))}
def multi_labels_to_ids(labels):
ids = [0.0] * len(label2idx)
for label in labels:
ids[label2idx[label]] = 1.0
return ids
def preprocess(examples):
result = {'sentence': [[0, 3, 4] for _ in range(len(examples['labels']))]}
print('"labels" are int', examples['labels'])
result['labels'] = [multi_labels_to_ids(l) for l in examples['labels']]
print('"labels" were converted to multi-label format with float values', result['labels'])
return result
preprocessed_dataset = dataset.map(preprocess, batched=True, remove_columns=['labels', 'text'])
print(preprocessed_dataset[0]['labels'])
# Output: "[1.0, 1.0, 1.0, 0.0, 0.0]"
# Expected: "[1.0, 1.0, 1.0, 0.0, 0.0]"
```
Note that the only difference between these two examples is
> 'labels': [[0, 1, 2], [3], [3, 4], [3]]
v.s
> 'labels': [['label1', 'label2', 'label3'], ['label4'], ['label4', 'label5'], ['label4']]
### Expected behavior
Even if the original dataset uses a list of int labels, the int->float casting during `.map` process should not be canceled as shown in the above example
### Environment info
OS Ubuntu 22.04 LTS
Python 3.10.11
datasets v3.4.1 | {
"avatar_url": "https://avatars.githubusercontent.com/u/11156001?v=4",
"events_url": "https://api.github.com/users/yoshitomo-matsubara/events{/privacy}",
"followers_url": "https://api.github.com/users/yoshitomo-matsubara/followers",
"following_url": "https://api.github.com/users/yoshitomo-matsubara/following{/other_user}",
"gists_url": "https://api.github.com/users/yoshitomo-matsubara/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yoshitomo-matsubara",
"id": 11156001,
"login": "yoshitomo-matsubara",
"node_id": "MDQ6VXNlcjExMTU2MDAx",
"organizations_url": "https://api.github.com/users/yoshitomo-matsubara/orgs",
"received_events_url": "https://api.github.com/users/yoshitomo-matsubara/received_events",
"repos_url": "https://api.github.com/users/yoshitomo-matsubara/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yoshitomo-matsubara/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yoshitomo-matsubara/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yoshitomo-matsubara",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7472/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7472/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7471 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7471/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7471/comments | https://api.github.com/repos/huggingface/datasets/issues/7471/events | https://github.com/huggingface/datasets/issues/7471 | 2,937,530,069 | I_kwDODunzps6vFybV | 7,471 | Adding argument to `_get_data_files_patterns` | {
"avatar_url": "https://avatars.githubusercontent.com/u/34004152?v=4",
"events_url": "https://api.github.com/users/SangbumChoi/events{/privacy}",
"followers_url": "https://api.github.com/users/SangbumChoi/followers",
"following_url": "https://api.github.com/users/SangbumChoi/following{/other_user}",
"gists_url": "https://api.github.com/users/SangbumChoi/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/SangbumChoi",
"id": 34004152,
"login": "SangbumChoi",
"node_id": "MDQ6VXNlcjM0MDA0MTUy",
"organizations_url": "https://api.github.com/users/SangbumChoi/orgs",
"received_events_url": "https://api.github.com/users/SangbumChoi/received_events",
"repos_url": "https://api.github.com/users/SangbumChoi/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/SangbumChoi/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/SangbumChoi/subscriptions",
"type": "User",
"url": "https://api.github.com/users/SangbumChoi",
"user_view_type": "public"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | null | [
"Hi ! The pattern can be specified in advance in YAML in the README.md of the dataset :)\n\nFor example\n\n```\n---\nconfigs:\n- config_name: default\n data_files:\n - split: train\n path: \"train/*\"\n - split: test\n path: \"test/*\"\n---\n```\n\nSee the docs at https://huggingface.co/docs/hub/en/datasets-manual-configuration",
"@lhoestq How can we choose in this case ? https://huggingface.co/datasets/datasets-examples/doc-image-5\n",
"choose what ? sorry I didn't get it ^^'"
] | 2025-03-21T07:17:53Z | 2025-03-27T12:30:52Z | 2025-03-26T07:26:27Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Feature request
How about adding if the user already know about the pattern?
https://github.com/huggingface/datasets/blob/a256b85cbc67aa3f0e75d32d6586afc507cf535b/src/datasets/data_files.py#L252
### Motivation
While using this load_dataset people might use 10M of images for the local files.
However, due to searching all the appropriate file pattern in fsspec, purely searching this pattern takes more than 10 hours (real use-case).
### Your contribution
Yeah I can make this happen if this seems valid. @lhoestq WDYT?
such like
```
def _get_data_files_patterns(pattern_resolver: Callable[[str], list[str]], patterns: PATTERNS) -> dict[str, list[str]]:
``` | {
"avatar_url": "https://avatars.githubusercontent.com/u/34004152?v=4",
"events_url": "https://api.github.com/users/SangbumChoi/events{/privacy}",
"followers_url": "https://api.github.com/users/SangbumChoi/followers",
"following_url": "https://api.github.com/users/SangbumChoi/following{/other_user}",
"gists_url": "https://api.github.com/users/SangbumChoi/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/SangbumChoi",
"id": 34004152,
"login": "SangbumChoi",
"node_id": "MDQ6VXNlcjM0MDA0MTUy",
"organizations_url": "https://api.github.com/users/SangbumChoi/orgs",
"received_events_url": "https://api.github.com/users/SangbumChoi/received_events",
"repos_url": "https://api.github.com/users/SangbumChoi/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/SangbumChoi/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/SangbumChoi/subscriptions",
"type": "User",
"url": "https://api.github.com/users/SangbumChoi",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7471/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7471/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7470 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7470/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7470/comments | https://api.github.com/repos/huggingface/datasets/issues/7470/events | https://github.com/huggingface/datasets/issues/7470 | 2,937,236,323 | I_kwDODunzps6vEqtj | 7,470 | Is it possible to shard a single-sharded IterableDataset? | {
"avatar_url": "https://avatars.githubusercontent.com/u/511073?v=4",
"events_url": "https://api.github.com/users/jonathanasdf/events{/privacy}",
"followers_url": "https://api.github.com/users/jonathanasdf/followers",
"following_url": "https://api.github.com/users/jonathanasdf/following{/other_user}",
"gists_url": "https://api.github.com/users/jonathanasdf/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jonathanasdf",
"id": 511073,
"login": "jonathanasdf",
"node_id": "MDQ6VXNlcjUxMTA3Mw==",
"organizations_url": "https://api.github.com/users/jonathanasdf/orgs",
"received_events_url": "https://api.github.com/users/jonathanasdf/received_events",
"repos_url": "https://api.github.com/users/jonathanasdf/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jonathanasdf/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jonathanasdf/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jonathanasdf",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! Maybe you can look for an option in your dataset to partition your data based on a deterministic filter ? For example each worker could stream the data based on `row.id % num_shards` or something like that ?",
"So the recommendation is to start out with multiple shards initially and re-sharding after is not expected to work? :(\n\nWould something like the following work? Some DiskCachingIterableDataset, where worker 0 streams from the datasource, but also writes to disk, and all of the other workers read from what worker 0 wrote? Then that would produce a stream with a deterministic order and we can subsample.",
"To be honest it would be cool to support native multiprocessing in `IterableDataset.map` so you can parallelize any specific processing step without having to rely on a torch Dataloader. What do you think ?\n\nrelated: https://github.com/huggingface/datasets/issues/7193 https://github.com/huggingface/datasets/issues/3444 \noriginal issue: https://github.com/huggingface/datasets/issues/2642\n\nAlternatively the DiskCachingIterableDataset idea works, just note that to make it work with a torch Dataloader with num_workers>0 you'll need:\n1. to make your own `torch.utils.data.IterableDataset` and have rank=0 stream the data and share them with the other workers (either via disk as suggested or IPC)\n2. take into account that`datasets.IterableDataset` will yield 0 examples for ranks with id>0 if there is only one shard, but in your case it's ok since you'd only stream from rank=0",
"Ohh that would be pretty cool!\n\nThanks for the suggestions, as there's no actionable items for this repo I'm going to close this issue now.",
"Another usecase for this resharding:\n\nIf we have a bunch of jsonl files, and we load it as an IterableDataset with multiple dataloader workers, each file gets naively assigned to a worker.\n\nIf the files were not carefully produced to be equally sized, eg if the very last file is significantly shorter, containing just a few examples, and it gets assigned onto a dataloader worker by itself, then the examples in that file will be significantly oversampled.\n\nIt would be nice if datasets had an internal way to rebalance this without requiring offline reprocessing of the data files"
] | 2025-03-21T04:33:37Z | 2025-05-09T22:51:46Z | 2025-03-26T06:49:28Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | I thought https://github.com/huggingface/datasets/pull/7252 might be applicable but looking at it maybe not.
Say we have a process, eg. a database query, that can return data in slightly different order each time. So, the initial query needs to be run by a single thread (not to mention running multiple times incurs more cost too). But the results are also big enough that we don't want to materialize it entirely and instead stream it with an IterableDataset.
But after we have the results we want to split it up across workers to parallelize processing.
Is something like this possible to do?
Here's a failed attempt. The end result should be that each of the shards has unique data, but unfortunately with this attempt the generator gets run once in each shard and the results end up with duplicates...
```
import random
import datasets
def gen():
print('RUNNING GENERATOR!')
items = list(range(10))
random.shuffle(items)
yield from items
ds = datasets.IterableDataset.from_generator(gen)
print('dataset contents:')
for item in ds:
print(item)
print()
print('dataset contents (2):')
for item in ds:
print(item)
print()
num_shards = 3
def sharded(shard_id):
for i, example in enumerate(ds):
if i % num_shards in shard_id:
yield example
ds1 = datasets.IterableDataset.from_generator(
sharded, gen_kwargs={'shard_id': list(range(num_shards))}
)
for shard in range(num_shards):
print('shard', shard)
for item in ds1.shard(num_shards, shard):
print(item)
``` | {
"avatar_url": "https://avatars.githubusercontent.com/u/511073?v=4",
"events_url": "https://api.github.com/users/jonathanasdf/events{/privacy}",
"followers_url": "https://api.github.com/users/jonathanasdf/followers",
"following_url": "https://api.github.com/users/jonathanasdf/following{/other_user}",
"gists_url": "https://api.github.com/users/jonathanasdf/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jonathanasdf",
"id": 511073,
"login": "jonathanasdf",
"node_id": "MDQ6VXNlcjUxMTA3Mw==",
"organizations_url": "https://api.github.com/users/jonathanasdf/orgs",
"received_events_url": "https://api.github.com/users/jonathanasdf/received_events",
"repos_url": "https://api.github.com/users/jonathanasdf/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jonathanasdf/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jonathanasdf/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jonathanasdf",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7470/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7470/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7469 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7469/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7469/comments | https://api.github.com/repos/huggingface/datasets/issues/7469/events | https://github.com/huggingface/datasets/issues/7469 | 2,936,606,080 | I_kwDODunzps6vCQ2A | 7,469 | Custom split name with the web interface | {
"avatar_url": "https://avatars.githubusercontent.com/u/15141326?v=4",
"events_url": "https://api.github.com/users/vince62s/events{/privacy}",
"followers_url": "https://api.github.com/users/vince62s/followers",
"following_url": "https://api.github.com/users/vince62s/following{/other_user}",
"gists_url": "https://api.github.com/users/vince62s/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/vince62s",
"id": 15141326,
"login": "vince62s",
"node_id": "MDQ6VXNlcjE1MTQxMzI2",
"organizations_url": "https://api.github.com/users/vince62s/orgs",
"received_events_url": "https://api.github.com/users/vince62s/received_events",
"repos_url": "https://api.github.com/users/vince62s/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/vince62s/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/vince62s/subscriptions",
"type": "User",
"url": "https://api.github.com/users/vince62s",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [] | 2025-03-20T20:45:59Z | 2025-03-21T07:20:37Z | 2025-03-21T07:20:37Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
According the doc here: https://huggingface.co/docs/hub/datasets-file-names-and-splits#custom-split-name
it should infer the split name from the subdir of data or the beg of the name of the files in data.
When doing this manually through web upload it does not work. it uses "train" as a unique split.
example: https://huggingface.co/datasets/eole-nlp/estimator_chatml
### Steps to reproduce the bug
follow the link above
### Expected behavior
there should be two splits "mlqe" and "1720_da"
### Environment info
website | {
"avatar_url": "https://avatars.githubusercontent.com/u/15141326?v=4",
"events_url": "https://api.github.com/users/vince62s/events{/privacy}",
"followers_url": "https://api.github.com/users/vince62s/followers",
"following_url": "https://api.github.com/users/vince62s/following{/other_user}",
"gists_url": "https://api.github.com/users/vince62s/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/vince62s",
"id": 15141326,
"login": "vince62s",
"node_id": "MDQ6VXNlcjE1MTQxMzI2",
"organizations_url": "https://api.github.com/users/vince62s/orgs",
"received_events_url": "https://api.github.com/users/vince62s/received_events",
"repos_url": "https://api.github.com/users/vince62s/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/vince62s/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/vince62s/subscriptions",
"type": "User",
"url": "https://api.github.com/users/vince62s",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7469/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7469/timeline | null | completed | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7468 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7468/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7468/comments | https://api.github.com/repos/huggingface/datasets/issues/7468/events | https://github.com/huggingface/datasets/issues/7468 | 2,934,094,103 | I_kwDODunzps6u4rkX | 7,468 | function `load_dataset` can't solve folder path with regex characters like "[]" | {
"avatar_url": "https://avatars.githubusercontent.com/u/89294013?v=4",
"events_url": "https://api.github.com/users/Hpeox/events{/privacy}",
"followers_url": "https://api.github.com/users/Hpeox/followers",
"following_url": "https://api.github.com/users/Hpeox/following{/other_user}",
"gists_url": "https://api.github.com/users/Hpeox/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Hpeox",
"id": 89294013,
"login": "Hpeox",
"node_id": "MDQ6VXNlcjg5Mjk0MDEz",
"organizations_url": "https://api.github.com/users/Hpeox/orgs",
"received_events_url": "https://api.github.com/users/Hpeox/received_events",
"repos_url": "https://api.github.com/users/Hpeox/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Hpeox/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Hpeox/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Hpeox",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! Have you tried escaping the glob special characters `[` and `]` ?\n\nbtw note that`AbstractFileSystem.glob` doesn't support regex, instead it supports glob patterns as in the python library [glob](https://docs.python.org/3/library/glob.html)\n"
] | 2025-03-20T05:21:59Z | 2025-03-25T10:18:12Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When using the `load_dataset` function with a folder path containing regex special characters (such as "[]"), the issue occurs due to how the path is handled in the `resolve_pattern` function. This function passes the unprocessed path directly to `AbstractFileSystem.glob`, which supports regular expressions. As a result, the globbing mechanism interprets these characters as regex patterns, leading to a traversal of the entire disk partition instead of confining the search to the intended directory.
### Steps to reproduce the bug
just create a folder like `E:\[D_DATA]\koch_test`, then `load_dataset("parquet", data_dir="E:\[D_DATA]\\test", split="train")`
it will keep searching the whole disk.
I add two `print` in `glob` and `resolve_pattern` to see the path
### Expected behavior
it should load the dataset as in normal folders
### Environment info
- `datasets` version: 3.3.2
- Platform: Windows-10-10.0.22631-SP0
- Python version: 3.10.16
- `huggingface_hub` version: 0.29.1
- PyArrow version: 19.0.1
- Pandas version: 2.2.3
- `fsspec` version: 2024.12.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7468/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7468/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7467 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7467/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7467/comments | https://api.github.com/repos/huggingface/datasets/issues/7467/events | https://github.com/huggingface/datasets/issues/7467 | 2,930,067,107 | I_kwDODunzps6upUaj | 7,467 | load_dataset with streaming hangs on parquet datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/10550252?v=4",
"events_url": "https://api.github.com/users/The0nix/events{/privacy}",
"followers_url": "https://api.github.com/users/The0nix/followers",
"following_url": "https://api.github.com/users/The0nix/following{/other_user}",
"gists_url": "https://api.github.com/users/The0nix/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/The0nix",
"id": 10550252,
"login": "The0nix",
"node_id": "MDQ6VXNlcjEwNTUwMjUy",
"organizations_url": "https://api.github.com/users/The0nix/orgs",
"received_events_url": "https://api.github.com/users/The0nix/received_events",
"repos_url": "https://api.github.com/users/The0nix/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/The0nix/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/The0nix/subscriptions",
"type": "User",
"url": "https://api.github.com/users/The0nix",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! The issue comes from `pyarrow`, I reported it here: https://github.com/apache/arrow/issues/45214 (feel free to comment / thumb up).\n\nAlternatively we can try to find something else than `ParquetFileFragment.to_batches()` to iterate on Parquet data and keep the option the pass `filters=`..."
] | 2025-03-18T23:33:54Z | 2025-03-25T10:28:04Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When I try to load a dataset with parquet files (e.g. "bigcode/the-stack") the dataset loads, but python interpreter can't exit and hangs
### Steps to reproduce the bug
```python3
import datasets
print('Start')
dataset = datasets.load_dataset("bigcode/the-stack", data_dir="data/yaml", streaming=True, split="train")
it = iter(dataset)
next(it)
print('Finish')
```
The program prints finish but doesn't exit and hangs indefinitely.
I tried this on two different machines and several datasets.
### Expected behavior
The program exits successfully
### Environment info
datasets==3.4.1
Python 3.12.9.
MacOS and Ubuntu Linux | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7467/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7467/timeline | null | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7466 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7466/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7466/comments | https://api.github.com/repos/huggingface/datasets/issues/7466/events | https://github.com/huggingface/datasets/pull/7466 | 2,928,661,327 | PR_kwDODunzps6PHQyp | 7,466 | Fix local pdf loading | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7466). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-03-18T14:09:06Z | 2025-03-18T14:11:52Z | 2025-03-18T14:09:21Z | MEMBER | null | null | null | fir this error when accessing a local pdf
```
File ~/.pyenv/versions/3.12.2/envs/hf-datasets/lib/python3.12/site-packages/pdfminer/psparser.py:220, in PSBaseParser.seek(self, pos)
218 """Seeks the parser to the given position."""
219 log.debug("seek: %r", pos)
--> 220 self.fp.seek(pos)
221 # reset the status for nextline()
222 self.bufpos = pos
ValueError: seek of closed file
``` | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7466/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7466/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7466.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7466",
"merged_at": "2025-03-18T14:09:21Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7466.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7466"
} | true |
https://api.github.com/repos/huggingface/datasets/issues/7464 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7464/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7464/comments | https://api.github.com/repos/huggingface/datasets/issues/7464/events | https://github.com/huggingface/datasets/pull/7464 | 2,926,478,838 | PR_kwDODunzps6PABJa | 7,464 | Minor fix for metadata files in extension counter | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7464). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-03-17T21:57:11Z | 2025-03-18T15:21:43Z | 2025-03-18T15:21:41Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7464/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7464/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7464.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7464",
"merged_at": "2025-03-18T15:21:41Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7464.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7464"
} | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.