url
stringlengths
61
61
repository_url
stringclasses
1 value
labels_url
stringlengths
75
75
comments_url
stringlengths
70
70
events_url
stringlengths
68
68
html_url
stringlengths
51
51
id
int64
1.92B
2.7B
node_id
stringlengths
18
18
number
int64
6.27k
7.3k
title
stringlengths
2
150
user
dict
labels
listlengths
0
2
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
1
milestone
null
comments
sequencelengths
0
23
created_at
timestamp[ns]
updated_at
int64
1.7k
1.73k
closed_at
timestamp[ns]
author_association
stringclasses
4 values
active_lock_reason
null
body
stringlengths
3
47.9k
closed_by
dict
reactions
dict
timeline_url
stringlengths
70
70
performed_via_github_app
null
state_reason
stringclasses
3 values
draft
null
pull_request
null
is_pull_request
bool
1 class
time_to_close
float64
0
0
https://api.github.com/repos/huggingface/datasets/issues/6328
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6328/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6328/comments
https://api.github.com/repos/huggingface/datasets/issues/6328/events
https://github.com/huggingface/datasets/issues/6328
1,955,857,904
I_kwDODunzps50lAXw
6,328
شبکه های متن به گفتار ابتدا متن داده شده را به بازنمایی میانی
{ "avatar_url": "https://avatars.githubusercontent.com/u/147399213?v=4", "events_url": "https://api.github.com/users/shabnam706/events{/privacy}", "followers_url": "https://api.github.com/users/shabnam706/followers", "following_url": "https://api.github.com/users/shabnam706/following{/other_user}", "gists_url": "https://api.github.com/users/shabnam706/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shabnam706", "id": 147399213, "login": "shabnam706", "node_id": "U_kgDOCMkiLQ", "organizations_url": "https://api.github.com/users/shabnam706/orgs", "received_events_url": "https://api.github.com/users/shabnam706/received_events", "repos_url": "https://api.github.com/users/shabnam706/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shabnam706/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shabnam706/subscriptions", "type": "User", "url": "https://api.github.com/users/shabnam706", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "شبکه های متن به گفتار ابتدا متن داده شده را به بازنمایی میانی" ]
1970-01-01T00:00:00.000001
1,698
1970-01-01T00:00:00.000001
NONE
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6328/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6328/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6327
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6327/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6327/comments
https://api.github.com/repos/huggingface/datasets/issues/6327/events
https://github.com/huggingface/datasets/issues/6327
1,955,470,755
I_kwDODunzps50jh2j
6,327
FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)`
{ "avatar_url": "https://avatars.githubusercontent.com/u/18402347?v=4", "events_url": "https://api.github.com/users/yzhangcs/events{/privacy}", "followers_url": "https://api.github.com/users/yzhangcs/followers", "following_url": "https://api.github.com/users/yzhangcs/following{/other_user}", "gists_url": "https://api.github.com/users/yzhangcs/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yzhangcs", "id": 18402347, "login": "yzhangcs", "node_id": "MDQ6VXNlcjE4NDAyMzQ3", "organizations_url": "https://api.github.com/users/yzhangcs/orgs", "received_events_url": "https://api.github.com/users/yzhangcs/received_events", "repos_url": "https://api.github.com/users/yzhangcs/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yzhangcs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yzhangcs/subscriptions", "type": "User", "url": "https://api.github.com/users/yzhangcs", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "You can clone the `togethercomputer/RedPajama-Data-1T-Sample` repo and load the dataset with `load_dataset(\"path/to/cloned_repo\")` to use it offline.", "@mariosasko Thank you for your kind reply! I'll try it as a workaround.\r\nDoes that mean that currently it's not supported to simply load with a short name?", "It is, but manually downloading repo files to the cache can easily lead to failure (the HF cache is not meant to be modified by a user besides deleting the files 🙂), as in your case. Hence, the clone + `load_dataset(\"path/to/cloned_repo\")` workflow should be used instead." ]
1970-01-01T00:00:00.000001
1,698
1970-01-01T00:00:00.000001
CONTRIBUTOR
null
### Describe the bug Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs. ### Steps to reproduce the bug I've downloaded the dataset and save it to the cache dir in advance. My hope is loading the files in offline environment and without taking too much hours to prepross the entire data before running into the training process. So I try the following code to load the files streamingly ```py dataset = load_dataset('togethercomputer/RedPajama-Data-1T-Sample', streaming=True) print(next(iter(dataset['train']))) ``` Sadly, it raises the following: ``` FileNotFoundError: [Errno 2] No such file or directory: 'CURRENT_CODE_PATH/arxiv_sample.jsonl' ``` I've noticed that the dataset can be properly found in the begining ``` Using the latest cached version of the module from /root/.cache/huggingface/modules/datasets_modules/datasets/togethercomputer--RedPajama-Data-1T-Sample/6ea3bc8ec2e84ec6d2df1930942e9028ace8c5b9d9143823cf911c50bbd92039 (last modified on Sat Oct 21 20:12:57 2023) since it couldn't be found locally at togethercomputer/RedPajama-Data-1T-Sample., or remotely on the Hugging Face Hub. ``` But it seems that the paths couldn't be properly parsed when loading iteratively. How should I fix this error. I've tried specifying `data_files` or `data_dir` as `.../arxiv_sample.jsonl` but none of them works. Thanks. ### Expected behavior Properly load the dataset. ### Environment info `datasets==2.14.5`
{ "avatar_url": "https://avatars.githubusercontent.com/u/18402347?v=4", "events_url": "https://api.github.com/users/yzhangcs/events{/privacy}", "followers_url": "https://api.github.com/users/yzhangcs/followers", "following_url": "https://api.github.com/users/yzhangcs/following{/other_user}", "gists_url": "https://api.github.com/users/yzhangcs/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yzhangcs", "id": 18402347, "login": "yzhangcs", "node_id": "MDQ6VXNlcjE4NDAyMzQ3", "organizations_url": "https://api.github.com/users/yzhangcs/orgs", "received_events_url": "https://api.github.com/users/yzhangcs/received_events", "repos_url": "https://api.github.com/users/yzhangcs/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yzhangcs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yzhangcs/subscriptions", "type": "User", "url": "https://api.github.com/users/yzhangcs", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6327/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6327/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6324
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6324/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6324/comments
https://api.github.com/repos/huggingface/datasets/issues/6324/events
https://github.com/huggingface/datasets/issues/6324
1,955,126,687
I_kwDODunzps50iN2f
6,324
Conversion to Arrow fails due to wrong type heuristic
{ "avatar_url": "https://avatars.githubusercontent.com/u/2862336?v=4", "events_url": "https://api.github.com/users/jphme/events{/privacy}", "followers_url": "https://api.github.com/users/jphme/followers", "following_url": "https://api.github.com/users/jphme/following{/other_user}", "gists_url": "https://api.github.com/users/jphme/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jphme", "id": 2862336, "login": "jphme", "node_id": "MDQ6VXNlcjI4NjIzMzY=", "organizations_url": "https://api.github.com/users/jphme/orgs", "received_events_url": "https://api.github.com/users/jphme/received_events", "repos_url": "https://api.github.com/users/jphme/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jphme/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jphme/subscriptions", "type": "User", "url": "https://api.github.com/users/jphme", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "Unlike Pandas, Arrow is strict with types, so converting the problematic strings to ints (or ints to strings) to ensure all the values have the same type is the only fix. \r\n\r\nJSON support has been requested in Arrow [here](https://github.com/apache/arrow/issues/32538), but I don't expect this to be implemented soon. \r\n\r\nAlso, this type could be represented with the Arrow Union type. However, due to low usage, the Union type has limited support in the Arrow ecosystem (e.g., IIRC Parquet still does not support it). So, we should probably wait a bit more before adding support for it in `datasets`", "> Unlike Pandas, Arrow is strict with types, so converting the problematic strings to ints (or ints to strings) to ensure all the values have the same type is the only fix.\r\n> \r\n> JSON support has been requested in Arrow [here](https://github.com/apache/arrow/issues/32538), but I don't expect this to be implemented soon.\r\n> \r\n> Also, this type could be represented with the Arrow Union type. However, due to low usage, the Union type has limited support in the Arrow ecosystem (e.g., IIRC Parquet still does not support it). So, we should probably wait a bit more before adding support for it in `datasets`\r\n\r\nOk many thanks, I was able to mitigate the problem by manually checking and converting all problematic fields now." ]
1970-01-01T00:00:00.000001
1,698
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug I have a list of dictionaries with valid/JSON-serializable values. One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on. If trying to convert this list to a dataset with `Dataset.from_list()`, I always get `ArrowInvalid: Could not convert '1' with type str: tried to convert to int64`, presumably because pyarrow tries to convert the keys to integers. Is there any way to circumvent this and fix dtypes? I didn't find anything in the documentation. ### Steps to reproduce the bug * create a list of dicts with one key being a string of an integer for the first few thousand occurences and try to convert to dataset. ### Expected behavior There shouldn't be an error (e.g. some flag to turn off automatic str to numeric conversion). ### Environment info - `datasets` version: 2.14.5 - Platform: Linux-5.15.0-84-generic-x86_64-with-glibc2.35 - Python version: 3.9.18 - Huggingface_hub version: 0.17.3 - PyArrow version: 13.0.0 - Pandas version: 2.1.1
{ "avatar_url": "https://avatars.githubusercontent.com/u/2862336?v=4", "events_url": "https://api.github.com/users/jphme/events{/privacy}", "followers_url": "https://api.github.com/users/jphme/followers", "following_url": "https://api.github.com/users/jphme/following{/other_user}", "gists_url": "https://api.github.com/users/jphme/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jphme", "id": 2862336, "login": "jphme", "node_id": "MDQ6VXNlcjI4NjIzMzY=", "organizations_url": "https://api.github.com/users/jphme/orgs", "received_events_url": "https://api.github.com/users/jphme/received_events", "repos_url": "https://api.github.com/users/jphme/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jphme/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jphme/subscriptions", "type": "User", "url": "https://api.github.com/users/jphme", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6324/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6324/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6323
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6323/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6323/comments
https://api.github.com/repos/huggingface/datasets/issues/6323/events
https://github.com/huggingface/datasets/issues/6323
1,954,245,980
I_kwDODunzps50e21c
6,323
Loading dataset from large GCS bucket very slow since 2.14
{ "avatar_url": "https://avatars.githubusercontent.com/u/6209990?v=4", "events_url": "https://api.github.com/users/jbcdnr/events{/privacy}", "followers_url": "https://api.github.com/users/jbcdnr/followers", "following_url": "https://api.github.com/users/jbcdnr/following{/other_user}", "gists_url": "https://api.github.com/users/jbcdnr/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jbcdnr", "id": 6209990, "login": "jbcdnr", "node_id": "MDQ6VXNlcjYyMDk5OTA=", "organizations_url": "https://api.github.com/users/jbcdnr/orgs", "received_events_url": "https://api.github.com/users/jbcdnr/received_events", "repos_url": "https://api.github.com/users/jbcdnr/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jbcdnr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jbcdnr/subscriptions", "type": "User", "url": "https://api.github.com/users/jbcdnr", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "I've also encountered this issue recently and want to ask if this has been seen.\r\n\r\n@albertvillanova for visibility - I'm not sure who the right person is to tag, but I noticed you were active recently so perhaps you can direct this to the right person.\r\n\r\nThanks!" ]
1970-01-01T00:00:00.000001
1,725
null
NONE
null
### Describe the bug Since updating to >2.14 we have very slow access to our parquet files on GCS when loading a dataset (>30 min vs 3s). Our GCS bucket has many objects and resolving globs is very slow. I could track down the problem to this change: https://github.com/huggingface/datasets/blame/bade7af74437347a760830466eb74f7a8ce0d799/src/datasets/data_files.py#L348 The underlying implementation with gcsfs is really slow. Could you go back to the old way if we are simply giving the parquet files and no glob pattern? Thank you. ### Steps to reproduce the bug Load a dataset from a GCS bucket that has many files. ### Expected behavior Used to be fast (3s) in 2.13 ### Environment info datasets==2.14.5
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6323/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6323/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6320
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6320/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6320/comments
https://api.github.com/repos/huggingface/datasets/issues/6320/events
https://github.com/huggingface/datasets/issues/6320
1,952,618,316
I_kwDODunzps50YpdM
6,320
Dataset slice splits can't load training and validation at the same time
{ "avatar_url": "https://avatars.githubusercontent.com/u/32488097?v=4", "events_url": "https://api.github.com/users/timlac/events{/privacy}", "followers_url": "https://api.github.com/users/timlac/followers", "following_url": "https://api.github.com/users/timlac/following{/other_user}", "gists_url": "https://api.github.com/users/timlac/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/timlac", "id": 32488097, "login": "timlac", "node_id": "MDQ6VXNlcjMyNDg4MDk3", "organizations_url": "https://api.github.com/users/timlac/orgs", "received_events_url": "https://api.github.com/users/timlac/received_events", "repos_url": "https://api.github.com/users/timlac/repos", "site_admin": false, "starred_url": "https://api.github.com/users/timlac/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/timlac/subscriptions", "type": "User", "url": "https://api.github.com/users/timlac", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The expression \"train+test\" concatenates the splits.\r\n\r\nThe individual splits as separate datasets can be obtained as follows:\r\n```python\r\ntrain_ds, test_ds = load_dataset(\"<dataset_name>\", split=[\"train\", \"test\"])\r\ntrain_10pct_ds, test_10pct_ds = load_dataset(\"<dataset_name>\", split=[\"train[:10%]\", \"test[:%10]\"])\r\n```" ]
1970-01-01T00:00:00.000001
1,701
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug According to the [documentation](https://huggingface.co/docs/datasets/v2.14.5/loading#slice-splits) is should be possible to run the following command: `train_test_ds = datasets.load_dataset("bookcorpus", split="train+test")` to load the train and test sets from the dataset. However executing the equivalent code: `speech_commands_v1 = load_dataset("superb", "ks", split="train+test")` only yields the following output: > Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 54175 > }) Where loading the dataset without the split argument yields: > DatasetDict({ > train: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 51094 > }) > validation: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 6798 > }) > test: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 3081 > }) > }) Thus, the API seems to be broken in this regard. This is a bit annoying since I want to be able to use the split argument with `split="train[:10%]+test[:10%]"` to have smaller dataset to work with when validating my model is working correctly. ### Steps to reproduce the bug `speech_commands_v1 = load_dataset("superb", "ks", split="train+test")` ### Expected behavior > DatasetDict({ > train: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 51094 > }) > test: Dataset({ > features: ['file', 'audio', 'label'], > num_rows: 3081 > }) > }) ### Environment info ``` import datasets print(datasets.__version__) ``` > 2.14.5 ``` import sys print(sys.version) ``` > 3.9.17 (main, Jul 5 2023, 20:41:20) > [GCC 11.2.0]
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6320/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6320/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6319
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6319/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6319/comments
https://api.github.com/repos/huggingface/datasets/issues/6319/events
https://github.com/huggingface/datasets/issues/6319
1,952,101,717
I_kwDODunzps50WrVV
6,319
Datasets.map is severely broken
{ "avatar_url": "https://avatars.githubusercontent.com/u/4603365?v=4", "events_url": "https://api.github.com/users/phalexo/events{/privacy}", "followers_url": "https://api.github.com/users/phalexo/followers", "following_url": "https://api.github.com/users/phalexo/following{/other_user}", "gists_url": "https://api.github.com/users/phalexo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/phalexo", "id": 4603365, "login": "phalexo", "node_id": "MDQ6VXNlcjQ2MDMzNjU=", "organizations_url": "https://api.github.com/users/phalexo/orgs", "received_events_url": "https://api.github.com/users/phalexo/received_events", "repos_url": "https://api.github.com/users/phalexo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/phalexo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/phalexo/subscriptions", "type": "User", "url": "https://api.github.com/users/phalexo", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Hi! Instead of processing a single example at a time, you should use the batched `map` for the best performance (with `num_proc=1`) - the fast tokenizers can process a batch's samples in parallel in that scenario.\r\n\r\nE.g., the following code in Colab takes an hour to complete:\r\n```python\r\n# !pip install datasets transformers\r\nfrom datasets import load_dataset\r\nfrom transformers import AutoTokenizer\r\ntokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\r\ndataset = dataset.map(lambda ex: tokenizer(ex[\"text\"]), batched=True, remove_columns=[\"text\", \"meta\"])\r\n```", "Batched is far worse. A single batch of 1000 took hours and that was only 1%\r\n\r\n\r\nOn Thu, Oct 19, 2023, 2:26 PM Mario Šaško ***@***.***> wrote:\r\n\r\n> Hi! You should use the batched map for the best performance (with\r\n> num_proc=1) - the fast tokenizers can process a batch's samples in\r\n> parallel.\r\n>\r\n> E.g., the following code in Colab takes an hour to complete:\r\n>\r\n> # !pip install datasets transformersfrom datasets import load_datasetfrom transformers import AutoTokenizertokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"]), batched=True, remove_columns=[\"text\", \"meta\"])\r\n>\r\n> —\r\n> Reply to this email directly, view it on GitHub\r\n> <https://github.com/huggingface/datasets/issues/6319#issuecomment-1771503757>,\r\n> or unsubscribe\r\n> <https://github.com/notifications/unsubscribe-auth/ABDD3ZJHPSRVDEXFNMXR2N3YAFWFZAVCNFSM6AAAAAA6HDKPSCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZRGUYDGNZVG4>\r\n> .\r\n> You are receiving this because you authored the thread.Message ID:\r\n> ***@***.***>\r\n>\r\n", "Can you please provide a self-contained reproducer?", "Which specific version of datasets are you using?\r\n\r\nWhat is the architecture of your colab setup? Ram? Cores? OS?\r\n\r\n\r\nOn Thu, Oct 19, 2023, 2:27 PM pensive introvert ***@***.***>\r\nwrote:\r\n\r\n> Batched is far worse. A single batch of 1000 took hours and that was only\r\n> 1%\r\n>\r\n>\r\n> On Thu, Oct 19, 2023, 2:26 PM Mario Šaško ***@***.***>\r\n> wrote:\r\n>\r\n>> Hi! You should use the batched map for the best performance (with\r\n>> num_proc=1) - the fast tokenizers can process a batch's samples in\r\n>> parallel.\r\n>>\r\n>> E.g., the following code in Colab takes an hour to complete:\r\n>>\r\n>> # !pip install datasets transformersfrom datasets import load_datasetfrom transformers import AutoTokenizertokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")dataset = dataset.map(lambda ex: tokenizer(ex[\"text\"]), batched=True, remove_columns=[\"text\", \"meta\"])\r\n>>\r\n>> —\r\n>> Reply to this email directly, view it on GitHub\r\n>> <https://github.com/huggingface/datasets/issues/6319#issuecomment-1771503757>,\r\n>> or unsubscribe\r\n>> <https://github.com/notifications/unsubscribe-auth/ABDD3ZJHPSRVDEXFNMXR2N3YAFWFZAVCNFSM6AAAAAA6HDKPSCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZRGUYDGNZVG4>\r\n>> .\r\n>> You are receiving this because you authored the thread.Message ID:\r\n>> ***@***.***>\r\n>>\r\n>\r\n", "from functools import partial\r\nimport transformers\r\nfrom datasets import load_dataset, concatenate_datasets, load_from_disk\r\n\r\nmodel_name_or_path=\"/opt/data/data/daryl149/llama-2-7b-chat-hf\"\r\noutput_dir=\"/opt/data/data/LongLoRA/checkpoints\"\r\ncache_dir=\"/opt/data/data/LongLoRA/cache\"\r\nmodel_max_length=16384\r\n\r\nIGNORE_INDEX = -100\r\nDEFAULT_PAD_TOKEN = \"[PAD]\"\r\nDEFAULT_EOS_TOKEN = \"</s>\"\r\nDEFAULT_BOS_TOKEN = \"<s>\"\r\nDEFAULT_UNK_TOKEN = \"<unk>\"\r\n\r\n\r\ntokenizer = transformers.LlamaTokenizerFast.from_pretrained(\r\n model_name_or_path,\r\n cache_dir=cache_dir,\r\n model_max_length=model_max_length,\r\n padding_side=\"right\",\r\n use_fast=True,\r\n #use_fast=False\r\n)\r\n\r\nspecial_tokens_dict = dict()\r\nif tokenizer.pad_token is None:\r\n special_tokens_dict[\"pad_token\"] = DEFAULT_PAD_TOKEN\r\nif tokenizer.eos_token is None:\r\n special_tokens_dict[\"eos_token\"] = DEFAULT_EOS_TOKEN\r\nif tokenizer.bos_token is None:\r\n special_tokens_dict[\"bos_token\"] = DEFAULT_BOS_TOKEN\r\nif tokenizer.unk_token is None:\r\n special_tokens_dict[\"unk_token\"] = DEFAULT_UNK_TOKEN\r\n\r\ntokenizer.add_special_tokens(special_tokens_dict)\r\n\r\ndef tokenize_fn(tokenizer, example):\r\n context_length = tokenizer.model_max_length\r\n outputs = tokenizer(\r\n tokenizer.eos_token.join(example[\"text\"]),\r\n #truncation=False,\r\n truncation=True,\r\n return_tensors=\"pt\",\r\n #return_tensors=\"np\",\r\n pad_to_multiple_of=context_length,\r\n padding=True,\r\n )\r\n return {\"input_ids\": outputs[\"input_ids\"].view(-1, context_length)}\r\n\r\nfor idx in range(100):\r\n dataset = load_dataset(\"togethercomputer/RedPajama-Data-1T-Sample\",\r\ncache_dir=cache_dir, split=f'train[{idx}%:{idx+1}%]')\r\n dataset = dataset.map(partial(tokenize_fn, tokenizer), batched=False,\r\nnum_proc=16, remove_columns=[\"text\", \"meta\"])\r\n dataset.save_to_disk(training_args.cache_dir + f\"/training_data_{idx}\")\r\n\r\n\r\nOn Thu, Oct 19, 2023 at 2:30 PM Mario Šaško ***@***.***>\r\nwrote:\r\n\r\n> Can you please provide a self-contained reproducer?\r\n>\r\n> —\r\n> Reply to this email directly, view it on GitHub\r\n> <https://github.com/huggingface/datasets/issues/6319#issuecomment-1771509229>,\r\n> or unsubscribe\r\n> <https://github.com/notifications/unsubscribe-auth/ABDD3ZNBZ3BE7Q4EQZZK6MLYAFWURAVCNFSM6AAAAAA6HDKPSCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZRGUYDSMRSHE>\r\n> .\r\n> You are receiving this because you authored the thread.Message ID:\r\n> ***@***.***>\r\n>\r\n", "I changed the tokenizer to one without \"Fast suffix, and something changed.\r\nThe fraction, although still slowed a lot at 80% was able to get over the\r\nfinish line of 100%\r\n\r\nI have to do more testng, see if the whole set can be processed\r\n\r\n\r\n\r\nOn Thu, Oct 19, 2023 at 3:03 PM pensive introvert <\r\n***@***.***> wrote:\r\n\r\n> from functools import partial\r\n> import transformers\r\n> from datasets import load_dataset, concatenate_datasets, load_from_disk\r\n>\r\n> model_name_or_path=\"/opt/data/data/daryl149/llama-2-7b-chat-hf\"\r\n> output_dir=\"/opt/data/data/LongLoRA/checkpoints\"\r\n> cache_dir=\"/opt/data/data/LongLoRA/cache\"\r\n> model_max_length=16384\r\n>\r\n> IGNORE_INDEX = -100\r\n> DEFAULT_PAD_TOKEN = \"[PAD]\"\r\n> DEFAULT_EOS_TOKEN = \"</s>\"\r\n> DEFAULT_BOS_TOKEN = \"<s>\"\r\n> DEFAULT_UNK_TOKEN = \"<unk>\"\r\n>\r\n>\r\n> tokenizer = transformers.LlamaTokenizerFast.from_pretrained(\r\n> model_name_or_path,\r\n> cache_dir=cache_dir,\r\n> model_max_length=model_max_length,\r\n> padding_side=\"right\",\r\n> use_fast=True,\r\n> #use_fast=False\r\n> )\r\n>\r\n> special_tokens_dict = dict()\r\n> if tokenizer.pad_token is None:\r\n> special_tokens_dict[\"pad_token\"] = DEFAULT_PAD_TOKEN\r\n> if tokenizer.eos_token is None:\r\n> special_tokens_dict[\"eos_token\"] = DEFAULT_EOS_TOKEN\r\n> if tokenizer.bos_token is None:\r\n> special_tokens_dict[\"bos_token\"] = DEFAULT_BOS_TOKEN\r\n> if tokenizer.unk_token is None:\r\n> special_tokens_dict[\"unk_token\"] = DEFAULT_UNK_TOKEN\r\n>\r\n> tokenizer.add_special_tokens(special_tokens_dict)\r\n>\r\n> def tokenize_fn(tokenizer, example):\r\n> context_length = tokenizer.model_max_length\r\n> outputs = tokenizer(\r\n> tokenizer.eos_token.join(example[\"text\"]),\r\n> #truncation=False,\r\n> truncation=True,\r\n> return_tensors=\"pt\",\r\n> #return_tensors=\"np\",\r\n> pad_to_multiple_of=context_length,\r\n> padding=True,\r\n> )\r\n> return {\"input_ids\": outputs[\"input_ids\"].view(-1, context_length)}\r\n>\r\n> for idx in range(100):\r\n> dataset = load_dataset(\"togethercomputer/RedPajama-Data-1T-Sample\",\r\n> cache_dir=cache_dir, split=f'train[{idx}%:{idx+1}%]')\r\n> dataset = dataset.map(partial(tokenize_fn, tokenizer), batched=False,\r\n> num_proc=16, remove_columns=[\"text\", \"meta\"])\r\n> dataset.save_to_disk(training_args.cache_dir + f\"/training_data_{idx}\")\r\n>\r\n>\r\n> On Thu, Oct 19, 2023 at 2:30 PM Mario Šaško ***@***.***>\r\n> wrote:\r\n>\r\n>> Can you please provide a self-contained reproducer?\r\n>>\r\n>> —\r\n>> Reply to this email directly, view it on GitHub\r\n>> <https://github.com/huggingface/datasets/issues/6319#issuecomment-1771509229>,\r\n>> or unsubscribe\r\n>> <https://github.com/notifications/unsubscribe-auth/ABDD3ZNBZ3BE7Q4EQZZK6MLYAFWURAVCNFSM6AAAAAA6HDKPSCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZRGUYDSMRSHE>\r\n>> .\r\n>> You are receiving this because you authored the thread.Message ID:\r\n>> ***@***.***>\r\n>>\r\n>\r\n", "So, using LlamaTokenizerFast was the problem. Changing it to LlamaTokenizer\r\nfixed things,\r\n\r\nOn Thu, Oct 19, 2023 at 4:04 PM pensive introvert <\r\n***@***.***> wrote:\r\n\r\n> I changed the tokenizer to one without \"Fast suffix, and something\r\n> changed. The fraction, although still slowed a lot at 80% was able to get\r\n> over the finish line of 100%\r\n>\r\n> I have to do more testng, see if the whole set can be processed\r\n>\r\n>\r\n>\r\n> On Thu, Oct 19, 2023 at 3:03 PM pensive introvert <\r\n> ***@***.***> wrote:\r\n>\r\n>> from functools import partial\r\n>> import transformers\r\n>> from datasets import load_dataset, concatenate_datasets, load_from_disk\r\n>>\r\n>> model_name_or_path=\"/opt/data/data/daryl149/llama-2-7b-chat-hf\"\r\n>> output_dir=\"/opt/data/data/LongLoRA/checkpoints\"\r\n>> cache_dir=\"/opt/data/data/LongLoRA/cache\"\r\n>> model_max_length=16384\r\n>>\r\n>> IGNORE_INDEX = -100\r\n>> DEFAULT_PAD_TOKEN = \"[PAD]\"\r\n>> DEFAULT_EOS_TOKEN = \"</s>\"\r\n>> DEFAULT_BOS_TOKEN = \"<s>\"\r\n>> DEFAULT_UNK_TOKEN = \"<unk>\"\r\n>>\r\n>>\r\n>> tokenizer = transformers.LlamaTokenizerFast.from_pretrained(\r\n>> model_name_or_path,\r\n>> cache_dir=cache_dir,\r\n>> model_max_length=model_max_length,\r\n>> padding_side=\"right\",\r\n>> use_fast=True,\r\n>> #use_fast=False\r\n>> )\r\n>>\r\n>> special_tokens_dict = dict()\r\n>> if tokenizer.pad_token is None:\r\n>> special_tokens_dict[\"pad_token\"] = DEFAULT_PAD_TOKEN\r\n>> if tokenizer.eos_token is None:\r\n>> special_tokens_dict[\"eos_token\"] = DEFAULT_EOS_TOKEN\r\n>> if tokenizer.bos_token is None:\r\n>> special_tokens_dict[\"bos_token\"] = DEFAULT_BOS_TOKEN\r\n>> if tokenizer.unk_token is None:\r\n>> special_tokens_dict[\"unk_token\"] = DEFAULT_UNK_TOKEN\r\n>>\r\n>> tokenizer.add_special_tokens(special_tokens_dict)\r\n>>\r\n>> def tokenize_fn(tokenizer, example):\r\n>> context_length = tokenizer.model_max_length\r\n>> outputs = tokenizer(\r\n>> tokenizer.eos_token.join(example[\"text\"]),\r\n>> #truncation=False,\r\n>> truncation=True,\r\n>> return_tensors=\"pt\",\r\n>> #return_tensors=\"np\",\r\n>> pad_to_multiple_of=context_length,\r\n>> padding=True,\r\n>> )\r\n>> return {\"input_ids\": outputs[\"input_ids\"].view(-1, context_length)}\r\n>>\r\n>> for idx in range(100):\r\n>> dataset = load_dataset(\"togethercomputer/RedPajama-Data-1T-Sample\",\r\n>> cache_dir=cache_dir, split=f'train[{idx}%:{idx+1}%]')\r\n>> dataset = dataset.map(partial(tokenize_fn, tokenizer), batched=False,\r\n>> num_proc=16, remove_columns=[\"text\", \"meta\"])\r\n>> dataset.save_to_disk(training_args.cache_dir +\r\n>> f\"/training_data_{idx}\")\r\n>>\r\n>>\r\n>> On Thu, Oct 19, 2023 at 2:30 PM Mario Šaško ***@***.***>\r\n>> wrote:\r\n>>\r\n>>> Can you please provide a self-contained reproducer?\r\n>>>\r\n>>> —\r\n>>> Reply to this email directly, view it on GitHub\r\n>>> <https://github.com/huggingface/datasets/issues/6319#issuecomment-1771509229>,\r\n>>> or unsubscribe\r\n>>> <https://github.com/notifications/unsubscribe-auth/ABDD3ZNBZ3BE7Q4EQZZK6MLYAFWURAVCNFSM6AAAAAA6HDKPSCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZRGUYDSMRSHE>\r\n>>> .\r\n>>> You are receiving this because you authored the thread.Message ID:\r\n>>> ***@***.***>\r\n>>>\r\n>>\r\n", "Indeed, the tokenizer is super slow. Perhaps @ArthurZucker knows the reason why.\r\n\r\n([This](https://colab.research.google.com/drive/1VgeurX-4Fl2X6aBQTwh_X4kuQKZ6K9L1?usp=sharing) simplified Colab can be used to reproduce the behavior)", "same issue here\r\nsample to reproduce: https://github.com/philschmid/document-ai-transformers/blob/main/training/donut_sroie.ipynb\r\nwith following map line\r\nhttps://github.com/philschmid/document-ai-transformers/blob/main/training/donut_sroie.ipynb\r\n\r\nIf I directly iterate over the dataset and call the mapping method, it is very fast\r\n```py\r\nfor sample in dataset:\r\n def preprocess_documents_for_donut(sample):\r\n```\r\n\r\nif i removed `.convert('RGB')` It can run to completion without getting stuck. I suspect it has something to do with the Image.\r\n\r\nIf I use batch, it's even slower.", "@ewfian \r\n\r\n> If I directly iterate over the dataset and call the mapping method, it is very fast\r\n\r\n`Dataset.map` must also convert the images into bytes to write them to an Arrow file (the write itself takes some time, too). \r\n\r\nYou can make the `map` faster by manually converting the images into an \"arrow-compatible\" representation. Otherwise, the Pillow defaults are used when saving an image, which seems particularly slow for the notebook's case.\r\n\r\n```python\r\ndef preprocess_documents_for_donut(sample):\r\n text = json.loads(sample[\"text\"])\r\n d_doc = task_start_token + json2token(text) + eos_token\r\n image = sample[\"image\"].convert('RGB')\r\n # convert image to bytes\r\n buffer = io.BytesIO()\r\n image.save(buffer, format=\"PNG\", compress_level=1)\r\n return {\"image\": {\"bytes\": buffer.getvalue()}, \"text\": d_doc}\r\n\r\nproc_dataset = dataset.map(preprocess_documents_for_donut, writer_batch_size=50)\r\n```", "The problem I had was to do with map using fork and copying locks from the\r\nparent process in acquired state. I ended up changing the context to use\r\nforkserver instead.\r\n\r\n\r\nOn Wed, Nov 29, 2023, 10:04 PM Mario Šaško ***@***.***> wrote:\r\n\r\n> @ewfian <https://github.com/ewfian>\r\n>\r\n> If I directly iterate over the dataset and call the mapping method, it is\r\n> very fast\r\n>\r\n> Dataset.map must also convert the images into bytes to write them to an\r\n> Arrow file (the write itself takes some time, too).\r\n>\r\n> You can make the map faster by manually converting the images into an\r\n> \"arrow-compatible\" representation. Otherwise, the Pillow defaults are used\r\n> when saving an image, which seems particularly slow for the notebook's case.\r\n>\r\n> def preprocess_documents_for_donut(sample):\r\n> text = json.loads(sample[\"text\"])\r\n> d_doc = task_start_token + json2token(text) + eos_token\r\n> image = sample[\"image\"].convert('RGB')\r\n> # convert image to bytes\r\n> buffer = io.BytesIO()\r\n> image.save(buffer, format=\"PNG\", compress_level=1)\r\n> return {\"image\": {\"bytes\": buffer.getvalue()}, \"text\": d_doc}\r\n> proc_dataset = dataset.map(preprocess_documents_for_donut, writer_batch_size=50)\r\n>\r\n> —\r\n> Reply to this email directly, view it on GitHub\r\n> <https://github.com/huggingface/datasets/issues/6319#issuecomment-1833033973>,\r\n> or unsubscribe\r\n> <https://github.com/notifications/unsubscribe-auth/ABDD3ZKKEKJVWBFH7QHLRJ3YG7ZUJAVCNFSM6AAAAAA6HDKPSCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQMZTGAZTGOJXGM>\r\n> .\r\n> You are receiving this because you authored the thread.Message ID:\r\n> ***@***.***>\r\n>\r\n", "I face the same issue many times.\r\n\r\nNot only when using the transformers' tokenizer, but also when applying nltk's [pos_tag](https://www.nltk.org/api/nltk.tag.pos_tag.html) to the entire English Wikipedia. So I suspect the cause is not in the tokenizer but in the Dataset.map\r\n\r\nMy case:\r\nAt the beginning of the run, the speed was 600 samples/s, but it slowed down to 20 samples/s at around 90% (after 3 hours). I am concerned that the CPU usage was only about 5% at the end of the run, even though there was still lots of data left.", "It is the interaction of fork() inside the map and tokenizer mutexes/locks.\r\n\r\nYou have to set up your own process pool and use fork server instead of\r\nfork.\r\n\r\n\r\nOn Tue, Aug 6, 2024, 11:44 AM yuji96 ***@***.***> wrote:\r\n\r\n> I face the same issue many times.\r\n>\r\n> Not only when using the transformers' tokenizer, but also when applying\r\n> nltk's pos_tag <https://www.nltk.org/api/nltk.tag.pos_tag.html> to the\r\n> entire English Wikipedia. So I suspect the cause is not in the tokenizer\r\n> but in the Dataset.map\r\n>\r\n> My case:\r\n> At the beginning of the run, the speed was 600 samples/s, but it slowed\r\n> down to 20 samples/s at around 90% (after 3 hours). I am concerned that the\r\n> CPU usage was only about 5% at the end of the run, even though there was\r\n> still lots of data left.\r\n>\r\n> #6319 (comment)\r\n> <https://github.com/huggingface/datasets/issues/6319#issuecomment-1771629160>\r\n> It's very nice to hear that the run is complete, but the original issue\r\n> has not been solved, which is that it gets slower and slower. As it is now,\r\n> Dataset.map will not be able to handle the large datasets that are getting\r\n> larger day by day.\r\n>\r\n> —\r\n> Reply to this email directly, view it on GitHub\r\n> <https://github.com/huggingface/datasets/issues/6319#issuecomment-2271603976>,\r\n> or unsubscribe\r\n> <https://github.com/notifications/unsubscribe-auth/ABDD3ZLFHSIGNNXAEJJIXWLZQDVPTAVCNFSM6AAAAABMCTVK2SVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDENZRGYYDGOJXGY>\r\n> .\r\n> You are receiving this because you authored the thread.Message ID:\r\n> ***@***.***>\r\n>\r\n", "Thank you for your advice!\r\n\r\nI added `multiprocess.set_start_method(\"forkserver\")` but the result seemed to be the same. In my case, it may be due to the very simple fact that about 10% of the process, which includes long text, never ends. I'll try shard by data size.\r\n![image](https://github.com/user-attachments/assets/03f4be67-1a6b-4d33-88f6-a46a9b1d37f4)\r\n", "Would recommend using `LlamaTokenizerFast` not `LlamaTokenizer` ! " ]
1970-01-01T00:00:00.000001
1,723
null
NONE
null
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing something, but it never ends. I saw some comments about fast tokenizers using Rust and all and tried different variations. NOTHING works. ### Steps to reproduce the bug Running it without breaking the dataset into parts results in the same behavior. The loop was an attempt to see if this was a RAM issue. for idx in range(100): dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=cache_dir, split=f'train[{idx}%:{idx+1}%]') dataset = dataset.map(partial(tokenize_fn, tokenizer), batched=False, num_proc=1, remove_columns=["text", "meta"]) dataset.save_to_disk(training_args.cache_dir + f"/training_data_{idx}") ### Expected behavior I expect map to run at more or less the same speed it starts with and FINISH its processing. ### Environment info Python 3.8, same with 3.10 makes no difference. Ubuntu 20.04,
null
{ "+1": 6, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 6, "url": "https://api.github.com/repos/huggingface/datasets/issues/6319/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6319/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6317
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6317/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6317/comments
https://api.github.com/repos/huggingface/datasets/issues/6317/events
https://github.com/huggingface/datasets/issues/6317
1,951,965,668
I_kwDODunzps50WKHk
6,317
sentiment140 dataset unavailable
{ "avatar_url": "https://avatars.githubusercontent.com/u/52670382?v=4", "events_url": "https://api.github.com/users/AndreasKarasenko/events{/privacy}", "followers_url": "https://api.github.com/users/AndreasKarasenko/followers", "following_url": "https://api.github.com/users/AndreasKarasenko/following{/other_user}", "gists_url": "https://api.github.com/users/AndreasKarasenko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AndreasKarasenko", "id": 52670382, "login": "AndreasKarasenko", "node_id": "MDQ6VXNlcjUyNjcwMzgy", "organizations_url": "https://api.github.com/users/AndreasKarasenko/orgs", "received_events_url": "https://api.github.com/users/AndreasKarasenko/received_events", "repos_url": "https://api.github.com/users/AndreasKarasenko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AndreasKarasenko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AndreasKarasenko/subscriptions", "type": "User", "url": "https://api.github.com/users/AndreasKarasenko", "user_view_type": "public" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" } ]
null
[ "Thanks for reporting. We are investigating the issue.", "We have opened an issue in the corresponding Hub dataset: https://huggingface.co/datasets/sentiment140/discussions/3\r\n\r\nLet's continue the discussion there." ]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug loading the dataset using load_dataset("sentiment140") returns the following error ConnectionError: Couldn't reach http://cs.stanford.edu/people/alecmgo/trainingandtestdata.zip (error 403) ### Steps to reproduce the bug Run the following code (version should not matter). ``` from datasets import load_dataset data = load_dataset("sentiment140") ``` ### Expected behavior The dataset should be loaded just like any other. The main issue is that it is no longer hosted by stanford. It is still available from a [Google Drive Link](https://docs.google.com/file/d/0B04GJPshIjmPRnZManQwWEdTZjg/edit). ### Environment info - `datasets` version: 2.14.5 - Platform: Windows-10-10.0.19045-SP0 - Python version: 3.10.8 - Huggingface_hub version: 0.17.3 - PyArrow version: 13.0.0 - Pandas version: 2.1.1
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6317/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6317/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6315
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6315/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6315/comments
https://api.github.com/repos/huggingface/datasets/issues/6315/events
https://github.com/huggingface/datasets/issues/6315
1,951,800,819
I_kwDODunzps50Vh3z
6,315
Hub datasets with CSV metadata raise ArrowInvalid: JSON parse error: Invalid value. in row 0
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" } ]
null
[]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
MEMBER
null
When trying to load a Hub dataset that contains a CSV metadata file, it raises an `ArrowInvalid` error: ``` E pyarrow.lib.ArrowInvalid: JSON parse error: Invalid value. in row 0 pyarrow/error.pxi:100: ArrowInvalid ``` See: https://huggingface.co/datasets/lukarape/public_small_papers/discussions/1
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6315/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6315/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6311
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6311/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6311/comments
https://api.github.com/repos/huggingface/datasets/issues/6311/events
https://github.com/huggingface/datasets/issues/6311
1,949,304,993
I_kwDODunzps50MAih
6,311
cast_column to Sequence with length=4 occur exception raise in datasets/table.py:2146
{ "avatar_url": "https://avatars.githubusercontent.com/u/16574677?v=4", "events_url": "https://api.github.com/users/neiblegy/events{/privacy}", "followers_url": "https://api.github.com/users/neiblegy/followers", "following_url": "https://api.github.com/users/neiblegy/following{/other_user}", "gists_url": "https://api.github.com/users/neiblegy/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/neiblegy", "id": 16574677, "login": "neiblegy", "node_id": "MDQ6VXNlcjE2NTc0Njc3", "organizations_url": "https://api.github.com/users/neiblegy/orgs", "received_events_url": "https://api.github.com/users/neiblegy/received_events", "repos_url": "https://api.github.com/users/neiblegy/repos", "site_admin": false, "starred_url": "https://api.github.com/users/neiblegy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neiblegy/subscriptions", "type": "User", "url": "https://api.github.com/users/neiblegy", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "Thanks for reporting! We've spotted the bugs with the `array.values` handling and are fixing them in https://github.com/huggingface/datasets/pull/6283 (should be part of the next release).", "> Thanks for reporting! We've spotted the bugs with the `array.values` handling and are fixing them in #6283 (should be part of the next release).\r\n\r\ni encounter another exception while cast_column to type `Sequence(feature={\"points\": Array2D(shape=(-1, 2), dtype=\"int64\"), \"label\": ClassLabel(num_classes=num_classes, names=names)})`\r\n\r\nwhile my data like this: '{\"points\": [[0.6,0.6], [0.7,0.7], [0.8,0.8]], \"label\": \"A1\"}'\r\n\r\nhere is the backtrace info:\r\n\r\n```\r\n out = func(dataset, *args, **kwargs)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py\", line 2110, in cast_column\r\n return self.cast(features)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py\", line 2055, in cast\r\n dataset = dataset.map(\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py\", line 592, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py\", line 557, in wrapper\r\n out: Union[\"Dataset\", \"DatasetDict\"] = func(self, *args, **kwargs)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py\", line 3097, in map\r\n for rank, done, content in Dataset._map_single(**dataset_kwargs):\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py\", line 3474, in _map_single\r\n batch = apply_function_on_filtered_inputs(\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py\", line 3353, in apply_function_on_filtered_inputs\r\n processed_inputs = function(*fn_args, *additional_args, **fn_kwargs)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 2329, in table_cast\r\n return cast_table_to_schema(table, schema)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 2288, in cast_table_to_schema\r\n arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()]\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 2288, in <listcomp>\r\n arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()]\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 1831, in wrapper\r\n return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 1831, in <listcomp>\r\n return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 2073, in cast_array_to_feature\r\n arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()]\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 2073, in <listcomp>\r\n arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()]\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 1833, in wrapper\r\n return func(array, *args, **kwargs)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 2095, in cast_array_to_feature\r\n casted_values = _c(array.values, feature.feature)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 1833, in wrapper\r\n return func(array, *args, **kwargs)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 2144, in cast_array_to_feature\r\n return array_cast(array, feature(), allow_number_to_str=allow_number_to_str)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 1833, in wrapper\r\n return func(array, *args, **kwargs)\r\n File \"/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py\", line 1967, in array_cast\r\n return pa_type.wrap_array(array)\r\n File \"pyarrow/types.pxi\", line 1369, in pyarrow.lib.BaseExtensionType.wrap_array\r\nTypeError: Incompatible storage type for extension<arrow.py_extension_type<Array2DExtensionType>>: expected list<item: list<item: double>>, got list<item: double>\r\n```\r\nand i print(array) in datasets/table.py:1967 indeed get 2D list. is that same issue in #6283 ?\r\n\r\nbesides this, hugging face datasets seems don't naturally support multi-labels which means `Sequence(ClassLabel)` illegal if data is [\"label1\", \"label2\"]. so i have to define a class derived from `ClassLabel`, like this:\r\n\r\n```\r\nclass AisClassLabels(ClassLabel):\r\n def encode_example(self, example_data):\r\n if self.num_classes is None:\r\n raise ValueError(\r\n \"Trying to use ClassLabel feature with undefined number of class. \"\r\n \"Please set ClassLabel.names or num_classes.\"\r\n )\r\n if not isinstance(example_data, list):\r\n example_data = [example_data]\r\n\r\n for i in range(len(example_data)):\r\n if isinstance(example_data[i], str):\r\n example_data[i] = self.str2int(example_data[i])\r\n if not -1 <= example_data[i] < self.num_classes:\r\n raise ValueError(f\"Class label {example_data:d} greater than configured num_classes {self.num_classes}\")\r\n return example_data\r\n```\r\nand it works well in my case. but is there any recommend way to implement multi-labels?", "`Incompatible storage type for extension<arrow.py_extension_type<Array2DExtensionType>>: expected list<item: list<item: double>>, got list<item: double>`\r\nif i change `Array2D(shape=(-1, 2), dtype=\"int64\")` to `Sequence(Value(\"int64\"))` , every thing goes well. but my data is 2D int list", "i test Sequence(ClassLabel) is ok if one column is label list. but it is not ok in nested column such as `Sequence(feature= {\"points\": Sequence(Value(\"int32\")), \"label\": Sequence(ClassLabel(num_classes....)))`. in this case i need override ClassLabels. encode_example as i given above." ]
1970-01-01T00:00:00.000001
1,707
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug i load a dataset from local csv file which has 187383612 examples, then use `map` to generate new columns for test. here is my code : ``` import os from datasets import load_dataset from datasets.features import Sequence, Value def add_new_path(example): example["ais_bbox"] = [100,100,200,200] example["ais_image_path"] = os.path.join("images", example["image_path"]) if example["image_path"] else "" return example ais_dataset = load_dataset("/data/ryan.gao/ais_dataset_cache/raw/1749/") hf_ds = ais_dataset.map(add_new_path, batched=False, num_proc=32) ds = hf_ds.cast_column("ais_bbox", Sequence(Value("int32"), length=4)) ``` and the `cast_column` raise an exception ``` Casting the dataset: 3%|███▉ ... File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 2110, in cast_column return self.cast(features) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 2055, in cast dataset = dataset.map( File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 592, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3097, in map for rank, done, content in Dataset._map_single(**dataset_kwargs): File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3474, in _map_single batch = apply_function_on_filtered_inputs( File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3353, in apply_function_on_filtered_inputs processed_inputs = function(*fn_args, *additional_args, **fn_kwargs) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2329, in table_cast return cast_table_to_schema(table, schema) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2288, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2288, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 1831, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 1831, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/protoss.gao/.local/lib/python3.9/site-packages/datasets/table.py", line 2145, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") TypeError: Couldn't cast array of type list<item: int64> to Sequence(feature=Value(dtype='int32', id=None), length=4, id=None) ``` i check the source code and make debug info: in datasets/table.py:2092 ``` 2091 if feature.length > -1: 2092 if feature.length * len(array) == len(array.values): 2093 return pa.FixedSizeListArray.from_arrays(_c(array.values, feature.feature), feature.length) 2094 print(len(array)) 2095 print(len(array.values)) ``` my feature.length is 4. but feature.length * len(array) == len(array.values) is false. print(len(array)) is 262 print(len(array.values)) is 4000 then I use "for item in array" to print each item then get 262 * [100,100,200,200] and use "for item in array.values" to print each item and get 4000 int32 which are 1000 * [100,100,200,200] i'm wondering the `chunk` in each `array.chunks`, the "chunk.values" may get all the chunks's value rather than single chunk? but i check the pyarrow's doc seems chunk.values is chunk's value not all. ### Steps to reproduce the bug code provided above. ### Expected behavior feature.length * len(array) == len(array.values) should be true. and there should not has Exception. ### Environment info python3.9 x86_64 datasets: 2.14.4 pyarrow: 13.0.0 or 10.0.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6311/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6311/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6308
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6308/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6308/comments
https://api.github.com/repos/huggingface/datasets/issues/6308/events
https://github.com/huggingface/datasets/issues/6308
1,946,810,625
I_kwDODunzps50CfkB
6,308
module 'resource' has no attribute 'error'
{ "avatar_url": "https://avatars.githubusercontent.com/u/48009681?v=4", "events_url": "https://api.github.com/users/NeoWang9999/events{/privacy}", "followers_url": "https://api.github.com/users/NeoWang9999/followers", "following_url": "https://api.github.com/users/NeoWang9999/following{/other_user}", "gists_url": "https://api.github.com/users/NeoWang9999/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NeoWang9999", "id": 48009681, "login": "NeoWang9999", "node_id": "MDQ6VXNlcjQ4MDA5Njgx", "organizations_url": "https://api.github.com/users/NeoWang9999/orgs", "received_events_url": "https://api.github.com/users/NeoWang9999/received_events", "repos_url": "https://api.github.com/users/NeoWang9999/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NeoWang9999/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NeoWang9999/subscriptions", "type": "User", "url": "https://api.github.com/users/NeoWang9999", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "This (Windows) issue was fixed in `fsspec` in https://github.com/fsspec/filesystem_spec/pull/1275. So, to avoid the error, update the `fsspec` installation with `pip install -U fsspec`.", "> This (Windows) issue was fixed in `fsspec` in [fsspec/filesystem_spec#1275](https://github.com/fsspec/filesystem_spec/pull/1275). So, to avoid the error, update the `fsspec` installation with `pip install -U fsspec`.\r\n\r\nafter I run `pip install -U fsspec`\r\n\r\nit occurs a new error:\r\n```\r\nERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflict\r\ns.\r\ndatasets 2.14.5 requires fsspec[http]<2023.9.0,>=2023.1.0, but you have fsspec 2023.9.2 which is incompatible.\r\n\r\n```", "The `fsspec<2023.9.0` upper bound will be removed in the next release. The `ResourceError` fix is also present in version 2023.6.0, so use that version in the meantime (`pip install fsspec==2023.6.0`).", "> The `fsspec<2023.9.0` upper bound will be removed in the next release. The `ResourceError` fix is also present in version 2023.6.0, so use that version in the meantime (`pip install fsspec==2023.6.0`).\r\n\r\nthanks for reply!" ]
1970-01-01T00:00:00.000001
1,698
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug just run import: `from datasets import load_dataset` and then: ``` File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\__init__.py", line 22, in <module> from .arrow_dataset import Dataset File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\arrow_dataset.py", line 66, in <module> from .arrow_reader import ArrowReader File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\arrow_reader.py", line 30, in <module> from .download.download_config import DownloadConfig File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\download\__init__.py", line 10, in <module> from .streaming_download_manager import StreamingDownloadManager File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\download\streaming_download_manager.py", line 21, in <module> from ..filesystems import COMPRESSION_FILESYSTEMS File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\datasets\filesystems\__init__.py", line 8, in <module> import fsspec.asyn File "C:\ProgramData\anaconda3\envs\py310\lib\site-packages\fsspec\asyn.py", line 157, in <module> ResourceEror = resource.error AttributeError: module 'resource' has no attribute 'error' Process finished with exit code 1 ``` and the error codes are: ``` try: import resource except ImportError: resource = None ResourceError = OSError else: ResourceEror = resource.error ``` 1. miss spelling : "ResourceEror " should be "ResourceErorr" 2. module 'resource' has no attribute 'error' ### Steps to reproduce the bug only one step: `from datasets import load_dataset` ### Expected behavior slove error: module 'resource' has no attribute 'error' ### Environment info python=3.10 datasets==2.14.5
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6308/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6308/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6306
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6306/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6306/comments
https://api.github.com/repos/huggingface/datasets/issues/6306/events
https://github.com/huggingface/datasets/issues/6306
1,946,363,452
I_kwDODunzps50AyY8
6,306
pyinstaller : OSError: could not get source code
{ "avatar_url": "https://avatars.githubusercontent.com/u/57702070?v=4", "events_url": "https://api.github.com/users/dusk877647949/events{/privacy}", "followers_url": "https://api.github.com/users/dusk877647949/followers", "following_url": "https://api.github.com/users/dusk877647949/following{/other_user}", "gists_url": "https://api.github.com/users/dusk877647949/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dusk877647949", "id": 57702070, "login": "dusk877647949", "node_id": "MDQ6VXNlcjU3NzAyMDcw", "organizations_url": "https://api.github.com/users/dusk877647949/orgs", "received_events_url": "https://api.github.com/users/dusk877647949/received_events", "repos_url": "https://api.github.com/users/dusk877647949/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dusk877647949/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dusk877647949/subscriptions", "type": "User", "url": "https://api.github.com/users/dusk877647949", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "more information:\r\n``` \r\nFile \"text2vec\\__init__.py\", line 8, in <module>\r\nFile \"<frozen importlib._bootstrap>\", line 1027, in _find_and_load\r\nFile \"<frozen importlib._bootstrap>\", line 1006, in _find_and_load_unlocked\r\nFile \"<frozen importlib._bootstrap>\", line 688, in _load_unlocked\r\nFile \"PyInstaller\\loader\\pyimod02_importers.py\", line 499, in exec_module\r\nFile \"text2vec\\bertmatching_model.py\", line 19, in <module>\r\nFile \"<frozen importlib._bootstrap>\", line 1027, in _find_and_load\r\nFile \"<frozen importlib._bootstrap>\", line 1006, in _find_and_load_unlocked\r\nFile \"<frozen importlib._bootstrap>\", line 688, in _load_unlocked\r\nFile \"PyInstaller\\loader\\pyimod02_importers.py\", line 499, in exec_module\r\nFile \"text2vec\\bertmatching_dataset.py\", line 7, in <module>\r\nFile \"<frozen importlib._bootstrap>\", line 1027, in _find_and_load\r\nFile \"<frozen importlib._bootstrap>\", line 1006, in _find_and_load_unlocked\r\nFile \"<frozen importlib._bootstrap>\", line 688, in _load_unlocked\r\nFile \"PyInstaller\\loader\\pyimod02_importers.py\", line 499, in exec_module\r\nFile \"datasets\\__init__.py\", line 52, in <module>\r\nFile \"<frozen importlib._bootstrap>\", line 1027, in _find_and_load\r\nFile \"<frozen importlib._bootstrap>\", line 1006, in _find_and_load_unlocked\r\nFile \"<frozen importlib._bootstrap>\", line 688, in _load_unlocked\r\nFile \"PyInstaller\\loader\\pyimod02_importers.py\", line 499, in exec_module\r\nFile \"datasets\\inspect.py\", line 30, in <module>\r\nFile \"<frozen importlib._bootstrap>\", line 1027, in _find_and_load\r\nFile \"<frozen importlib._bootstrap>\", line 1006, in _find_and_load_unlocked\r\nFile \"<frozen importlib._bootstrap>\", line 688, in _load_unlocked\r\nFile \"PyInstaller\\loader\\pyimod02_importers.py\", line 499, in exec_module\r\nFile \"datasets\\load.py\", line 58, in <module>\r\nFile \"<frozen importlib._bootstrap>\", line 1027, in _find_and_load\r\nFile \"<frozen importlib._bootstrap>\", line 1006, in _find_and_load_unlocked\r\nFile \"<frozen importlib._bootstrap>\", line 688, in _load_unlocked\r\nFile \"PyInstaller\\loader\\pyimod02_importers.py\", line 499, in exec_module\r\nFile \"datasets\\packaged_modules\\__init__.py\", line 31, in <module>\r\nFile \"inspect.py\", line 1147, in getsource\r\nFile \"inspect.py\", line 1129, in getsourcelines\r\nFile \"inspect.py\", line 958, in findsource\r\nOSError: could not get source code\r\n```\r\n", "Can you share a reproducer? I haven't been able to reproduce the error myself.", "> '\r\n\r\nthanks,I solve it.it's about pyinstaller.", "1", "> > '\r\n> \r\n> thanks,I solve it.it's about pyinstaller.\r\n\r\nI encountered the same error, how to solve it?" ]
1970-01-01T00:00:00.000001
1,698
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug I ran a package with pyinstaller and got the following error: ### Steps to reproduce the bug ``` ... File "datasets\__init__.py", line 52, in <module> File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "PyInstaller\loader\pyimod02_importers.py", line 499, in exec_module File "datasets\inspect.py", line 30, in <module> File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "PyInstaller\loader\pyimod02_importers.py", line 499, in exec_module File "datasets\load.py", line 58, in <module> File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "PyInstaller\loader\pyimod02_importers.py", line 499, in exec_module File "datasets\packaged_modules\__init__.py", line 31, in <module> File "inspect.py", line 1147, in getsource File "inspect.py", line 1129, in getsourcelines File "inspect.py", line 958, in findsource OSError: could not get source code ``` ### Expected behavior I have looked up the relevant information, but I can't find a suitable reason ### Environment info ```python python 3.10 datasets 2.14.4 pyinstaller 5.6.2 ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/57702070?v=4", "events_url": "https://api.github.com/users/dusk877647949/events{/privacy}", "followers_url": "https://api.github.com/users/dusk877647949/followers", "following_url": "https://api.github.com/users/dusk877647949/following{/other_user}", "gists_url": "https://api.github.com/users/dusk877647949/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dusk877647949", "id": 57702070, "login": "dusk877647949", "node_id": "MDQ6VXNlcjU3NzAyMDcw", "organizations_url": "https://api.github.com/users/dusk877647949/orgs", "received_events_url": "https://api.github.com/users/dusk877647949/received_events", "repos_url": "https://api.github.com/users/dusk877647949/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dusk877647949/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dusk877647949/subscriptions", "type": "User", "url": "https://api.github.com/users/dusk877647949", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6306/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6306/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6305
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6305/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6305/comments
https://api.github.com/repos/huggingface/datasets/issues/6305/events
https://github.com/huggingface/datasets/issues/6305
1,946,010,912
I_kwDODunzps5z_cUg
6,305
Cannot load dataset with `2.14.5`: `FileNotFound` error
{ "avatar_url": "https://avatars.githubusercontent.com/u/167943?v=4", "events_url": "https://api.github.com/users/finiteautomata/events{/privacy}", "followers_url": "https://api.github.com/users/finiteautomata/followers", "following_url": "https://api.github.com/users/finiteautomata/following{/other_user}", "gists_url": "https://api.github.com/users/finiteautomata/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/finiteautomata", "id": 167943, "login": "finiteautomata", "node_id": "MDQ6VXNlcjE2Nzk0Mw==", "organizations_url": "https://api.github.com/users/finiteautomata/orgs", "received_events_url": "https://api.github.com/users/finiteautomata/received_events", "repos_url": "https://api.github.com/users/finiteautomata/repos", "site_admin": false, "starred_url": "https://api.github.com/users/finiteautomata/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/finiteautomata/subscriptions", "type": "User", "url": "https://api.github.com/users/finiteautomata", "user_view_type": "public" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" } ]
null
[ "Thanks for reporting, @finiteautomata.\r\n\r\nWe are investigating it. ", "There is a bug in `datasets`. You can see our proposed fix:\r\n- #6309 " ]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug I'm trying to load [piuba-bigdata/articles_and_comments] and I'm stumbling with this error on `2.14.5`. However, this works on `2.10.0`. ### Steps to reproduce the bug [Colab link](https://colab.research.google.com/drive/1SAftFMQnFE708ikRnJJHIXZV7R5IBOCE#scrollTo=r2R2ipCCDmsg) ```python Downloading readme: 100% 1.19k/1.19k [00:00<00:00, 30.9kB/s] --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) [<ipython-input-2-807c3583d297>](https://localhost:8080/#) in <cell line: 3>() 1 from datasets import load_dataset 2 ----> 3 load_dataset("piuba-bigdata/articles_and_comments", split="train") 2 frames [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://localhost:8080/#) in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, ignore_verifications, keep_in_memory, save_infos, revision, token, use_auth_token, task, streaming, num_proc, storage_options, **config_kwargs) 2127 2128 # Create a dataset builder -> 2129 builder_instance = load_dataset_builder( 2130 path=path, 2131 name=name, [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://localhost:8080/#) in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, token, use_auth_token, storage_options, **config_kwargs) 1813 download_config = download_config.copy() if download_config else DownloadConfig() 1814 download_config.storage_options.update(storage_options) -> 1815 dataset_module = dataset_module_factory( 1816 path, 1817 revision=revision, [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://localhost:8080/#) in dataset_module_factory(path, revision, download_config, download_mode, dynamic_modules_path, data_dir, data_files, **download_kwargs) 1506 raise e1 from None 1507 if isinstance(e1, FileNotFoundError): -> 1508 raise FileNotFoundError( 1509 f"Couldn't find a dataset script at {relative_to_absolute_path(combined_path)} or any data file in the same directory. " 1510 f"Couldn't find '{path}' on the Hugging Face Hub either: {type(e1).__name__}: {e1}" FileNotFoundError: Couldn't find a dataset script at /content/piuba-bigdata/articles_and_comments/articles_and_comments.py or any data file in the same directory. Couldn't find 'piuba-bigdata/articles_and_comments' on the Hugging Face Hub either: FileNotFoundError: No (supported) data files or dataset script found in piuba-bigdata/articles_and_comments. ``` ### Expected behavior It should load normally. ### Environment info ``` - `datasets` version: 2.14.5 - Platform: Linux-5.15.120+-x86_64-with-glibc2.35 - Python version: 3.10.12 - Huggingface_hub version: 0.18.0 - PyArrow version: 9.0.0 - Pandas version: 1.5.3 ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6305/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6305/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6303
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6303/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6303/comments
https://api.github.com/repos/huggingface/datasets/issues/6303/events
https://github.com/huggingface/datasets/issues/6303
1,943,466,532
I_kwDODunzps5z1vIk
6,303
Parquet uploads off-by-one naming scheme
{ "avatar_url": "https://avatars.githubusercontent.com/u/1981179?v=4", "events_url": "https://api.github.com/users/ZachNagengast/events{/privacy}", "followers_url": "https://api.github.com/users/ZachNagengast/followers", "following_url": "https://api.github.com/users/ZachNagengast/following{/other_user}", "gists_url": "https://api.github.com/users/ZachNagengast/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZachNagengast", "id": 1981179, "login": "ZachNagengast", "node_id": "MDQ6VXNlcjE5ODExNzk=", "organizations_url": "https://api.github.com/users/ZachNagengast/orgs", "received_events_url": "https://api.github.com/users/ZachNagengast/received_events", "repos_url": "https://api.github.com/users/ZachNagengast/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZachNagengast/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZachNagengast/subscriptions", "type": "User", "url": "https://api.github.com/users/ZachNagengast", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "You can find the reasoning behind this naming scheme [here](https://github.com/huggingface/transformers/pull/16343#discussion_r931182168).\r\n\r\nThis point has been raised several times, so I'd be okay with starting with `00001-` (also to be consistent with the `transformers` sharding), but I'm not sure @lhoestq agrees.", "We start at 0 in `datasets` for consistency with Apache Spark, Apache Beam, Dask and others.\r\n\r\nAlso note `transformers` isn't a good reference on this topic. I talked with the maintainers when they added shards but it was already released this way. Though we found that there is a backward-compatible way in `transformers` to start at 0, but no request from `transformers` users to changes this AFAIK.", "not sure it would be a good idea to break the consistency now, IMO", "Makes sense to start at 0 for plenty of good reasons so I'm on board.\r\n\r\nWhat about the second part `-of-0000X`? With single commit PR #6269 just getting merged, there was a note about issues with 100+ file edits https://github.com/huggingface/datasets/pull/6269#issuecomment-1755428581.\r\n\r\nThat would be my last remaining concern in the context of the `push_to_hub(..., append=True)` work to be done, where appending a single file to the full dataset will require renaming every other existing file in the dataset. If it doesn't seem like a big issue for this work then all the better 👍" ]
1970-01-01T00:00:00.000001
1,697
null
CONTRIBUTOR
null
### Describe the bug I noticed this numbering scheme not matching up in a different project and wanted to raise it as an issue for discussion, what is the actual proper way to have these stored? <img width="425" alt="image" src="https://github.com/huggingface/datasets/assets/1981179/3ffa2144-7c9a-446f-b521-a5e9db71e7ce"> The `-SSSSS-of-NNNNN` seems to be used widely across the codebase. The section that creates the part in my screenshot is here https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L5287 There are also some edits to this section in the single commit branch. ### Steps to reproduce the bug 1. Upload a dataset that requires at least two parquet files in it 2. Observe the naming scheme ### Expected behavior The couple options here are of course **1. keeping it as is** **2. Starting the index at 1:** train-00001-of-00002-{hash}.parquet train-00002-of-00002-{hash}.parquet **3. My preferred option** (which would solve my specific issue), dropping the total entirely: train-00000-{hash}.parquet train-00001-{hash}.parquet This also solves an issue that will occur with an `append` variable for `push_to_hub` (see https://github.com/huggingface/datasets/issues/6290) where as you add a new parquet file, you need to rename everything in the repo as well. However, I know there are parts of the repo that use 0 as the starting file or may require the total, so raising the question for discussion. ### Environment info - `datasets` version: 2.14.6.dev0 - Platform: macOS-14.0-arm64-arm-64bit - Python version: 3.10.12 - Huggingface_hub version: 0.18.0 - PyArrow version: 12.0.1 - Pandas version: 1.5.3
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6303/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6303/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6302
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6302/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6302/comments
https://api.github.com/repos/huggingface/datasets/issues/6302/events
https://github.com/huggingface/datasets/issues/6302
1,942,096,078
I_kwDODunzps5zwgjO
6,302
ArrowWriter/ParquetWriter `write` method does not increase `_num_bytes` and hence datasets not sharding at `max_shard_size`
{ "avatar_url": "https://avatars.githubusercontent.com/u/2855550?v=4", "events_url": "https://api.github.com/users/Rassibassi/events{/privacy}", "followers_url": "https://api.github.com/users/Rassibassi/followers", "following_url": "https://api.github.com/users/Rassibassi/following{/other_user}", "gists_url": "https://api.github.com/users/Rassibassi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rassibassi", "id": 2855550, "login": "Rassibassi", "node_id": "MDQ6VXNlcjI4NTU1NTA=", "organizations_url": "https://api.github.com/users/Rassibassi/orgs", "received_events_url": "https://api.github.com/users/Rassibassi/received_events", "repos_url": "https://api.github.com/users/Rassibassi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rassibassi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rassibassi/subscriptions", "type": "User", "url": "https://api.github.com/users/Rassibassi", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "`writer._num_bytes` is updated every `writer_batch_size`-th call to the `write` method (default `writer_batch_size` is 1000 (examples)). You should be able to see the update by passing a smaller `writer_batch_size` to the `load_dataset_builder`.\r\n\r\nWe could improve this by supporting the string `writer_batch_size` version as we do with `max_shard_size`, and capping `writer_batch_size` to `max_shard_size` in scenarios where the default `writer_batch_size` > `max_shard_size`. ", "Thanks, reducing `writer_batch_size` solved my problem :)" ]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug An example from [1], does not work when limiting shards with `max_shard_size`. Try the following example with low `max_shard_size`, such as: ```python builder.download_and_prepare(output_dir, storage_options=storage_options, file_format="parquet", max_shard_size="10MB") ``` The reason for this is that, in line [2] `writer._num_bytes > max_shard_size` is never true, because the `write` method of `ArrowWriter` [3] does not increase `self._num_bytes`. Such that respective Arrow/Parquet shards are only written to file based on the `writer_batch_size` or `config.DEFAULT_MAX_BATCH_SIZE`, but not based on `max_shard_size`. [1] https://huggingface.co/docs/datasets/filesystems#download-and-prepare-a-dataset-into-a-cloud-storage [2] https://github.com/huggingface/datasets/blob/3e8d420808718c9a1453a2e7ee3484ca12c9c70d/src/datasets/builder.py#L1677 [3] https://github.com/huggingface/datasets/blob/3e8d420808718c9a1453a2e7ee3484ca12c9c70d/src/datasets/arrow_writer.py#L459 ### Steps to reproduce the bug Get example from: https://huggingface.co/docs/datasets/filesystems#download-and-prepare-a-dataset-into-a-cloud-storage Call `builder.download_and_prepare` with low `max_shard_size` such as `10MB`, e.g.: ```python builder.download_and_prepare(output_dir, storage_options=storage_options, file_format="parquet", max_shard_size="10MB") ``` ### Expected behavior Shards should be written based on `max_shard_size` instead of batch size. ### Environment info ``` >>> import datasets >>> datasets.__version__ '2.14.6.dev0 ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/2855550?v=4", "events_url": "https://api.github.com/users/Rassibassi/events{/privacy}", "followers_url": "https://api.github.com/users/Rassibassi/followers", "following_url": "https://api.github.com/users/Rassibassi/following{/other_user}", "gists_url": "https://api.github.com/users/Rassibassi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rassibassi", "id": 2855550, "login": "Rassibassi", "node_id": "MDQ6VXNlcjI4NTU1NTA=", "organizations_url": "https://api.github.com/users/Rassibassi/orgs", "received_events_url": "https://api.github.com/users/Rassibassi/received_events", "repos_url": "https://api.github.com/users/Rassibassi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rassibassi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rassibassi/subscriptions", "type": "User", "url": "https://api.github.com/users/Rassibassi", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6302/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6302/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6299
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6299/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6299/comments
https://api.github.com/repos/huggingface/datasets/issues/6299/events
https://github.com/huggingface/datasets/issues/6299
1,939,649,238
I_kwDODunzps5znLLW
6,299
Support for newer versions of JAX
{ "avatar_url": "https://avatars.githubusercontent.com/u/25456859?v=4", "events_url": "https://api.github.com/users/ddrous/events{/privacy}", "followers_url": "https://api.github.com/users/ddrous/followers", "following_url": "https://api.github.com/users/ddrous/following{/other_user}", "gists_url": "https://api.github.com/users/ddrous/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ddrous", "id": 25456859, "login": "ddrous", "node_id": "MDQ6VXNlcjI1NDU2ODU5", "organizations_url": "https://api.github.com/users/ddrous/orgs", "received_events_url": "https://api.github.com/users/ddrous/received_events", "repos_url": "https://api.github.com/users/ddrous/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ddrous/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ddrous/subscriptions", "type": "User", "url": "https://api.github.com/users/ddrous", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
NONE
null
### Feature request Hi, I like your idea of adapting the datasets library to be usable with JAX. Thank you for that. However, in your [setup.py](https://github.com/huggingface/datasets/blob/main/setup.py), you enforce old versions of JAX <= 0.3... It is very cumbersome ! What is the rationale for such a limitation ? Can you remove it please ? Thanks, ### Motivation This library is unusable with new versions of JAX ? ### Your contribution Yes.
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6299/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6299/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6294
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6294/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6294/comments
https://api.github.com/repos/huggingface/datasets/issues/6294/events
https://github.com/huggingface/datasets/issues/6294
1,937,359,605
I_kwDODunzps5zecL1
6,294
IndexError: Invalid key is out of bounds for size 0 despite having a populated dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/61892155?v=4", "events_url": "https://api.github.com/users/ZYM66/events{/privacy}", "followers_url": "https://api.github.com/users/ZYM66/followers", "following_url": "https://api.github.com/users/ZYM66/following{/other_user}", "gists_url": "https://api.github.com/users/ZYM66/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZYM66", "id": 61892155, "login": "ZYM66", "node_id": "MDQ6VXNlcjYxODkyMTU1", "organizations_url": "https://api.github.com/users/ZYM66/orgs", "received_events_url": "https://api.github.com/users/ZYM66/received_events", "repos_url": "https://api.github.com/users/ZYM66/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZYM66/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZYM66/subscriptions", "type": "User", "url": "https://api.github.com/users/ZYM66", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "It looks to be the same issue as the one reported in https://discuss.huggingface.co/t/indexerror-invalid-key-16-is-out-of-bounds-for-size-0.\r\n\r\nCan you check the length of `train_dataset` before the `train_sampler = self._get_train_sampler()` (and after `_remove_unused_columns`) line?" ]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug I am encountering an `IndexError` when trying to access data from a DataLoader which wraps around a dataset I've loaded using the `datasets` library. The error suggests that the dataset size is `0`, but when I check the length and print the dataset, it's clear that it has `1166` entries. ### Steps to reproduce the bug 1. Load a dataset with `1166` entries. 2. Create a DataLoader using this dataset. 3. Try iterating over the DataLoader. code: ```python def get_train_dataloader(self) -> DataLoader: if self.train_dataset is None: raise ValueError("Trainer: training requires a train_dataset.") train_dataset = self.train_dataset data_collator = self.data_collator print(len(train_dataset)) print(train_dataset) if is_datasets_available() and isinstance(train_dataset, datasets.Dataset): train_dataset = self._remove_unused_columns(train_dataset, description="training") else: data_collator = self._get_collator_with_removed_columns(data_collator, description="training") train_sampler = self._get_train_sampler() dl = DataLoader( train_dataset, batch_size=self._train_batch_size, sampler=train_sampler, collate_fn=data_collator, drop_last=self.args.dataloader_drop_last, num_workers=self.args.dataloader_num_workers, pin_memory=self.args.dataloader_pin_memory, worker_init_fn=seed_worker, ) print(dl) print(len(dl)) for i in dl: print(i) break return dl ``` output : ``` 1166 Dataset({ features: ['input_ids', 'special_tokens_mask'], num_rows: 1166 }) <torch.utils.data.dataloader.DataLoader object ...> 146 ``` Error: ``` Traceback (most recent call last): File "/home/dl/zym/llamaJP/TestUseContinuePretrainLlama.py", line 266, in <module> train() File "/home/dl/zym/llamaJP/TestUseContinuePretrainLlama.py", line 260, in train trainer.train() File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/transformers/trainer.py", line 1506, in train return inner_training_loop( File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/transformers/trainer.py", line 1520, in _inner_training_loop train_dataloader = self.get_train_dataloader() File "/home/dl/zym/llamaJP/TestUseContinuePretrainLlama.py", line 80, in get_train_dataloader for i in dl: File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 630, in __next__ data = self._next_data() File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 674, in _next_data data = self._dataset_fetcher.fetch(index) # may raise StopIteration File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/torch/utils/data/_utils/fetch.py", line 49, in fetch data = self.dataset.__getitems__(possibly_batched_index) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2807, in __getitems__ batch = self.__getitem__(keys) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2803, in __getitem__ return self._getitem(key) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2787, in _getitem pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/formatting/formatting.py", line 583, in query_table _check_valid_index_key(key, size) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/formatting/formatting.py", line 536, in _check_valid_index_key _check_valid_index_key(int(max(key)), size=size) File "/root/miniconda3/envs/LLM/lib/python3.10/site-packages/datasets/formatting/formatting.py", line 526, in _check_valid_index_key raise IndexError(f"Invalid key: {key} is out of bounds for size {size}") IndexError: Invalid key: 1116 is out of bounds for size 0 ``` ### Expected behavior I expect to be able to iterate over the DataLoader without encountering an IndexError since the dataset is populated. ### Environment info - `datasets` library version: [2.14.5] - Platform: [Linux] - Python version: 3.10 - Other libraries involved: HuggingFace Transformers
{ "avatar_url": "https://avatars.githubusercontent.com/u/61892155?v=4", "events_url": "https://api.github.com/users/ZYM66/events{/privacy}", "followers_url": "https://api.github.com/users/ZYM66/followers", "following_url": "https://api.github.com/users/ZYM66/following{/other_user}", "gists_url": "https://api.github.com/users/ZYM66/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZYM66", "id": 61892155, "login": "ZYM66", "node_id": "MDQ6VXNlcjYxODkyMTU1", "organizations_url": "https://api.github.com/users/ZYM66/orgs", "received_events_url": "https://api.github.com/users/ZYM66/received_events", "repos_url": "https://api.github.com/users/ZYM66/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZYM66/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZYM66/subscriptions", "type": "User", "url": "https://api.github.com/users/ZYM66", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6294/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6294/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6293
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6293/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6293/comments
https://api.github.com/repos/huggingface/datasets/issues/6293/events
https://github.com/huggingface/datasets/issues/6293
1,937,238,047
I_kwDODunzps5zd-gf
6,293
Choose columns to stream parquet data in streaming mode
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
MEMBER
null
Currently passing columns= to load_dataset in streaming mode fails ``` Tried to load parquet data with columns '['link']' with mismatching features '{'caption': Value(dtype='string', id=None), 'image': {'bytes': Value(dtype='binary', id=None), 'path': Value(dtype='null', id=None)}, 'link': Value(dtype='string', id=None), 'message_id': Value(dtype='string', id=None), 'timestamp': Value(dtype='string', id=None)}' ``` similar to https://github.com/huggingface/datasets/issues/6039 reported at https://huggingface.co/datasets/laion/dalle-3-dataset/discussions/3#65259a09617407d4520f4ad9
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6293/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6293/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6292
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6292/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6292/comments
https://api.github.com/repos/huggingface/datasets/issues/6292/events
https://github.com/huggingface/datasets/issues/6292
1,937,050,470
I_kwDODunzps5zdQtm
6,292
how to load the image of dtype float32 or float64
{ "avatar_url": "https://avatars.githubusercontent.com/u/26437644?v=4", "events_url": "https://api.github.com/users/wanglaofei/events{/privacy}", "followers_url": "https://api.github.com/users/wanglaofei/followers", "following_url": "https://api.github.com/users/wanglaofei/following{/other_user}", "gists_url": "https://api.github.com/users/wanglaofei/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wanglaofei", "id": 26437644, "login": "wanglaofei", "node_id": "MDQ6VXNlcjI2NDM3NjQ0", "organizations_url": "https://api.github.com/users/wanglaofei/orgs", "received_events_url": "https://api.github.com/users/wanglaofei/received_events", "repos_url": "https://api.github.com/users/wanglaofei/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wanglaofei/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wanglaofei/subscriptions", "type": "User", "url": "https://api.github.com/users/wanglaofei", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Hi! Can you provide a code that reproduces the issue?\r\n\r\nAlso, which version of `datasets` are you using? You can check this by running `python -c \"import datasets; print(datasets.__version__)\"` inside the env. We added support for \"float images\" in `datasets 2.9`." ]
1970-01-01T00:00:00.000001
1,697
null
NONE
null
_FEATURES = datasets.Features( { "image": datasets.Image(), "text": datasets.Value("string"), }, ) The datasets builder seems only support the unit8 data. How to load the float dtype data?
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6292/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6292/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6291
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6291/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6291/comments
https://api.github.com/repos/huggingface/datasets/issues/6291/events
https://github.com/huggingface/datasets/issues/6291
1,936,129,871
I_kwDODunzps5zZv9P
6,291
Casting type from Array2D int to Array2D float crashes
{ "avatar_url": "https://avatars.githubusercontent.com/u/22567306?v=4", "events_url": "https://api.github.com/users/AlanBlanchet/events{/privacy}", "followers_url": "https://api.github.com/users/AlanBlanchet/followers", "following_url": "https://api.github.com/users/AlanBlanchet/following{/other_user}", "gists_url": "https://api.github.com/users/AlanBlanchet/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AlanBlanchet", "id": 22567306, "login": "AlanBlanchet", "node_id": "MDQ6VXNlcjIyNTY3MzA2", "organizations_url": "https://api.github.com/users/AlanBlanchet/orgs", "received_events_url": "https://api.github.com/users/AlanBlanchet/received_events", "repos_url": "https://api.github.com/users/AlanBlanchet/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AlanBlanchet/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AlanBlanchet/subscriptions", "type": "User", "url": "https://api.github.com/users/AlanBlanchet", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "Thanks for reporting! I've opened a PR with a fix" ]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug I am on a school project and the initial type for feature annotations are `Array2D(shape=(None, 4))`. I am trying to cast this type to a `float64` and pyarrow gives me this error : ``` Traceback (most recent call last): File "/home/alan/dev/ClassezDesImagesAvecDesAlgorithmesDeDeeplearning/src/sdd/data/dataset.py", line 141, in <module> dataset = StanfordDogsDataset(size, 5).original(True).demo() File "<attrs generated init __main__.StanfordDogsDataset>", line 4, in __init__ File "/home/alan/dev/ClassezDesImagesAvecDesAlgorithmesDeDeeplearning/src/sdd/data/dataset.py", line 33, in __attrs_post_init__ self.dataset = self.dataset.cast_column( File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/fingerprint.py", line 511, in wrapper out = func(dataset, *args, **kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2110, in cast_column return self.cast(features) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2055, in cast dataset = dataset.map( File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 592, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 3097, in map for rank, done, content in Dataset._map_single(**dataset_kwargs): File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 3474, in _map_single batch = apply_function_on_filtered_inputs( File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 3353, in apply_function_on_filtered_inputs processed_inputs = function(*fn_args, *additional_args, **fn_kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 2328, in table_cast return cast_table_to_schema(table, schema) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 2287, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 2287, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 1831, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 1831, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 2143, in cast_array_to_feature return array_cast(array, feature(), allow_number_to_str=allow_number_to_str) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 1833, in wrapper return func(array, *args, **kwargs) File "/home/alan/.cache/pypoetry/virtualenvs/sdd-2XWLAjSi-py3.10/lib/python3.10/site-packages/datasets/table.py", line 1967, in array_cast return pa_type.wrap_array(array) File "pyarrow/types.pxi", line 1369, in pyarrow.lib.BaseExtensionType.wrap_array TypeError: Incompatible storage type for extension<arrow.py_extension_type<Array2DExtensionType>>: expected list<item: list<item: double>>, got list<item: list<item: int32>> ``` ### Steps to reproduce the bug ```python dataset = datasets.load_dataset("Alanox/stanford-dogs", split="full") dataset = dataset.cast_column("annotations", Array2D((None, 4), "float64")) ``` ### Expected behavior It should simply cast the column feature type to a `float64` without error ### Environment info datasets == 2.14.5
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6291/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6291/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6290
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6290/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6290/comments
https://api.github.com/repos/huggingface/datasets/issues/6290/events
https://github.com/huggingface/datasets/issues/6290
1,935,629,679
I_kwDODunzps5zX11v
6,290
Incremental dataset (e.g. `.push_to_hub(..., append=True)`)
{ "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Wauplin", "id": 11801849, "login": "Wauplin", "node_id": "MDQ6VXNlcjExODAxODQ5", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "repos_url": "https://api.github.com/users/Wauplin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "type": "User", "url": "https://api.github.com/users/Wauplin", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[ "Yea I think waiting for #6269 would be best, or branching from it. For reference, this [PR](https://github.com/LAION-AI/Discord-Scrapers/pull/2) is progressing pretty well which will do similar using the hf hub for our LAION dataset bot https://github.com/LAION-AI/Discord-Scrapers/pull/2. ", "Is there any update on this?", "Is there any update on this?", "No update so far on this feature request but for broader context, this announce will help with incremental datasets https://huggingface.co/blog/xethub-joins-hf :)" ]
1970-01-01T00:00:00.000001
1,725
null
CONTRIBUTOR
null
### Feature request Have the possibility to do `ds.push_to_hub(..., append=True)`. ### Motivation Requested in this [comment](https://huggingface.co/datasets/laion/dalle-3-dataset/discussions/3#65252597c4edc168202a5eaa) and this [comment](https://huggingface.co/datasets/laion/dalle-3-dataset/discussions/4#6524f675c9607bdffb208d8f). Discussed internally on [slack](https://huggingface.slack.com/archives/C02EMARJ65P/p1696950642610639?thread_ts=1690554266.830949&cid=C02EMARJ65P). ### Your contribution What I suggest to do for parquet datasets is to use `CommitOperationCopy` + `CommitOperationDelete` from `huggingface_hub`: 1. list files 2. copy files from parquet-0001-of-0004 to parquet-0001-of-0005 3. delete files like parquet-0001-of-0004 4. generate + add last parquet file parquet-0005-of-0005 => make a single commit with all commit operations at once I think it should be quite straightforward to implement. Happy to review a PR (maybe conflicting with the ongoing "1 commit push_to_hub" PR https://github.com/huggingface/datasets/pull/6269)
null
{ "+1": 5, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 5, "url": "https://api.github.com/repos/huggingface/datasets/issues/6290/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6290/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6288
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6288/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6288/comments
https://api.github.com/repos/huggingface/datasets/issues/6288/events
https://github.com/huggingface/datasets/issues/6288
1,935,005,457
I_kwDODunzps5zVdcR
6,288
Dataset.from_pandas with a DataFrame of PIL.Images
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[ "A duplicate of https://github.com/huggingface/datasets/issues/4796.\r\n\r\nWe could get this for free by implementing the `Image` feature as an extension type, as shown in [this](https://colab.research.google.com/drive/1Uzm_tXVpGTwbzleDConWcNjacwO1yxE4?usp=sharing) Colab (example with UUIDs).\r\n", "+1 to this\r\nCalling this line with a df that contains a PIL image (as they are returned from load_dataset)\r\n`ds = Dataset.from_pandas(df)`\r\nResults in this error:\r\n`ArrowInvalid: ('Could not convert <PIL.PngImagePlugin.PngImageFile image mode=RGB size=1024x1024 at 0x2B41F2D70> with type PngImageFile: did not recognize Python value type when inferring an Arrow data type', 'Conversion failed for column image with type object')`" ]
1970-01-01T00:00:00.000001
1,697
null
MEMBER
null
Currently type inference doesn't know what to do with a Pandas Series of PIL.Image objects, though it would be nice to get a Dataset with the Image type this way
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6288/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6288/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6287
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6287/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6287/comments
https://api.github.com/repos/huggingface/datasets/issues/6287/events
https://github.com/huggingface/datasets/issues/6287
1,932,758,192
I_kwDODunzps5zM4yw
6,287
map() not recognizing "text"
{ "avatar_url": "https://avatars.githubusercontent.com/u/5688359?v=4", "events_url": "https://api.github.com/users/EngineerKhan/events{/privacy}", "followers_url": "https://api.github.com/users/EngineerKhan/followers", "following_url": "https://api.github.com/users/EngineerKhan/following{/other_user}", "gists_url": "https://api.github.com/users/EngineerKhan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/EngineerKhan", "id": 5688359, "login": "EngineerKhan", "node_id": "MDQ6VXNlcjU2ODgzNTk=", "organizations_url": "https://api.github.com/users/EngineerKhan/orgs", "received_events_url": "https://api.github.com/users/EngineerKhan/received_events", "repos_url": "https://api.github.com/users/EngineerKhan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/EngineerKhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EngineerKhan/subscriptions", "type": "User", "url": "https://api.github.com/users/EngineerKhan", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "There is no \"text\" column in the `amazon_reviews_multi`, hence the `KeyError`. You can get the column names by running `dataset.column_names`." ]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug The [map() documentation](https://huggingface.co/docs/datasets/v2.14.5/en/package_reference/main_classes#datasets.Dataset.map) reads: ` ds = ds.map(lambda x: tokenizer(x['text'], truncation=True, padding=True), batched=True)` I have been trying to reproduce it in my code as: `tokenizedDataset = dataset.map(lambda x: tokenizer(x['text']), batched=True)` But it doesn't work as it throws the error: > KeyError: 'text' Can you please guide me on how to fix it? ### Steps to reproduce the bug 1. `from datasets import load_dataset dataset = load_dataset("amazon_reviews_multi")` 2. Then this code: `from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("bert-base-cased")` 3. The line I quoted above (which I have been trying) ### Expected behavior As mentioned in the documentation, it should run without any error and map the tokenization on the whole dataset. ### Environment info Python 3.10.2
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6287/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6287/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6285
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6285/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6285/comments
https://api.github.com/repos/huggingface/datasets/issues/6285/events
https://github.com/huggingface/datasets/issues/6285
1,932,306,325
I_kwDODunzps5zLKeV
6,285
TypeError: expected str, bytes or os.PathLike object, not dict
{ "avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4", "events_url": "https://api.github.com/users/andysingal/events{/privacy}", "followers_url": "https://api.github.com/users/andysingal/followers", "following_url": "https://api.github.com/users/andysingal/following{/other_user}", "gists_url": "https://api.github.com/users/andysingal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/andysingal", "id": 20493493, "login": "andysingal", "node_id": "MDQ6VXNlcjIwNDkzNDkz", "organizations_url": "https://api.github.com/users/andysingal/orgs", "received_events_url": "https://api.github.com/users/andysingal/received_events", "repos_url": "https://api.github.com/users/andysingal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/andysingal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andysingal/subscriptions", "type": "User", "url": "https://api.github.com/users/andysingal", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "You should be able to load the images by modifying the `load_dataset` call like this:\r\n```python\r\ndataset = load_dataset(\"imagefolder\", data_dir=\"/content/datasets/PotholeDetectionYOLOv8-1\")\r\n```\r\n\r\nThe `imagefolder` builder expects the image files to be in `path/label/image_file` (e.g. .`.../train/dog/image_1.jpg`), so the solution for the labels in your case is to create metadata files (one for each split; as explained [here](https://huggingface.co/docs/datasets/image_dataset#imagefolder)) that map the images to their labels.", "> You should be able to load the images by modifying the `load_dataset` call like this:\r\n> \r\n> ```python\r\n> dataset = load_dataset(\"imagefolder\", data_dir=\"/content/datasets/PotholeDetectionYOLOv8-1\")\r\n> ```\r\n> \r\n> The `imagefolder` builder expects the image files to be in `path/label/image_file` (e.g. .`.../train/dog/image_1.jpg`), so the solution for the labels in your case is to create metadata files (one for each split; as explained [here](https://huggingface.co/docs/datasets/image_dataset#imagefolder)) that map the images to their labels.\r\n\r\nI tried like this but only uploads images and not labels, Andyrasika/potholes-dataset", "As explained in my previous comment, you need to define metadata files to load the labels or update the paths to be in the format `train/label/image` (`train- image /n -labels` is not supported by the loader).", "I downloaded my file after annotating using roboflow . It gives train-\r\nimages, labels , test- images, labels , valid- images, labels . I hope it\r\ngives you an idea of the dataset . Please advise on this dataset\r\n\r\nOn Tue, Oct 10, 2023 at 18:12 Mario Šaško ***@***.***> wrote:\r\n\r\n> As explained in my previous comment, you need to define metadata files to\r\n> load the labels or update the paths to be in the format train/label/image\r\n> (train- image /n -labels is not supported by the loader).\r\n>\r\n> —\r\n> Reply to this email directly, view it on GitHub\r\n> <https://github.com/huggingface/datasets/issues/6285#issuecomment-1755335215>,\r\n> or unsubscribe\r\n> <https://github.com/notifications/unsubscribe-auth/AE4LJNN56FWWTSBYTSTUWHLX6U7CVAVCNFSM6AAAAAA5YHCSTGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONJVGMZTKMRRGU>\r\n> .\r\n> You are receiving this because you authored the thread.Message ID:\r\n> ***@***.***>\r\n>\r\n" ]
1970-01-01T00:00:00.000001
1,696
null
NONE
null
### Describe the bug my dataset is in form : train- image /n -labels and tried the code: ``` from datasets import load_dataset data_files = { "train": "/content/datasets/PotholeDetectionYOLOv8-1/train/", "validation": "/content/datasets/PotholeDetectionYOLOv8-1/valid/", "test": "/content/datasets/PotholeDetectionYOLOv8-1/test/" } dataset = load_dataset("imagefolder", data_dir=data_files) dataset ``` got error: ``` --------------------------------------------------------------------------- TypeError Traceback (most recent call last) [<ipython-input-29-2ef1926f73d9>](https://localhost:8080/#) in <cell line: 8>() 6 "test": "/content/datasets/PotholeDetectionYOLOv8-1/test/" 7 } ----> 8 dataset = load_dataset("imagefolder", data_dir=data_files) 9 dataset 6 frames [/usr/lib/python3.10/pathlib.py](https://localhost:8080/#) in _parse_args(cls, args) 576 parts += a._parts 577 else: --> 578 a = os.fspath(a) 579 if isinstance(a, str): 580 # Force-cast str subclasses to str (issue #21127) TypeError: expected str, bytes or os.PathLike object, not dict ``` ### Steps to reproduce the bug as share above ### Expected behavior load images and labels , but my dataset only uploads images - https://huggingface.co/datasets/Andyrasika/potholes-dataset ### Environment info colab pro
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6285/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6285/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6284
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6284/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6284/comments
https://api.github.com/repos/huggingface/datasets/issues/6284/events
https://github.com/huggingface/datasets/issues/6284
1,929,551,712
I_kwDODunzps5zAp9g
6,284
Add Belebele multiple-choice machine reading comprehension (MRC) dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/64583161?v=4", "events_url": "https://api.github.com/users/rajveer43/events{/privacy}", "followers_url": "https://api.github.com/users/rajveer43/followers", "following_url": "https://api.github.com/users/rajveer43/following{/other_user}", "gists_url": "https://api.github.com/users/rajveer43/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rajveer43", "id": 64583161, "login": "rajveer43", "node_id": "MDQ6VXNlcjY0NTgzMTYx", "organizations_url": "https://api.github.com/users/rajveer43/orgs", "received_events_url": "https://api.github.com/users/rajveer43/received_events", "repos_url": "https://api.github.com/users/rajveer43/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rajveer43/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rajveer43/subscriptions", "type": "User", "url": "https://api.github.com/users/rajveer43", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[ "This dataset is already available on the Hub: https://huggingface.co/datasets/facebook/belebele.\r\n" ]
1970-01-01T00:00:00.000001
1,696
1970-01-01T00:00:00.000001
NONE
null
### Feature request Belebele is a multiple-choice machine reading comprehension (MRC) dataset spanning 122 language variants. This dataset enables the evaluation of mono- and multi-lingual models in high-, medium-, and low-resource languages. Each question has four multiple-choice answers and is linked to a short passage from the [FLORES-200](https://github.com/facebookresearch/flores/tree/main/flores200) dataset. The human annotation procedure was carefully curated to create questions that discriminate between different levels of generalizable language comprehension and is reinforced by extensive quality checks. While all questions directly relate to the passage, the English dataset on its own proves difficult enough to challenge state-of-the-art language models. Being fully parallel, this dataset enables direct comparison of model performance across all languages. Belebele opens up new avenues for evaluating and analyzing the multilingual abilities of language models and NLP systems. Please refer to paper for more details, [The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants](https://arxiv.org/abs/2308.16884). ## Composition - 900 questions per language variant - 488 distinct passages, there are 1-2 associated questions for each. - For each question, there is 4 multiple-choice answers, exactly 1 of which is correct. - 122 language/language variants (including English). - 900 x 122 = 109,800 total questions. ### Motivation official repo https://github.com/facebookresearch/belebele ### Your contribution -
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6284/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6284/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6280
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6280/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6280/comments
https://api.github.com/repos/huggingface/datasets/issues/6280/events
https://github.com/huggingface/datasets/issues/6280
1,928,215,278
I_kwDODunzps5y7jru
6,280
Couldn't cast array of type fixed_size_list to Sequence(Value(float64))
{ "avatar_url": "https://avatars.githubusercontent.com/u/1000442?v=4", "events_url": "https://api.github.com/users/jmif/events{/privacy}", "followers_url": "https://api.github.com/users/jmif/followers", "following_url": "https://api.github.com/users/jmif/following{/other_user}", "gists_url": "https://api.github.com/users/jmif/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jmif", "id": 1000442, "login": "jmif", "node_id": "MDQ6VXNlcjEwMDA0NDI=", "organizations_url": "https://api.github.com/users/jmif/orgs", "received_events_url": "https://api.github.com/users/jmif/received_events", "repos_url": "https://api.github.com/users/jmif/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jmif/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmif/subscriptions", "type": "User", "url": "https://api.github.com/users/jmif", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "Thanks for reporting! I've opened a PR with a fix.", "Thanks for the quick response @mariosasko! I just installed your branch via `poetry add 'git+https://github.com/huggingface/datasets#fix-array_values'` and I can confirm it works on the example provided.\r\n\r\nFollow up question for you, should `None`s be supported in these types of features as they are in others?\r\n\r\nFor example, the following script:\r\n\r\n```\r\nfrom datasets import Features, Value, Sequence, ClassLabel, Dataset\r\n\r\ndataset_features = Features({\r\n 'text': Value('string'),\r\n 'embedding': Sequence(Value('double'), length=2),\r\n 'categories': Sequence(ClassLabel(names=sorted([\r\n 'one',\r\n 'two',\r\n 'three'\r\n ]))),\r\n})\r\n\r\ndataset = Dataset.from_dict(\r\n {\r\n 'text': ['A'] * 10000,\r\n \"embedding\": [None] * 10000, # THIS LINE CHANGED\r\n 'categories': [[0]] * 10000,\r\n },\r\n features=dataset_features\r\n)\r\n\r\ndef test_mapper(r):\r\n r['text'] = list(map(lambda t: t + ' b', r['text']))\r\n return r\r\n\r\n\r\ndataset = dataset.map(test_mapper, batched=True, batch_size=10, features=dataset_features, num_proc=2)\r\n```\r\n\r\nfails with\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"/home/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/multiprocess/pool.py\", line 125, in worker\r\n result = (True, func(*args, **kwds))\r\n File \"/home/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/utils/py_utils.py\", line 1354, in _write_generator_to_queue\r\n for i, result in enumerate(func(**kwargs)):\r\n File \"/home/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/arrow_dataset.py\", line 3493, in _map_single\r\n writer.write_batch(batch)\r\n File \"/home/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/arrow_writer.py\", line 549, in write_batch\r\n array = cast_array_to_feature(col_values, col_type) if col_type is not None else col_values\r\n File \"/home/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/table.py\", line 1831, in wrapper\r\n return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])\r\n File \"/home/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/table.py\", line 1831, in <listcomp>\r\n return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])\r\n File \"/home/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/table.py\", line 2160, in cast_array_to_feature\r\n raise TypeError(f\"Couldn't cast array of type\\n{array.type}\\nto\\n{feature}\")\r\nTypeError: Couldn't cast array of type\r\nfixed_size_list<item: double>[2]\r\nto\r\nSequence(feature=Value(dtype='float64', id=None), length=2, id=None)\r\n```\r\n\r\nIdeally we can have empty embedding columns as well!", "This part of PyArrow is buggy and inconsistent regarding features implemented across the types, so the only option is to operate on the Arrow buffer level to fix issues such as the above one.", "Ok - can you take the POC I did [here](https://github.com/huggingface/datasets/commit/15443098e9ce053943172f7ec6fce3769d7dff6e)? Happy to turn this into an actual PR but would appreciate feedback on the implementation before I take another pass!" ]
1970-01-01T00:00:00.000001
1,707
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug I have a dataset with an embedding column, when I try to map that dataset I get the following exception: ``` Traceback (most recent call last): File "/Users/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 3189, in map for rank, done, content in iflatmap_unordered( File "/Users/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/utils/py_utils.py", line 1387, in iflatmap_unordered [async_result.get(timeout=0.05) for async_result in async_results] File "/Users/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/datasets/utils/py_utils.py", line 1387, in <listcomp> [async_result.get(timeout=0.05) for async_result in async_results] File "/Users/jmif/.virtualenvs/llm-training/lib/python3.10/site-packages/multiprocess/pool.py", line 774, in get raise self._value TypeError: Couldn't cast array of type fixed_size_list<item: float>[2] to Sequence(feature=Value(dtype='float32', id=None), length=2, id=None) ``` ### Steps to reproduce the bug Here's a simple repro script: ``` from datasets import Features, Value, Sequence, ClassLabel, Dataset dataset_features = Features({ 'text': Value('string'), 'embedding': Sequence(Value('double'), length=2), 'categories': Sequence(ClassLabel(names=sorted([ 'one', 'two', 'three' ]))), }) dataset = Dataset.from_dict( { 'text': ['A'] * 10000, 'embedding': [[0.0, 0.1]] * 10000, 'categories': [[0]] * 10000, }, features=dataset_features ) def test_mapper(r): r['text'] = list(map(lambda t: t + ' b', r['text'])) return r dataset = dataset.map(test_mapper, batched=True, batch_size=10, features=dataset_features, num_proc=2) ``` Removing the embedding column fixes the issue! ### Expected behavior The mapping completes successfully. ### Environment info - `datasets` version: 2.14.4 - Platform: macOS-14.0-arm64-arm-64bit - Python version: 3.10.12 - Huggingface_hub version: 0.17.1 - PyArrow version: 13.0.0 - Pandas version: 2.0.3
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6280/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6280/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6279
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6279/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6279/comments
https://api.github.com/repos/huggingface/datasets/issues/6279/events
https://github.com/huggingface/datasets/issues/6279
1,928,028,226
I_kwDODunzps5y62BC
6,279
Batched IterableDataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/7010688?v=4", "events_url": "https://api.github.com/users/lneukom/events{/privacy}", "followers_url": "https://api.github.com/users/lneukom/followers", "following_url": "https://api.github.com/users/lneukom/following{/other_user}", "gists_url": "https://api.github.com/users/lneukom/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lneukom", "id": 7010688, "login": "lneukom", "node_id": "MDQ6VXNlcjcwMTA2ODg=", "organizations_url": "https://api.github.com/users/lneukom/orgs", "received_events_url": "https://api.github.com/users/lneukom/received_events", "repos_url": "https://api.github.com/users/lneukom/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lneukom/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lneukom/subscriptions", "type": "User", "url": "https://api.github.com/users/lneukom", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[ "This is exactly what I was looking for. It would also be very useful for me :-)", "This issue is really smashing the selling point of HF datasets... The only workaround I've found so far is to create a customized IterableDataloader which improves the loading speed to some extent.\r\n\r\nFor example I've a HF dataset `dt_train` with `len(dt_train) == 1M`. Using plain DataLoader is extremely slow:\r\n```\r\n%%time\r\ndl_train = DataLoader(dt_train, batch_size=128, shuffle = True)\r\nfor batch in dl_train:\r\n pass\r\n``` \r\n\r\n```\r\nCPU times: user 24min 35s, sys: 704 ms, total: 24min 36s\r\nWall time: 24min 37s\r\n```\r\nAnd DataLoader works even worse with HF's iterable_dataset:\r\n```\r\n%%time\r\ndt_train_ = dt_train.with_format(None).to_iterable_dataset(num_shards=64).shuffle(buffer_size=10_000)\r\ndl_train = DataLoader(dt_train_, batch_size=128)\r\nfor batch in dl_train:\r\n pass\r\n```\r\n```\r\nCPU times: user 1h 6min 2s, sys: 4.28 s, total: 1h 6min 6s\r\nWall time: 1h 7min 53s\r\n```\r\nWorkaround by running a customized wrapper:\r\n```\r\n%%time\r\nfrom torch.utils.data import DataLoader, IterableDataset\r\n\r\nclass Dataset2Iterable(IterableDataset):\r\n \"\"\"\r\n Wrapper to use a HF dataset as pytorch IterableDataset to speed up data loading.\r\n \"\"\"\r\n def __init__(self, dataset, batch_size=1, shuffle=True):\r\n super(Dataset2Iterable).__init__()\r\n self.dataset = dataset\r\n self.batch_size = batch_size\r\n self.shuffle = shuffle\r\n\r\n def __iter__(self):\r\n if self.shuffle: self.dataset.shuffle()\r\n return self.dataset.iter(batch_size=self.batch_size)\r\n\r\ndl_train = DataLoader(Dataset2Iterable(dt_train, batch_size = 128), batch_size=1, num_workers=0)\r\nfor n in range(2):\r\n for batch in dl_train:\r\n pass\r\n```\r\nThe speed still is slower than using tensorflow's loader but improved a lot than previous code:\r\n```\r\nCPU times: user 4min 18s, sys: 0 ns, total: 4min 18s\r\nWall time: 4min 20s\r\n```\r\nNote that the way I implemented `Dataset2Iterable` will only work with `num_workers=0`.", "I can confirm that @zhh210's solution works with `num_workers=0`. However, for my use case, this was still slower than tokenizing on the fly through a collator and leveraging multiple workers in the dataloder.\r\n\r\n@lhoestq I think this is an important use case (e.g., streaming from a large dataset, online or stored on disk). What do you think might be the best solution to move forward?", "I guess it can be implemented using a batched`.map()` under the hood that returns a single item containing the input batch.\r\n\r\nIn the meantime you can use this:\r\n\r\n```python\r\ndef batch(unbatched: dict[str, list]) -> dict[str, list]:\r\n return {k: [v] for k, v in unbatched}\r\n\r\nbatched_dataset = dataset.map(batch, batched=True, batch_size=batch_size)\r\n```\r\n\r\nThough it would be great to have a `.batch()` method indeed, I'd be happy to help with anyone wants to open a PR", "If no one else is planning to work on this, I can take it on. I'll wait until next week, and if no one has started a PR by then, I'll go ahead and open one.", "It looks like the implementation of IterableDataset is still using a hardcoded batch size of 1. For example in line 2063 in [`/datsets/src/datasets/iterable_dataset.py`](https://github.com/huggingface/datasets/blob/3.0.2/src/datasets/iterable_dataset.py#L2063). Iterating over IterableDataset with large batch sizes therefore remains slow, even when using `batch()`. I guess then the data are not being read from one contiguous chunk of memory. Instead every example is retrieved one by one, leading to long dataloading times. As a minimal example: Load c4 dataset and iterate over it with a large batch size.\r\n```python\r\nimport datasets\r\nfrom timeit import default_timer as timer\r\nc4 = datasets.load_dataset(\"allenai/c4\", \"en\", streaming=True, split=\"train\")\r\nc4_batched = c4.batch(512**2) # use large batch size\r\niterator = iter(c4_batched)\r\nfor i in range(5):\r\n start_time=timer()\r\n next(iterator) # get next batch\r\n end_time = timer()\r\n print(f\"time for one batch: {end_time-start_time}\")\r\n```\r\nThis results in the following output for me:\r\ntime for one batch: 12.615376660600305\r\ntime for one batch: 13.011422813870013\r\ntime for one batch: 14.157325950451195\r\ntime for one batch: 14.225894245319068\r\ntime for one batch: 13.898222777992487\r\n\r\nBecause I want to use my IterableDataset with the pytorch dataloader I rewrote the `__iter_pytorch__` and the `__iter__` functions like so and am getting much faster dataloading times. I marked the lines I changed with \"# changed here\":\r\n\r\n```python\r\nfrom datasets.iterable_dataset import _convert_to_arrow\r\nfrom datasets.formatting import TensorFormatter, get_formatter\r\nfrom datasets.features.features import cast_to_python_objects\r\nimport sys\r\nimport fsspec.asyn\r\nfrom itertools import islice\r\nfrom datasets.utils.logging import get_logger\r\nfrom datasets.iterable_dataset import _examples_to_batch, _apply_feature_types_on_batch, _apply_feature_types_on_example\r\n\r\nlogger = get_logger(__name__)\r\n\r\ndef __iter__(self):\r\n if \"torch\" in sys.modules:\r\n import torch.utils.data\r\n\r\n worker_info = torch.utils.data.get_worker_info()\r\n if isinstance(self, torch.utils.data.IterableDataset) and worker_info is not None:\r\n # We're a torch.utils.data.IterableDataset in a PyTorch worker process\r\n yield from self._iter_pytorch()\r\n return\r\n\r\n ex_iterable = self._prepare_ex_iterable_for_iteration(batch_size=self.batch_size, drop_last_batch=self.drop_last_batch) # changed here\r\n if self._formatting:\r\n formatter = get_formatter(self._formatting.format_type, features=self.features)\r\n format_dict = (\r\n formatter.recursive_tensorize if isinstance(formatter, TensorFormatter) else cast_to_python_objects\r\n )\r\n else:\r\n format_dict = None\r\n\r\n if self._formatting and (ex_iterable.iter_arrow or self._formatting.format_type == \"arrow\"):\r\n if ex_iterable.iter_arrow:\r\n iterator = ex_iterable.iter_arrow()\r\n else:\r\n iterator = _convert_to_arrow(ex_iterable, batch_size=self.batch_size) # changed here\r\n for key, pa_table in iterator:\r\n yield formatter.format_row(pa_table)\r\n return\r\n\r\n for key, example in ex_iterable:\r\n if self.features and not ex_iterable.is_typed:\r\n # `IterableDataset` automatically fills missing columns with None.\r\n # This is done with `_apply_feature_types_on_example`.\r\n example = _apply_feature_types_on_example(\r\n example, self.features, token_per_repo_id=self._token_per_repo_id\r\n )\r\n yield format_dict(example) if format_dict else example\r\n\r\n\r\n\r\ndef _iter_pytorch(self):\r\n ex_iterable = self._prepare_ex_iterable_for_iteration(batch_size=self.batch_size, drop_last_batch=self.drop_last_batch) # changed here\r\n # Fix for fsspec when using multiprocess to avoid hanging in the ML training loop. (only required for fsspec >= 0.9.0)\r\n # See https://github.com/fsspec/gcsfs/issues/379\r\n fsspec.asyn.reset_lock()\r\n # check if there aren't too many workers\r\n import torch.utils.data\r\n\r\n worker_info = torch.utils.data.get_worker_info()\r\n if self._is_main_process() and ex_iterable.n_shards < worker_info.num_workers:\r\n logger.warning(\r\n f\"Too many dataloader workers: {worker_info.num_workers} (max is dataset.n_shards={ex_iterable.n_shards}). \"\r\n f\"Stopping {worker_info.num_workers - ex_iterable.n_shards} dataloader workers.\"\r\n )\r\n logger.info(\r\n f\"To parallelize data loading, we give each process some shards (or data sources) to process. \"\r\n f\"Therefore it's unnecessary to have a number of workers greater than dataset.n_shards={ex_iterable.n_shards}. \"\r\n f\"To enable more parallelism, please split the dataset in more files than {ex_iterable.n_shards}.\"\r\n )\r\n # split workload\r\n _log_prefix = f\"node#{self._distributed.rank} \" if self._distributed else \"\"\r\n shards_indices = ex_iterable.split_shard_indices_by_worker(worker_info.id, worker_info.num_workers)\r\n if shards_indices:\r\n logger.debug(\r\n f\"{_log_prefix}dataloader worker#{worker_info.id}, ': Starting to iterate over {len(shards_indices)}/{ex_iterable.n_shards} shards.\"\r\n )\r\n ex_iterable = ex_iterable.shard_data_sources(worker_id=worker_info.id, num_workers=worker_info.num_workers)\r\n self._state_dict = ex_iterable._init_state_dict()\r\n if self._starting_state_dict:\r\n ex_iterable.load_state_dict(self._starting_state_dict)\r\n\r\n if self._formatting:\r\n formatter = get_formatter(self._formatting.format_type, features=self.features)\r\n format_dict = (\r\n formatter.recursive_tensorize if isinstance(formatter, TensorFormatter) else cast_to_python_objects\r\n )\r\n else:\r\n format_dict = None\r\n\r\n if self._formatting and (ex_iterable.iter_arrow or self._formatting == \"arrow\"):\r\n if ex_iterable.iter_arrow:\r\n iterator = ex_iterable.iter_arrow()\r\n else:\r\n iterator = _convert_to_arrow(ex_iterable, batch_size=self.batch_size) # changed here\r\n if self.batch_size > 1: # changed here until end of file\r\n for key, pa_table in iterator:\r\n yield formatter.format_batch(pa_table)\r\n return\r\n else:\r\n for key, pa_table in iterator:\r\n yield formatter.format_row(pa_table)\r\n return\r\n\r\n iterator = iter(ex_iterable)\r\n if self.batch_size > 1:\r\n for key, example in iterator:\r\n # If batched, first build the batch\r\n examples = [example] + [example for key, example in islice(iterator, self.batch_size - 1)]\r\n if self.drop_last_batch and len(examples) < self.batch_size: # ignore last batch\r\n return\r\n batch = _examples_to_batch(examples)\r\n if self.features and not ex_iterable.is_typed:\r\n # `IterableDataset` automatically fills missing columns with None.\r\n # This is done with `_apply_feature_types_on_batch`.\r\n batch = _apply_feature_types_on_batch(batch, self.features, token_per_repo_id=self._token_per_repo_id)\r\n yield format_dict(batch) if format_dict else batch\r\n else:\r\n for key, example in ex_iterable:\r\n if self.features and not ex_iterable.is_typed:\r\n # `IterableDataset` automatically fills missing columns with None.\r\n # This is done with `_apply_feature_types_on_example`.\r\n example = _apply_feature_types_on_example(\r\n example, self.features, token_per_repo_id=self._token_per_repo_id\r\n )\r\n yield format_dict(example) if format_dict else example\r\n logger.debug(\r\n f\"{_log_prefix}dataloader worker#{worker_info.id}, ': Finished iterating over {len(shards_indices)}/{ex_iterable.n_shards} shards.\"\r\n )\r\n else:\r\n logger.debug(\r\n f\"{_log_prefix}dataloader worker#{worker_info.id}, ': Stopping... Number of dataset shards < num_workers ({ex_iterable.n_shards}<{worker_info.num_workers}).\"\r\n )\r\n```\r\n\r\nFor anyone wanting to try it you can patch it into datasets by overwriting the function via `setattr(datasets.IterableDataset, '_iter_pytorch', _iter_pytorch)`\r\n\r\nI don't really know what most of the rest of the code is doing so no idea if this is a valid fix or not, but it seems to work for me.\r\nExample of running the fix:\r\n```python\r\nfrom torch.utils.data.dataloader import DataLoader\r\nc4.batch_size = 512**2 # set batch size here\r\ndataloader = Dataloader(c4, batch_size=None) # use custom batching from IterableDataset\r\niterator = iter(dataloader)\r\nfor i in range(5):\r\n start_time=timer()\r\n next(iterator) #get the batch\r\n end_time = timer()\r\n print(f\"time for one batch: {end_time-start_time}\")\r\n```\r\nI now get \r\ntime for one batch: 0.6047679269686341\r\ntime for one batch: 0.000248616561293602\r\ntime for one batch: 0.00017435848712921143\r\ntime for one batch: 0.00015910807996988297\r\ntime for one batch: 0.00015317369252443314\r\n\r\nI love the datasets library and it would be great if iterating with large batch sizes would be supported directly, either with a similar fix to mine or in some other way :)", "Hi @taczin , thanks for reporting !\r\n\r\nIndeed the `IterableDataset.batch()` implementation is quite naive is manipulates python objects:\r\n\r\nhttps://github.com/huggingface/datasets/blob/d37ed46ebf45981131bd3678173dbb4b7e2b2f1a/src/datasets/iterable_dataset.py#L3026-L3029\r\n\r\nHowever it can be much faster if it can be applied on the Arrow data, maybe using something like this (untested)\r\n\r\n```python\r\ndef batch_fn(unbatched): \r\n return {k: [v] for k, v in unbatched.items()} \r\n\r\ndef batch_fn_arrow(unbatched_pa_table): \r\n offsets = pa.array([0, len(unbatched_pa_table)])\r\n return pa.Table.from_arrays([\r\n pa.ListArray.from_arrays(offsets, unbatched_pa_table[k])\r\n for k in unbatched_pa_table.column_names\r\n ], unbatched_pa_table.column_names)\r\n\r\nif self._ex_iterabe.iter_arrow:\r\n return self.with_format(\"arrow\").map(\r\n batch_fn_arrow, batched=True, batch_size=batch_size, drop_last_batch=drop_last_batch\r\n ).with_format(self._formatting.format_type if self._formatting else None)\r\nelse:\r\n return self.map(batch_fn, batched=True, batch_size=batch_size, drop_last_batch=drop_last_batch)\r\n```", "Hi @lhoestq , thanks for your answer. I was wondering: is there a reason why the internal call to `ex_iterable = self._prepare_ex_iterable_for_iteration()` in the IterableDataset code does not pass the batch size even though it could? If not passed the default of 1 is used, leading to the observed slow loading.", "After calling `.batch()`, `_prepare_ex_iterable_for_iteration` should use batch_size=1 since now each row in the dataset is actually a batch of the original dataset." ]
1970-01-01T00:00:00.000001
1,730
null
NONE
null
### Feature request Hi, could you add an implementation of a batched `IterableDataset`. It already support an option to do batch iteration via `.iter(batch_size=...)` but this cannot be used in combination with a torch `DataLoader` since it just returns an iterator. ### Motivation The current implementation loads each element of a batch individually which can be very slow in cases of a big batch_size. I did some experiments [here](https://discuss.huggingface.co/t/slow-dataloader-with-big-batch-size/57224) and using a batched iteration would speed up data loading significantly. ### Your contribution N/A
null
{ "+1": 7, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 7, "url": "https://api.github.com/repos/huggingface/datasets/issues/6279/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6279/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6277
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6277/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6277/comments
https://api.github.com/repos/huggingface/datasets/issues/6277/events
https://github.com/huggingface/datasets/issues/6277
1,927,044,546
I_kwDODunzps5y3F3C
6,277
FileNotFoundError: Couldn't find a module script at /content/paws-x/paws-x.py. Module 'paws-x' doesn't exist on the Hugging Face Hub either.
{ "avatar_url": "https://avatars.githubusercontent.com/u/66733346?v=4", "events_url": "https://api.github.com/users/diegogonzalezc/events{/privacy}", "followers_url": "https://api.github.com/users/diegogonzalezc/followers", "following_url": "https://api.github.com/users/diegogonzalezc/following{/other_user}", "gists_url": "https://api.github.com/users/diegogonzalezc/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/diegogonzalezc", "id": 66733346, "login": "diegogonzalezc", "node_id": "MDQ6VXNlcjY2NzMzMzQ2", "organizations_url": "https://api.github.com/users/diegogonzalezc/orgs", "received_events_url": "https://api.github.com/users/diegogonzalezc/received_events", "repos_url": "https://api.github.com/users/diegogonzalezc/repos", "site_admin": false, "starred_url": "https://api.github.com/users/diegogonzalezc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/diegogonzalezc/subscriptions", "type": "User", "url": "https://api.github.com/users/diegogonzalezc", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "`evaluate.load(\"paws-x\", \"es\")` throws the error because there is no such metric in the `evaluate` lib.\r\n\r\nSo, this is unrelated to our lib." ]
1970-01-01T00:00:00.000001
1,696
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug I'm encountering a "FileNotFoundError" while attempting to use the "paws-x" dataset to retrain the DistilRoBERTa-base model. The error message is as follows: FileNotFoundError: Couldn't find a module script at /content/paws-x/paws-x.py. Module 'paws-x' doesn't exist on the Hugging Face Hub either. ### Steps to reproduce the bug https://colab.research.google.com/drive/11xUUFxloClpmqLvDy_Xxfmo3oUzjY5nx#scrollTo=kUn74FigzhHm ### Expected behavior The the trained model ### Environment info colab, "paws-x" dataset , DistilRoBERTa-base model
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6277/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6277/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6276
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6276/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6276/comments
https://api.github.com/repos/huggingface/datasets/issues/6276/events
https://github.com/huggingface/datasets/issues/6276
1,925,961,878
I_kwDODunzps5yy9iW
6,276
I'm trying to fine tune the openai/whisper model from huggingface using jupyter notebook and i keep getting this error
{ "avatar_url": "https://avatars.githubusercontent.com/u/50768065?v=4", "events_url": "https://api.github.com/users/valaofficial/events{/privacy}", "followers_url": "https://api.github.com/users/valaofficial/followers", "following_url": "https://api.github.com/users/valaofficial/following{/other_user}", "gists_url": "https://api.github.com/users/valaofficial/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/valaofficial", "id": 50768065, "login": "valaofficial", "node_id": "MDQ6VXNlcjUwNzY4MDY1", "organizations_url": "https://api.github.com/users/valaofficial/orgs", "received_events_url": "https://api.github.com/users/valaofficial/received_events", "repos_url": "https://api.github.com/users/valaofficial/repos", "site_admin": false, "starred_url": "https://api.github.com/users/valaofficial/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/valaofficial/subscriptions", "type": "User", "url": "https://api.github.com/users/valaofficial", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Since you are using Windows, maybe moving the `map` call inside `if __name__ == \"__main__\"` can fix the issue:\r\n```python\r\nif __name__ == \"__main__\":\r\n common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names[\"train\"], num_proc=4)\r\n```\r\n\r\nOtherwise, the only solution is to set `num_proc=1`.", "> Since you are using Windows, maybe moving the `map` call inside `if __name__ == \"__main__\"` can fix the issue:\r\n> \r\n> ```python\r\n> if __name__ == \"__main__\":\r\n> common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names[\"train\"], num_proc=4)\r\n> ```\r\n> \r\n> Otherwise, the only solution is to set `num_proc=1`.\r\n\r\nThank you very much for the response, i eventually tried setting `num_proc=1` and now the jupyter notebook kernel keers dying after running the command, what do you think the issue could be, could it be that my system is not capable of running the command \"i'm using a Lenovo Thinkpad T440 with no GPU\"", "Firstly, you didn't define feature_extractor variable. Secondly, it is large nlp model. Hence you should use proper gpu, otherwise your machine's cpu will be overclock and you can do nothing." ]
1970-01-01T00:00:00.000001
1,701
null
NONE
null
### Describe the bug I'm trying to fine tune the openai/whisper model from huggingface using jupyter notebook and i keep getting this error, i'm following the steps in this blog post https://huggingface.co/blog/fine-tune-whisper I tried google collab and it works but because I'm on the free version the training doesn't complete the error comes in jupyter notebook when i run this line `common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names["train"], num_proc=4)` here is the error message ``` Map (num_proc=4): 0% 0/2506 [00:52<?, ? examples/s] The above exception was the direct cause of the following exception: NameError Traceback (most recent call last) Cell In[19], line 1 ----> 1 common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names["train"], num_proc=4) File ~\anaconda\Lib\site-packages\datasets\dataset_dict.py:853, in DatasetDict.map(self, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, desc) 850 if cache_file_names is None: 851 cache_file_names = {k: None for k in self} 852 return DatasetDict( --> 853 { 854 k: dataset.map( 855 function=function, 856 with_indices=with_indices, 857 with_rank=with_rank, 858 input_columns=input_columns, 859 batched=batched, 860 batch_size=batch_size, 861 drop_last_batch=drop_last_batch, 862 remove_columns=remove_columns, 863 keep_in_memory=keep_in_memory, 864 load_from_cache_file=load_from_cache_file, 865 cache_file_name=cache_file_names[k], 866 writer_batch_size=writer_batch_size, 867 features=features, 868 disable_nullable=disable_nullable, 869 fn_kwargs=fn_kwargs, 870 num_proc=num_proc, 871 desc=desc, 872 ) 873 for k, dataset in self.items() 874 } 875 ) File ~\anaconda\Lib\site-packages\datasets\dataset_dict.py:854, in <dictcomp>(.0) 850 if cache_file_names is None: 851 cache_file_names = {k: None for k in self} 852 return DatasetDict( 853 { --> 854 k: dataset.map( 855 function=function, 856 with_indices=with_indices, 857 with_rank=with_rank, 858 input_columns=input_columns, 859 batched=batched, 860 batch_size=batch_size, 861 drop_last_batch=drop_last_batch, 862 remove_columns=remove_columns, 863 keep_in_memory=keep_in_memory, 864 load_from_cache_file=load_from_cache_file, 865 cache_file_name=cache_file_names[k], 866 writer_batch_size=writer_batch_size, 867 features=features, 868 disable_nullable=disable_nullable, 869 fn_kwargs=fn_kwargs, 870 num_proc=num_proc, 871 desc=desc, 872 ) 873 for k, dataset in self.items() 874 } 875 ) File ~\anaconda\Lib\site-packages\datasets\arrow_dataset.py:592, in transmit_tasks.<locals>.wrapper(*args, **kwargs) 590 self: "Dataset" = kwargs.pop("self") 591 # apply actual function --> 592 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 593 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 594 for dataset in datasets: 595 # Remove task templates if a column mapping of the template is no longer valid File ~\anaconda\Lib\site-packages\datasets\arrow_dataset.py:557, in transmit_format.<locals>.wrapper(*args, **kwargs) 550 self_format = { 551 "type": self._format_type, 552 "format_kwargs": self._format_kwargs, 553 "columns": self._format_columns, 554 "output_all_columns": self._output_all_columns, 555 } 556 # apply actual function --> 557 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 558 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 559 # re-apply format to the output File ~\anaconda\Lib\site-packages\datasets\arrow_dataset.py:3189, in Dataset.map(self, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 3182 logger.info(f"Spawning {num_proc} processes") 3183 with logging.tqdm( 3184 disable=not logging.is_progress_bar_enabled(), 3185 unit=" examples", 3186 total=pbar_total, 3187 desc=(desc or "Map") + f" (num_proc={num_proc})", 3188 ) as pbar: -> 3189 for rank, done, content in iflatmap_unordered( 3190 pool, Dataset._map_single, kwargs_iterable=kwargs_per_job 3191 ): 3192 if done: 3193 shards_done += 1 File ~\anaconda\Lib\site-packages\datasets\utils\py_utils.py:1394, in iflatmap_unordered(pool, func, kwargs_iterable) 1391 finally: 1392 if not pool_changed: 1393 # we get the result in case there's an error to raise -> 1394 [async_result.get(timeout=0.05) for async_result in async_results] File ~\anaconda\Lib\site-packages\datasets\utils\py_utils.py:1394, in <listcomp>(.0) 1391 finally: 1392 if not pool_changed: 1393 # we get the result in case there's an error to raise -> 1394 [async_result.get(timeout=0.05) for async_result in async_results] File ~\anaconda\Lib\site-packages\multiprocess\pool.py:774, in ApplyResult.get(self, timeout) 772 return self._value 773 else: --> 774 raise self._value NameError: name 'feature_extractor' is not defined ``` ### Steps to reproduce the bug 1. follow the steps in this blog post https://huggingface.co/blog/fine-tune-whisper 2. run this line of code `common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names["train"], num_proc=4)` 3. I'm using jupyter notebook from anaconda ### Expected behavior No error message ### Environment info datasets version: 2.8.0 Python version: 3.11 Windows 10
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6276/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6276/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6275
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6275/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6275/comments
https://api.github.com/repos/huggingface/datasets/issues/6275/events
https://github.com/huggingface/datasets/issues/6275
1,921,354,680
I_kwDODunzps5yhYu4
6,275
Would like to Contribute a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/97907750?v=4", "events_url": "https://api.github.com/users/vikas70607/events{/privacy}", "followers_url": "https://api.github.com/users/vikas70607/followers", "following_url": "https://api.github.com/users/vikas70607/following{/other_user}", "gists_url": "https://api.github.com/users/vikas70607/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vikas70607", "id": 97907750, "login": "vikas70607", "node_id": "U_kgDOBdX0Jg", "organizations_url": "https://api.github.com/users/vikas70607/orgs", "received_events_url": "https://api.github.com/users/vikas70607/received_events", "repos_url": "https://api.github.com/users/vikas70607/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vikas70607/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vikas70607/subscriptions", "type": "User", "url": "https://api.github.com/users/vikas70607", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "Hi! The process of contributing a dataset is explained here: https://huggingface.co/docs/datasets/upload_dataset. Also, check https://huggingface.co/docs/datasets/image_dataset for a more detailed explanation of how to share an image dataset." ]
1970-01-01T00:00:00.000001
1,696
1970-01-01T00:00:00.000001
NONE
null
I have a dataset of 2500 images that can be used for color-blind machine-learning algorithms. Since , there was no dataset available online , I made this dataset myself and would like to contribute this now to community
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6275/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6275/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6274
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6274/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6274/comments
https://api.github.com/repos/huggingface/datasets/issues/6274/events
https://github.com/huggingface/datasets/issues/6274
1,921,036,328
I_kwDODunzps5ygLAo
6,274
FileNotFoundError for dataset with multiple builder config
{ "avatar_url": "https://avatars.githubusercontent.com/u/97120485?v=4", "events_url": "https://api.github.com/users/LouisChen15/events{/privacy}", "followers_url": "https://api.github.com/users/LouisChen15/followers", "following_url": "https://api.github.com/users/LouisChen15/following{/other_user}", "gists_url": "https://api.github.com/users/LouisChen15/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LouisChen15", "id": 97120485, "login": "LouisChen15", "node_id": "U_kgDOBcnw5Q", "organizations_url": "https://api.github.com/users/LouisChen15/orgs", "received_events_url": "https://api.github.com/users/LouisChen15/received_events", "repos_url": "https://api.github.com/users/LouisChen15/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LouisChen15/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LouisChen15/subscriptions", "type": "User", "url": "https://api.github.com/users/LouisChen15", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "Please tell me if the above info is not enough for solving the problem. I will then make my dataset public temporarily so that you can really reproduce the bug. ", "Hi! \r\nCould you share how to solve this problem? \r\nI faced this same error. " ]
1970-01-01T00:00:00.000001
1,723
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug When there is only one config and only the dataset name is entered when using datasets.load_dataset(), it works fine. But if I create a second builder_config for my dataset and enter the config name when using datasets.load_dataset(), the following error will happen. FileNotFoundError: [Errno 2] No such file or directory: 'C:/Users/chenx/.cache/huggingface/datasets/my_dataset/0_shot_multiple_choice/1.0.0/97c3854a012cfd6b045e3be4c864739902af2d818bb9235b047baa94c302e9a2.incomplete/my_dataset-test-00000-00000-of-NNNNN.arrow' The "XXX.incomplete folder" in the cache folder of my dataset will disappear before "generating test split", which does not happen when config name is not entered and the config name is "default" C:\Users\chenx\.cache\huggingface\datasets\my_dataset\0_shot_multiple_choice\1.0.0 The folder that is supposed to remain under the above directory will disappear, and the data generator will not have a place to generate data into. ### Steps to reproduce the bug test = load_dataset('my_dataset', '0_shot_multiple_choice') ### Expected behavior FileNotFoundError: [Errno 2] No such file or directory: 'C:/Users/chenx/.cache/huggingface/datasets/my_dataset/0_shot_multiple_choice/1.0.0/97c3854a012cfd6b045e3be4c864739902af2d818bb9235b047baa94c302e9a2.incomplete/my_dataset-test-00000-00000-of-NNNNN.arrow' ### Environment info datasets 2.14.5 python 3.8.18
{ "avatar_url": "https://avatars.githubusercontent.com/u/97120485?v=4", "events_url": "https://api.github.com/users/LouisChen15/events{/privacy}", "followers_url": "https://api.github.com/users/LouisChen15/followers", "following_url": "https://api.github.com/users/LouisChen15/following{/other_user}", "gists_url": "https://api.github.com/users/LouisChen15/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LouisChen15", "id": 97120485, "login": "LouisChen15", "node_id": "U_kgDOBcnw5Q", "organizations_url": "https://api.github.com/users/LouisChen15/orgs", "received_events_url": "https://api.github.com/users/LouisChen15/received_events", "repos_url": "https://api.github.com/users/LouisChen15/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LouisChen15/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LouisChen15/subscriptions", "type": "User", "url": "https://api.github.com/users/LouisChen15", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6274/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6274/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6273
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6273/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6273/comments
https://api.github.com/repos/huggingface/datasets/issues/6273/events
https://github.com/huggingface/datasets/issues/6273
1,920,922,260
I_kwDODunzps5yfvKU
6,273
Broken Link to PubMed Abstracts dataset .
{ "avatar_url": "https://avatars.githubusercontent.com/u/100606327?v=4", "events_url": "https://api.github.com/users/sameemqureshi/events{/privacy}", "followers_url": "https://api.github.com/users/sameemqureshi/followers", "following_url": "https://api.github.com/users/sameemqureshi/following{/other_user}", "gists_url": "https://api.github.com/users/sameemqureshi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sameemqureshi", "id": 100606327, "login": "sameemqureshi", "node_id": "U_kgDOBf8hdw", "organizations_url": "https://api.github.com/users/sameemqureshi/orgs", "received_events_url": "https://api.github.com/users/sameemqureshi/received_events", "repos_url": "https://api.github.com/users/sameemqureshi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sameemqureshi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sameemqureshi/subscriptions", "type": "User", "url": "https://api.github.com/users/sameemqureshi", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "This has already been reported in the HF Course repo (https://github.com/huggingface/course/issues/623).", "@lhoestq @albertvillanova @lewtun I don't think we are allowed to host these data files on the Hub (due to DMCA), which means the only option is to use a different dataset in the course (and to re-record the video 🙂), no?", "Keeping the video is maybe fine, we can add a note on youtube to suggest to load a dataset with a different name. Maybe C4 ? And update the code snippets on the website ?", "Maybe you want to try it with the PUBMED dataset that I reproduced based on the The [PubMed Abstract GitHub Site](http://github.com/thoppe/The-Pile-PubMed) and uploaded on the HuggingFace:\r\n\r\n```\r\nfrom datasets import load_dataset\r\npubmed_dataset = load_dataset(\"hwang2006/PUBMED_title_abstracts_2020_baseline\")\r\npubmed_dataset\r\n\r\n#Downloading data: 100%\r\n#7.98G/7.98G [11:47<00:00, 9.68MB/s]\r\n#Generating train split: 17722096/0 [00:36<00:00, 505376.37 examples/s]\r\n\r\n#DatasetDict({\r\n# train: Dataset({\r\n# features: ['meta', 'text'],\r\n# num_rows: 17722096\r\n# })\r\n#})\r\n```", "孔令涛说感谢感谢" ]
1970-01-01T00:00:00.000001
1,714
null
NONE
null
### Describe the bug The link provided for the dataset is broken, data_files = [https://the-eye.eu/public/AI/pile_preliminary_components/PUBMED_title_abstracts_2019_baseline.jsonl.zst](url) The ### Steps to reproduce the bug Steps to reproduce: 1) Head over to [https://huggingface.co/learn/nlp-course/chapter5/4?fw=pt#big-data-datasets-to-the-rescue](url) 2) In the Section "What is the Pile?", you can see a code snippet that contains the broken link. ### Expected behavior The link should Redirect to the "PubMed Abstracts dataset" as expected . ### Environment info .
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6273/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6273/timeline
null
null
null
null
false
null
https://api.github.com/repos/huggingface/datasets/issues/6272
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6272/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6272/comments
https://api.github.com/repos/huggingface/datasets/issues/6272/events
https://github.com/huggingface/datasets/issues/6272
1,920,831,487
I_kwDODunzps5yfY__
6,272
Duplicate `data_files` when named `<split>/<split>.parquet`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[ "Also reported in https://github.com/huggingface/datasets/issues/6259", "I think it's best to drop duplicates with a `set` (as a temporary fix) and improve the patterns when/if https://github.com/fsspec/filesystem_spec/pull/1382 gets merged. @lhoestq Do you have some other ideas?", "Alternatively we could just use this no ?\r\n\r\n```python\r\nif config.FSSPEC_VERSION < version.parse(\"2023.9.0\"):\r\n KEYWORDS_IN_PATH_NAME_BASE_PATTERNS = [\r\n \"{keyword}[{sep}/]**\",\r\n \"**[{sep}]{keyword}[{sep}/]**\",\r\n \"**/{keyword}[{sep}/]**\",\r\n ]\r\nelse:\r\n KEYWORDS_IN_PATH_NAME_BASE_PATTERNS = [\r\n \"{keyword}[{sep}/]**\",\r\n \"**/*[{sep}]{keyword}[{sep}/]**\",\r\n \"**/*/{keyword}[{sep}/]**\",\r\n ]\r\n```\r\n\r\nThis way no need to implement sets, which would require a bit of work since we've always considered a list of pattern to be resolved as the concatenated list of resolved files for each pattern (including duplicates)\r\n", "Arf `\"**/*/{keyword}[{sep}/]**\"` does return `data/keyword.txt` in latest `fsspec` but not in `glob.glob`\r\n\r\nEDIT: actually forgot to set `recursive=True`", "Actually `glob.glob` does return it with `recursive=True` ! my bad", "Pff just tested and my idea sucks, pattern 1 and 3 obviously give duplicates ", "> I think it's best to drop duplicates with a set (as a temporary fix)\r\n\r\nI started https://github.com/huggingface/datasets/pull/6278 to use DataFilesSet objects instead of DataFilesList" ]
1970-01-01T00:00:00.000001
1,710
1970-01-01T00:00:00.000001
MEMBER
null
e.g. with `u23429/stock_1_minute_ticker` ```ipython In [1]: from datasets import * In [2]: b = load_dataset_builder("u23429/stock_1_minute_ticker") Downloading readme: 100%|██████████████████████████| 627/627 [00:00<00:00, 246kB/s] In [3]: b.config.data_files Out[3]: {NamedSplit('train'): ['hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/train/train.parquet', 'hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/train/train.parquet'], NamedSplit('validation'): ['hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/validation/validation.parquet', 'hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/validation/validation.parquet'], NamedSplit('test'): ['hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/test/test.parquet', 'hf://datasets/u23429/stock_1_minute_ticker@65c973cf4ec061f01a363b40da4c1bb128ba4166/test/test.parquet']} ``` This bug issue is present in the current `datasets` 2.14.5 and also on `main` even after https://github.com/huggingface/datasets/pull/6244 cc @mariosasko
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6272/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6272/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6271
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6271/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6271/comments
https://api.github.com/repos/huggingface/datasets/issues/6271/events
https://github.com/huggingface/datasets/issues/6271
1,920,420,295
I_kwDODunzps5yd0nH
6,271
Overwriting Split overwrites data but not metadata, corrupting dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/13859249?v=4", "events_url": "https://api.github.com/users/govindrai/events{/privacy}", "followers_url": "https://api.github.com/users/govindrai/followers", "following_url": "https://api.github.com/users/govindrai/following{/other_user}", "gists_url": "https://api.github.com/users/govindrai/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/govindrai", "id": 13859249, "login": "govindrai", "node_id": "MDQ6VXNlcjEzODU5MjQ5", "organizations_url": "https://api.github.com/users/govindrai/orgs", "received_events_url": "https://api.github.com/users/govindrai/received_events", "repos_url": "https://api.github.com/users/govindrai/repos", "site_admin": false, "starred_url": "https://api.github.com/users/govindrai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/govindrai/subscriptions", "type": "User", "url": "https://api.github.com/users/govindrai", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
NONE
null
### Describe the bug I want to be able to overwrite/update/delete splits in my dataset. Currently the only way to do is to manually go into the dataset and delete the split. If I try to overwrite programmatically I end up in an error state and (somewhat) corrupting the dataset. Read below. **Current Behavior** When I push to an existing split I get this error: `ValueError: Split complexRoofLocation_01Apr2023_to_31May2023test already present` This seems to suggest that the library doesn't support overwriting splits. **Potential Bug** What’s strange is that datasets, despite the operation erroring out with the ValueError above, does, in fact, overwrite the split: `Pushing dataset shards to the dataset hub: 100% [.....................] 1/1 [00:00<00:00, 55.04it/s]` Even though you got an error message and your code fails, your dataset is now changed. That seems like a bug. Either don't change the dataset, or don't throw the error and allow the script to proceed. Additional Bug While it overwrites the split, it doesn’t overwrite the split’s information. Because of this when you pull down the dataset you may end up getting a `NonMatchingSplitsSizesError` if the size of the dataset during the overwrite is different. For example, my original split had 5 rows, but on my overwrite, I only had 4. Then when I try to download the dataset, I get a `NonMatchingSplitsSizesError` because the dataset's data.json states there’s 5 but only 4 exist in the split. Expected Behavior This corrupts the dataset rendering it unusable (until you take manual intervention). Either the library should let the overwrite happen (which it does but should also update the metadata) or it shouldn’t do anything. ### Steps to reproduce the bug [Colab Notebook](https://colab.research.google.com/drive/1bqVkD06Ngs9MQNdSk_ygCG6y1UqXA4pC?usp=sharing) ### Expected behavior The split should be overwritten and I should be able to use the new version of the dataset without issue. ### Environment info - `datasets` version: 2.14.5 - Platform: Linux-5.15.120+-x86_64-with-glibc2.35 - Python version: 3.10.12 - Huggingface_hub version: 0.17.3 - PyArrow version: 9.0.0 - Pandas version: 1.5.3
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6271/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6271/timeline
null
completed
null
null
false
0
https://api.github.com/repos/huggingface/datasets/issues/6270
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6270/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6270/comments
https://api.github.com/repos/huggingface/datasets/issues/6270/events
https://github.com/huggingface/datasets/issues/6270
1,920,329,373
I_kwDODunzps5ydead
6,270
Dataset.from_generator raises with sharded gen_args
{ "avatar_url": "https://avatars.githubusercontent.com/u/53510?v=4", "events_url": "https://api.github.com/users/hartmans/events{/privacy}", "followers_url": "https://api.github.com/users/hartmans/followers", "following_url": "https://api.github.com/users/hartmans/following{/other_user}", "gists_url": "https://api.github.com/users/hartmans/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hartmans", "id": 53510, "login": "hartmans", "node_id": "MDQ6VXNlcjUzNTEw", "organizations_url": "https://api.github.com/users/hartmans/orgs", "received_events_url": "https://api.github.com/users/hartmans/received_events", "repos_url": "https://api.github.com/users/hartmans/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hartmans/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hartmans/subscriptions", "type": "User", "url": "https://api.github.com/users/hartmans", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "`gen_kwargs` should be a `dict`, as stated in the docstring, but you are passing a `list`.\r\n\r\nSo, to fix the error, replace the list of dicts with a dict of lists (and slightly modify the generator function):\r\n```python\r\nfrom pathlib import Path\r\nimport datasets\r\n\r\ndef process_yaml(files):\r\n for f in files:\r\n # process\r\n yield dict(...)\r\n\r\n\r\nif __name__ == '__main__':\r\n import sys\r\n dir = Path(sys.argv[0]).parent\r\n ds = datasets.Dataset.from_generator(process_yaml, gen_kwargs={'files': [f for f in dir.glob('*.yml')]})\r\n ds.to_json('training.jsonl')\r\n```", "That runs, and because my dataset is small, it's what I did to get past the problem.\r\nHowever, it does not produce a sharded dataset. From the doc string I expect there ought to be a way to call from_generator such that num_shards in the resulting data set is equal to the number of items in the list.\r\nThe part of the doc string that your suggestion is not responsive to is:\r\n` You can define a sharded dataset by passing the list of shards in *g\r\nen_kwargs*.\r\n`\r\n\r\nWhat your suggestion does is calls the generator once, with the list argument, and produces a single shard dataset.\r\n", "The sharding mentioned here refers to using this function with `num_proc` (multiprocessing splits the `kwargs` into shards and passes them to the generator function)\r\n\r\n> That runs, and because my dataset is small, it's what I did to get past the problem.\r\n\r\n`from_generator` generates a memory-mapped dataset (can be larger than RAM), so the dataset size should not be an issue unless the generator function's implementation does not properly free the memory.\r\n", "It sounds like you are saying that num_proc affects the form of gen_kwargs.\r\nAre you saying that for non-zero num_proc gen_kwargs should be a list whose length is the same as num_proc?\r\nOr are you saying that for non-zero num_proc, gen_kwargs should be a dict whose elements are lists the length of num_proc?\r\n", "I ran some tests. So, it looks like with num_proc greater than 1, gen_kwargs is expected to be a dict of lists. It calls the generator also with a dict of lists, but the lists are split.\r\nI.E. if my original has `gen_kwargs=dict(a=[0,1,2])`, then my generator might get called with `gen_kwalrgs=dict([0])`.\r\nThat all makes sense, but I definitely think there is room for improvement in the doc string here.\r\nIn order to suggest improvements to the doc string, I need to look at how the gen_kwargs are split, and figure out if:\r\n* num_proc needs to exactly equal the length of the lists\r\n* num_proc needs to evenly divide the length of the lists\r\n* Or there's no required relationship.\r\nI'll look into that and then propose an improved doc string if no one else gets to it first.", "Okay, that was fun; I took a dive through the dataset code and feel like I have a much better understanding.\r\nHere is my understanding of the behavior:\r\n* max_proc is an upper limit on the number of shards that `from_generator` produces\r\n* If `max_proc` is greater than 1, then all lists in *gen_kwargs* must be the same length\r\n* If the lists in *gen_kwargs* are shorter than *num_proc* elements, *num_proc* will be reduced and a warning produced. Put another way, `min(list_length, num_shards)` shards will be produced\r\n* The members of the lists in *gen_kwargs* will be partitioned among the created jobs.\r\nTo validate the above, take a look at\r\n`_number_of_shards_in_gen_kwargs` and `_distribute_shards` and `_split_gen_kwargs` in utils/sharding.py.\r\nI've also chased down starting at *from_generator* all the way through to GeneratorBuilder and the calls to the functions in sharding.py.\r\nTomorrow I'll take a look at the contributing guidelines and see what's involved in putting together a PR to improve the doc string." ]
1970-01-01T00:00:00.000001
1,697
1970-01-01T00:00:00.000001
CONTRIBUTOR
null
### Describe the bug According to the docs of Datasets.from_generator: ``` gen_kwargs(`dict`, *optional*): Keyword arguments to be passed to the `generator` callable. You can define a sharded dataset by passing the list of shards in `gen_kwargs`. ``` So I'd expect that if gen_kwargs was a list, then my generator would be called once for each element in the list with the dict in the list for that element. It doesn't work that way though. ### Steps to reproduce the bug ```python #!/usr/bin/python from pathlib import Path import datasets def process_yaml(file): yield dict(example=42) if __name__ == '__main__': import sys dir = Path(sys.argv[0]).parent ds = datasets.Dataset.from_generator(process_yaml, gen_kwargs=[{'file':f} for f in dir.glob('*.yml')], ) ds.to_json('training.jsonl') ``` ``` Generating train split: 0 examples [00:00, ? examples/s] Traceback (most recent call last): File "/tmp/dataset_bug.py", line 13, in <module> ds = datasets.Dataset.from_generator(process_yaml, gen_kwargs=[{'file':f} for f in dir.glob('*.yml')], ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/arrow_dataset.py", line 1072, in from_generator ).read() ^^^^^^ File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/io/generator.py", line 47, in read self.builder.download_and_prepare( File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 954, in download_and_prepare self._download_and_prepare( File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 1717, in _download_and_prepare super()._download_and_prepare( File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 1049, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 1555, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/home/hartmans/ai/venv/lib/python3.11/site-packages/datasets/builder.py", line 1656, in _prepare_split_single generator = self._generate_examples(**gen_kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ TypeError: datasets.packaged_modules.generator.generator.Generator._generate_examples() argument after ** must be a ``` mapping, not list ### Expected behavior I would expect that process_yaml would be called once for each yaml file in the directory where the script is run. I also tried with the list being in gen_kwargs, but in that case process_yaml gets called with a list. ### Environment info - `datasets` version: 2.14.6.dev0 (git commit 0cc77d7f45c7369; also tested with 2.14.0) - Platform: Linux-6.1.0-10-amd64-x86_64-with-glibc2.36 - Python version: 3.11.2 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6270/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6270/timeline
null
completed
null
null
false
0