Datasets:
mb23
/

Languages:
English
License:
mickylan2367 commited on
Commit
6038f66
·
1 Parent(s): 6236b5b

chaged loading script

Browse files
Files changed (1) hide show
  1. GraySpectrogram.py +195 -76
GraySpectrogram.py CHANGED
@@ -34,14 +34,126 @@ Using for Project Learning...
34
  # DatasetInfo : https://huggingface.co/docs/datasets/package_reference/main_classes
35
 
36
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
 
38
  class GraySpectrogram2(datasets.GeneratorBasedBuilder):
39
 
40
  # データのサブセットはここで用意
41
  BUILDER_CONFIGS = [
42
  datasets.BuilderConfig(
43
- name="train",
44
  description=_DESCRIPTION,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  )
46
  ]
47
 
@@ -78,85 +190,92 @@ class GraySpectrogram2(datasets.GeneratorBasedBuilder):
78
  )
79
 
80
  def _split_generators(self, dl_manager: DownloadManager):
81
- # huggingfaceのディレクトリからデータを取ってくる
82
- hfh_dataset_info = HfApi().dataset_info(_NAME, revision=_REVISION, timeout=100.0)
83
-
84
- # メタデータであるjsonlファイルのURLを取得
85
- # ここの抽出方法変えられないかな?
86
- train_metadata_url = DataFilesDict.from_hf_repo(
87
- {datasets.Split.TRAIN: ["data/train/**"]},
88
- dataset_info=hfh_dataset_info,
89
- allowed_extensions=["jsonl", ".jsonl"],
90
- )
91
 
92
- test_metadata_url = DataFilesDict.from_hf_repo(
93
- {datasets.Split.TEST: ["data/test/**"]},
94
- dataset_info=hfh_dataset_info,
95
- allowed_extensions=["jsonl", ".jsonl"],
96
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97
 
98
- metadata_urls = dict()
99
- metadata_urls["train"] = train_metadata_url["train"]
100
- metadata_urls["test"] = test_metadata_url["test"]
101
-
102
- # 画像データは**.zipのURLをDict型として取得?
103
- # **.zipのURLをDict型として取得?
104
- train_data_url = DataFilesDict.from_hf_repo(
105
- {datasets.Split.TRAIN: ["data/train/**"]},
106
- dataset_info=hfh_dataset_info,
107
- allowed_extensions=["zip", ".zip"],
108
- )
109
 
110
- test_data_url = DataFilesDict.from_hf_repo(
111
- {datasets.Split.TEST: ["data/test/**"]},
112
- dataset_info=hfh_dataset_info,
113
- allowed_extensions=["zip", ".zip"]
114
- )
115
- data_urls = dict()
116
- data_urls["train"] = train_data_url["train"]
117
- data_urls["test"] = test_data_url["test"]
 
 
 
 
 
 
 
 
118
 
119
- gs = []
120
-
121
- # for split, file_list in data_urls.items():
122
- # metadata_list = metadata_urls[split]
123
- # for i, file_ in enumerate(file_list):
124
- # '''
125
- # split : "train" or "test" or "val"
126
- # files : zip files
127
- # '''
128
- # # print(file_)
129
- # # print(metadata_list[0])
130
- # # # リポジトリからダウンロードしてとりあえずキャッシュしたURLリストを取得
131
- # metadata_path = dl_manager.download_and_extract(metadata_list[i])
132
- # downloaded_files = dl_manager.download(file_)
133
- # # # 元のコードではzipファイルの中身を"filepath"としてそのまま_generate_exampleに引き渡している?
134
- # gs.append(
135
- # datasets.SplitGenerator(
136
- # name = split,
137
- # gen_kwargs = {
138
- # # "images" : iter(iter_archive[split]),
139
- # "images" : dl_manager.iter_archive(downloaded_files),
140
- # "metadata_path": metadata_path # メタデータパスを渡す
141
- # }
142
- # )
143
- # )
144
-
145
- gs = []
146
-
147
- for split, file_list in data_urls.items():
148
- metadata_list = metadata_urls[split]
149
- gs.append(
150
- datasets.SplitGenerator(
151
- name = split,
152
- gen_kwargs = {
153
- # "images" : iter(iter_archive[split]),
154
- "images" : dl_manager.iter_archive(dl_manager.download([file_ for file_ in file_list])),
155
- "metadata_path": dl_manager.download_and_extract(metadata_list)# メタデータパスを渡す
156
- }
157
- )
158
- )
159
- return gs
160
 
161
 
162
  def _generate_examples(self, images, metadata_path):
 
34
  # DatasetInfo : https://huggingface.co/docs/datasets/package_reference/main_classes
35
 
36
 
37
+ # データを整理?
38
+ dl_manager = DownloadManager()
39
+ hfh_dataset_info = HfApi().dataset_info(_NAME, revision=_REVISION, timeout=100.0)
40
+
41
+ # メタデータであるjsonlファイルのURLを取得
42
+ # ここの抽出方法変えられないかな?
43
+ train_metadata_url = DataFilesDict.from_hf_repo(
44
+ {datasets.Split.TRAIN: ["data/train/**"]},
45
+ dataset_info=hfh_dataset_info,
46
+ allowed_extensions=["jsonl", ".jsonl"],
47
+ )
48
+
49
+ test_metadata_url = DataFilesDict.from_hf_repo(
50
+ {datasets.Split.TEST: ["data/test/**"]},
51
+ dataset_info=hfh_dataset_info,
52
+ allowed_extensions=["jsonl", ".jsonl"],
53
+ )
54
+
55
+
56
+ metadata_urls = dict()
57
+ metadata_urls["train"] = train_metadata_url["train"]
58
+ metadata_urls["test"] = test_metadata_url["test"]
59
+
60
+ # 画像データは**.zipのURLをDict型として取得?
61
+ # **.zipのURLをDict型として取得?
62
+ train_data_url = DataFilesDict.from_hf_repo(
63
+ {datasets.Split.TRAIN: ["data/train/**"]},
64
+ dataset_info=hfh_dataset_info,
65
+ allowed_extensions=["zip", ".zip"],
66
+ )
67
+
68
+ test_data_url = DataFilesDict.from_hf_repo(
69
+ {datasets.Split.TEST: ["data/test/**"]},
70
+ dataset_info=hfh_dataset_info,
71
+ allowed_extensions=["zip", ".zip"]
72
+ )
73
+ data_urls = dict()
74
+ data_urls["train"] = train_data_url["train"]
75
+ data_urls["test"] = test_data_url["test"]
76
+
77
+ class GraySpectrogramConfig(datasets.BuilderConfig):
78
+ """BuilderConfig for Imagette."""
79
+
80
+ def __init__(self, data_url, metadata_urls, **kwargs):
81
+ """BuilderConfig for Imagette.
82
+ Args:
83
+ data_url: `string`, url to download the zip file from.
84
+ matadata_urls: dictionary with keys 'train' and 'validation' containing the archive metadata URLs
85
+ **kwargs: keyword arguments forwarded to super.
86
+ """
87
+ super(GraySpectrogramConfig, self).__init__(version=datasets.Version("1.0.0"), **kwargs)
88
+ self.data_url = data_url
89
+ self.metadata_urls = metadata_urls
90
 
91
  class GraySpectrogram2(datasets.GeneratorBasedBuilder):
92
 
93
  # データのサブセットはここで用意
94
  BUILDER_CONFIGS = [
95
  datasets.BuilderConfig(
96
+ name="data 0-200",
97
  description=_DESCRIPTION,
98
+ data_url = {
99
+ "train" : data_urls["train"][0],
100
+ "test" : data_urls["test"][0]
101
+ },
102
+ metadata_urls = {
103
+ "train" : metadata_urls["train"][0],
104
+ "test" : metadata_urls["test"][0]
105
+ }
106
+
107
+ ),
108
+ datasets.BuilderConfig(
109
+ name="data 200-600",
110
+ description=_DESCRIPTION,
111
+ data_url ={
112
+ "train" : data_urls["train"][1],
113
+ "test" : data_urls["test"][1]
114
+ },
115
+ metadata_urls = {
116
+ "train": metadata_urls["train"][1],
117
+ "test" : metadata_urls["test"][1]
118
+ }
119
+
120
+ ),
121
+ datasets.BuilderConfig(
122
+ name="data 600-1000",
123
+ description=_DESCRIPTION,
124
+ data_url = {
125
+ "train" : data_urls["train"][2],
126
+ "test" : data_urls["test"][2]
127
+ },
128
+ metadata_urls = {
129
+ "train" : metadata_urls["train"][2],
130
+ "test" : metadata_urls["test"][2]
131
+ }
132
+ ),
133
+ datasets.BuilderConfig(
134
+ name="data 1000-1300",
135
+ description=_DESCRIPTION,
136
+ data_url = {
137
+ "train" : data_urls["train"][3],
138
+ "test" : data_urls["test"][3]
139
+ },
140
+ metadata_urls = {
141
+ "train" : metadata_urls["train"][3],
142
+ "test" : metadata_urls["test"][3]
143
+ }
144
+
145
+ ),
146
+ datasets.BuilderConfig(
147
+ name="data 1300-1600",
148
+ description=_DESCRIPTION,
149
+ data_url = {
150
+ "train" : data_urls["train"][4],
151
+ "test" : data_urls["test"][4]
152
+ },
153
+ metadata_urls = {
154
+ "train" : metadata_urls["train"][4],
155
+ "test" : metadata_urls["test"][4]
156
+ }
157
  )
158
  ]
159
 
 
190
  )
191
 
192
  def _split_generators(self, dl_manager: DownloadManager):
 
 
 
 
 
 
 
 
 
 
193
 
194
+ metadata_paths = dl_manager.download(self.config.metadata_urls)
195
+ data_paths = dl_manager.download(self.config.data_urls)
196
+
197
+ return [
198
+ datasets.SplitGenerator(
199
+ name=datasets.Split.TRAIN,
200
+ gen_kwargs={
201
+ "images": dl_manager.iter_archive(data_paths["train"]),
202
+ "metadata_path": metadata_paths["train"],
203
+ }
204
+ ),
205
+ datasets.SplitGenerator(
206
+ name=datasets.Split.TEST,
207
+ gen_kwargs={
208
+ "images": dl_manager.iter_archive(data_paths["test"]),
209
+ "metadata_path": metadata_paths["test"],
210
+ }
211
+ ),
212
+ ]
213
+
214
+ # # huggingfaceのディレクトリからデータを取ってくる
215
+ # hfh_dataset_info = HfApi().dataset_info(_NAME, revision=_REVISION, timeout=100.0)
216
+
217
+ # # メタデータであるjsonlファイルのURLを取得
218
+ # # ここの抽出方法変えられないかな?
219
+ # train_metadata_url = DataFilesDict.from_hf_repo(
220
+ # {datasets.Split.TRAIN: ["data/train/**"]},
221
+ # dataset_info=hfh_dataset_info,
222
+ # allowed_extensions=["jsonl", ".jsonl"],
223
+ # )
224
+
225
+ # test_metadata_url = DataFilesDict.from_hf_repo(
226
+ # {datasets.Split.TEST: ["data/test/**"]},
227
+ # dataset_info=hfh_dataset_info,
228
+ # allowed_extensions=["jsonl", ".jsonl"],
229
+ # )
230
 
231
+ # metadata_urls = dict()
232
+ # metadata_urls["train"] = train_metadata_url["train"]
233
+ # metadata_urls["test"] = test_metadata_url["test"]
 
 
 
 
 
 
 
 
234
 
235
+ # # 画像データは**.zipのURLをDict型として取得?
236
+ # # **.zipのURLをDict型として取得?
237
+ # train_data_url = DataFilesDict.from_hf_repo(
238
+ # {datasets.Split.TRAIN: ["data/train/**"]},
239
+ # dataset_info=hfh_dataset_info,
240
+ # allowed_extensions=["zip", ".zip"],
241
+ # )
242
+
243
+ # test_data_url = DataFilesDict.from_hf_repo(
244
+ # {datasets.Split.TEST: ["data/test/**"]},
245
+ # dataset_info=hfh_dataset_info,
246
+ # allowed_extensions=["zip", ".zip"]
247
+ # )
248
+ # data_urls = dict()
249
+ # data_urls["train"] = train_data_url["train"]
250
+ # data_urls["test"] = test_data_url["test"]
251
 
252
+ # gs = []
253
+
254
+ # # for split, file_list in data_urls.items():
255
+ # # metadata_list = metadata_urls[split]
256
+ # # for i, file_ in enumerate(file_list):
257
+ # # '''
258
+ # # split : "train" or "test" or "val"
259
+ # # files : zip files
260
+ # # '''
261
+ # # # print(file_)
262
+ # # # print(metadata_list[0])
263
+ # # # # リポジトリからダウンロードしてとりあえずキャッシュしたURLリストを取得
264
+ # # metadata_path = dl_manager.download_and_extract(metadata_list[i])
265
+ # # downloaded_files = dl_manager.download(file_)
266
+ # # # # 元のコードではzipファイルの中身を"filepath"としてそのまま_generate_exampleに引き渡している?
267
+ # # gs.append(
268
+ # # datasets.SplitGenerator(
269
+ # # name = split,
270
+ # # gen_kwargs = {
271
+ # # # "images" : iter(iter_archive[split]),
272
+ # # "images" : dl_manager.iter_archive(downloaded_files),
273
+ # # "metadata_path": metadata_path # メタデータパスを渡す
274
+ # # }
275
+ # # )
276
+ # # )
277
+
278
+ # return gs
 
 
 
 
 
 
 
 
 
 
 
 
 
 
279
 
280
 
281
  def _generate_examples(self, images, metadata_path):