DancingPrismPJ commited on
Commit
b37a55d
·
verified ·
1 Parent(s): a85d83b

Rename files to work as expected

Browse files
Wikipedia-Horse-Dataset.py ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from __future__ import annotations
2
+
3
+ import json
4
+ import random
5
+ from typing import Generator
6
+
7
+ import datasets # type: ignore
8
+
9
+ _CITATION = ""
10
+ _DESCRIPTION = "This is a dataset of Wikipadia articles about Japan-trained horse manually labeled with the named entity label. I used DBpedia Japan searched Wikipedia Articles."
11
+ _HOMEPAGE = "https://ja.dbpedia.org"
12
+ _LICENSE = "This work is licensed under CC-BY-SA 3.0"
13
+ _URL = "https://huggingface.co/datasets/DancingPrismPJ/Wikipedia-Horse-Dataset/raw/main/horse-dataset-600.json"
14
+
15
+ class wikipediaHorseDatasetConfig(datasets.BuilderConfig):
16
+ def __init__(
17
+ self,
18
+ name: str = "default",
19
+ version: datasets.Version | str | None = datasets.Version("0.0.0"),
20
+ data_dir: str | None = None,
21
+ data_files: datasets.data_files.DataFilesDict | None = None,
22
+ description: str | None = _DESCRIPTION,
23
+ shuffle: bool = True,
24
+ seed: int = 42,
25
+ train_ratio: float = 0.8,
26
+ validation_ratio: float = 0.1,
27
+ ) -> None:
28
+ super().__init__(
29
+ name=name,
30
+ version=version,
31
+ data_dir=data_dir,
32
+ data_files=data_files,
33
+ description=description,
34
+ )
35
+ self.shuffle = shuffle
36
+ self.seed = seed
37
+ self.train_ratio = train_ratio
38
+ self.validation_ratio = validation_ratio
39
+
40
+ class wikipediaHorseDataset(datasets.GeneratorBasedBuilder):
41
+ BUILDER_CONFIG_CLASS = wikipediaHorseDatasetConfig
42
+
43
+ def _info(self) -> datasets.DatasetInfo:
44
+ return datasets.DatasetInfo(
45
+ description=_DESCRIPTION,
46
+ features=datasets.Features(
47
+ {
48
+ "curid": datasets.Value("string"),
49
+ "text": datasets.Value("string"),
50
+ "entities": [
51
+ {
52
+ "name": datasets.Value("string"),
53
+ "span": datasets.Sequence(
54
+ datasets.Value("int64"), length=3
55
+ ),
56
+ "type": datasets.Value("string"),
57
+ }
58
+ ],
59
+ }
60
+ ),
61
+ homepage=_HOMEPAGE,
62
+ license=_LICENSE,
63
+ citation=_CITATION,
64
+ )
65
+
66
+ def _split_generators(
67
+ self, dl_manager: datasets.DownloadManager
68
+ ) -> list[datasets.SplitGenerator]:
69
+ dataset_dir = str(dl_manager.download_and_extract(_URL))
70
+ with open(dataset_dir, "r", encoding="utf-8") as f:
71
+ data = json.load(f)
72
+
73
+ if self.config.shuffle == True:
74
+ random.seed(self.config.seed)
75
+ random.shuffle(data)
76
+
77
+ num_data = len(data)
78
+ num_train_data = int(num_data * self.config.train_ratio)
79
+ num_validation_data = int(num_data * self.config.validation_ratio)
80
+ train_data = data[:num_train_data]
81
+ validation_data = data[num_train_data : num_train_data + num_validation_data]
82
+ test_data = data[num_train_data + num_validation_data :]
83
+ return [
84
+ datasets.SplitGenerator(
85
+ name=datasets.Split.TRAIN,
86
+ gen_kwargs={"data": train_data},
87
+ ),
88
+ datasets.SplitGenerator(
89
+ name=datasets.Split.VALIDATION,
90
+ gen_kwargs={"data": validation_data},
91
+ ),
92
+ datasets.SplitGenerator(
93
+ name=datasets.Split.TEST,
94
+ gen_kwargs={"data": test_data},
95
+ ),
96
+ ]
97
+
98
+ def _generate_examples(self, data: list[dict[str, str]]) -> Generator:
99
+ for key, d in enumerate(data):
100
+ yield key, {
101
+ "curid": d["curid"],
102
+ "text": d["text"],
103
+ "entities": d["entities"],
104
+ }
105
+
106
+
horse-dataset-600.json ADDED
The diff for this file is too large to render. See raw diff