File size: 4,784 Bytes
b8bb974
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os

import datasets
import pandas as pd


_DESCRIPTION = """\
TMMLU2 data loader
"""
_DATA_PATH = "data"

task_list = [
             'dentistry', 'traditional_chinese_medicine_clinical_medicine', 'clinical_psychology',
             'technical', 'culinary_skills', 'mechanical', 'logic_reasoning', 'real_estate',
             'general_principles_of_law', 'finance_banking', 'anti_money_laundering', 'ttqav2',
             'marketing_management', 'business_management', 'organic_chemistry', 'advance_chemistry',
             'physics', 'secondary_physics', 'human_behavior', 'national_protection', 'jce_humanities',
             'politic_science', 'agriculture', 'official_document_management',
             'financial_analysis', 'pharmacy', 'educational_psychology', 'statistics_and_machine_learning',
             'management_accounting', 'introduction_to_law', 'computer_science', 'veterinary_pathology',
             'accounting', 'fire_science', 'optometry', 'insurance_studies', 'pharmacology', 'taxation',
             'education_(profession_level)', 'economics',
             'veterinary_pharmacology', 'nautical_science', 'occupational_therapy_for_psychological_disorders',
             'trust_practice', 'geography_of_taiwan', 'physical_education', 'auditing', 'administrative_law',
             'basic_medical_science', 'macroeconomics', 'trade', 'chinese_language_and_literature',
             'tve_design', 'junior_science_exam', 'junior_math_exam', 'junior_chinese_exam',
             'junior_social_studies', 'tve_mathematics', 'tve_chinese_language',
             'tve_natural_sciences', 'junior_chemistry', 'music', 'education',
             'three_principles_of_people', 'taiwanese_hokkien',
             'engineering_math'
            ]

_URLs = {
    task_name: {
        split_name: [
            os.path.join(
                _DATA_PATH, task_name+"_"+split_name+".csv"
            ),  # TODO -> handle multiple shards
        ]
        for split_name in ['dev', 'test', 'val']
    }
    for task_name in task_list
}


class TMMLU2Config(datasets.BuilderConfig):
    def __init__(self, **kwargs):
        super().__init__(version=datasets.Version("1.0.0"), **kwargs)


class TMMLU2(datasets.GeneratorBasedBuilder):
    BUILDER_CONFIGS = [
        TMMLU2Config(
            name=task_name,
        )
        for task_name in task_list
    ]

    def _info(self):
        features = datasets.Features(
            {
                "question": datasets.Value("string"),
                "A": datasets.Value("string"),
                "B": datasets.Value("string"),
                "C": datasets.Value("string"),
                "D": datasets.Value("string"),
                "answer": datasets.Value("string"),
            }
        )
        return datasets.DatasetInfo(
            description=_DESCRIPTION,
            features=features,
        )

    def _split_generators(self, dl_manager):
        task_name = self.config.name
        data_dir = dl_manager.download(_URLs[task_name])
        return [
            datasets.SplitGenerator(
                name=datasets.Split.TEST,
                gen_kwargs={
                    "filepath": data_dir['test'],
                },
            ),
            datasets.SplitGenerator(
                name=datasets.Split.VALIDATION,
                gen_kwargs={
                    "filepath": data_dir['val'],
                },
            ),
            datasets.SplitGenerator(
                name=datasets.Split.TRAIN,
                gen_kwargs={
                    "filepath": data_dir['dev'],
                },
            ),
        ]

    def _generate_examples(self, filepath):
        if isinstance(filepath, list):
            filepath = filepath[0]
        df = pd.read_csv(filepath)

        for i, instance in enumerate(df.to_dict(orient="records")):
            yield i, {'question': instance['question'],
                      'A': instance['A'],
                      'B': instance['B'],
                      'C': instance['C'],
                      'D': instance['D'],
                      'answer': instance['answer']
                      }