ygorg commited on
Commit
d75aedb
·
verified ·
1 Parent(s): 8198e3f

Update scripts to use data.zip instead of local data.

Browse files
Files changed (4) hide show
  1. CAS.py +37 -29
  2. data.zip +2 -2
  3. data/CAS_neg.txt +0 -0
  4. data/CAS_spec.txt +0 -0
CAS.py CHANGED
@@ -59,17 +59,24 @@ _HOMEPAGE = "https://clementdalloux.fr/?page_id=28"
59
 
60
  _LICENSE = 'Data User Agreement'
61
 
 
 
 
62
  class CAS(datasets.GeneratorBasedBuilder):
63
 
64
- DEFAULT_CONFIG_NAME = "pos_spec"
65
 
66
  BUILDER_CONFIGS = [
67
- datasets.BuilderConfig(name="pos", version="1.0.0", description="The CAS corpora - POS Speculation task"),
 
68
 
69
- datasets.BuilderConfig(name="cls", version="1.0.0", description="The CAS corpora - CLS Negation / Speculation task"),
 
70
 
71
- datasets.BuilderConfig(name="ner_spec", version="1.0.0", description="The CAS corpora - NER Speculation task"),
72
- datasets.BuilderConfig(name="ner_neg", version="1.0.0", description="The CAS corpora - NER Negation task"),
 
 
73
  ]
74
 
75
  def _info(self):
@@ -82,13 +89,20 @@ class CAS(datasets.GeneratorBasedBuilder):
82
  "document_id": datasets.Value("string"),
83
  "tokens": [datasets.Value("string")],
84
  "lemmas": [datasets.Value("string")],
85
- # "pos_tags": [datasets.Value("string")],
86
  "pos_tags": [datasets.features.ClassLabel(
87
- names = ['B-INT', 'B-PRO:DEM', 'B-VER:impf', 'B-VER:ppre', 'B-PRP:det', 'B-KON', 'B-VER:pper', 'B-PRP', 'B-PRO:IND', 'B-VER:simp', 'B-VER:con', 'B-SENT', 'B-VER:futu', 'B-PRO:PER', 'B-VER:infi', 'B-ADJ', 'B-NAM', 'B-NUM', 'B-PUN:cit', 'B-PRO:REL', 'B-VER:subi', 'B-ABR', 'B-NOM', 'B-VER:pres', 'B-DET:ART', 'B-VER:cond', 'B-VER:subp', 'B-DET:POS', 'B-ADV', 'B-SYM', 'B-PUN'],
 
 
 
 
 
 
 
 
88
  )],
89
  }
90
  )
91
-
92
  elif self.config.name.find("cls") != -1:
93
 
94
  features = datasets.Features(
@@ -96,13 +110,12 @@ class CAS(datasets.GeneratorBasedBuilder):
96
  "id": datasets.Value("string"),
97
  "document_id": datasets.Value("string"),
98
  "tokens": [datasets.Value("string")],
99
- # "label": datasets.Value("string"),
100
  "label": datasets.features.ClassLabel(
101
- names = ['negation_speculation', 'speculation', 'neutral', 'negation'],
102
  ),
103
  }
104
  )
105
-
106
  elif self.config.name.find("ner") != -1:
107
 
108
  if self.config.name.find("_spec") != -1:
@@ -116,9 +129,8 @@ class CAS(datasets.GeneratorBasedBuilder):
116
  "document_id": datasets.Value("string"),
117
  "tokens": [datasets.Value("string")],
118
  "lemmas": [datasets.Value("string")],
119
- # "ner_tags": [datasets.Value("string")],
120
  "ner_tags": [datasets.features.ClassLabel(
121
- names = names,
122
  )],
123
  }
124
  )
@@ -134,12 +146,8 @@ class CAS(datasets.GeneratorBasedBuilder):
134
 
135
  def _split_generators(self, dl_manager):
136
 
137
- if self.config.data_dir is None:
138
- raise ValueError("This is a local dataset. Please pass the data_dir kwarg to load_dataset.")
139
-
140
- else:
141
- data_dir = self.config.data_dir
142
-
143
  return [
144
  datasets.SplitGenerator(
145
  name=datasets.Split.TRAIN,
@@ -203,7 +211,7 @@ class CAS(datasets.GeneratorBasedBuilder):
203
  id_doc, id_word, word, lemma, tag = splitted[0:5]
204
  if len(splitted) >= 8:
205
  tag = splitted[6]
206
-
207
  if tag == "@card@":
208
  print(splitted)
209
 
@@ -224,13 +232,13 @@ class CAS(datasets.GeneratorBasedBuilder):
224
  id_words.append(id_word)
225
  words.append(word)
226
  lemmas.append(lemma)
227
- POS_tags.append('B-'+tag)
228
 
229
  dic = {
230
- "id_docs": np.array(list(map(int, id_docs))),
231
  "id_words": id_words,
232
- "words": words,
233
- "lemmas": lemmas,
234
  "POS_tags": POS_tags,
235
  }
236
 
@@ -270,7 +278,7 @@ class CAS(datasets.GeneratorBasedBuilder):
270
  continue
271
 
272
  id_doc, id_word, word, lemma, _ = line.split("\t")[0:5]
273
- tag = line.replace("\n","").split("\t")[-1]
274
 
275
  if tag == "***" or tag == "_":
276
  tag = "O"
@@ -288,10 +296,10 @@ class CAS(datasets.GeneratorBasedBuilder):
288
  ner_tags.append(tag)
289
 
290
  dic = {
291
- "id_docs": np.array(list(map(int, id_docs))),
292
  "id_words": id_words,
293
- "words": words,
294
- "lemmas": lemmas,
295
  "ner_tags": ner_tags,
296
  }
297
 
@@ -384,4 +392,4 @@ class CAS(datasets.GeneratorBasedBuilder):
384
 
385
  for r in all_res:
386
  if r["id"] in allowed_ids:
387
- yield r["id"], r
 
59
 
60
  _LICENSE = 'Data User Agreement'
61
 
62
+ _URL = "data.zip"
63
+
64
+
65
  class CAS(datasets.GeneratorBasedBuilder):
66
 
67
+ DEFAULT_CONFIG_NAME = "pos"
68
 
69
  BUILDER_CONFIGS = [
70
+ datasets.BuilderConfig(name="pos", version="1.0.0",
71
+ description="The CAS corpora - POS Speculation task"),
72
 
73
+ datasets.BuilderConfig(name="cls", version="1.0.0",
74
+ description="The CAS corpora - CLS Negation / Speculation task"),
75
 
76
+ datasets.BuilderConfig(name="ner_spec", version="1.0.0",
77
+ description="The CAS corpora - NER Speculation task"),
78
+ datasets.BuilderConfig(name="ner_neg", version="1.0.0",
79
+ description="The CAS corpora - NER Negation task"),
80
  ]
81
 
82
  def _info(self):
 
89
  "document_id": datasets.Value("string"),
90
  "tokens": [datasets.Value("string")],
91
  "lemmas": [datasets.Value("string")],
 
92
  "pos_tags": [datasets.features.ClassLabel(
93
+ names=[
94
+ 'B-ABR', 'B-ADJ', 'B-ADV', 'B-DET:ART', 'B-DET:POS', 'B-INT',
95
+ 'B-KON', 'B-NAM', 'B-NOM', 'B-NUM',
96
+ 'B-PRO:DEM', 'B-PRO:IND', 'B-PRO:PER',
97
+ 'B-PRO:REL', 'B-PRP', 'B-PRP:det', 'B-PUN', 'B-PUN:cit',
98
+ 'B-SENT', 'B-SYM', 'B-VER:con', 'B-VER:cond', 'B-VER:futu',
99
+ 'B-VER:impf', 'B-VER:infi', 'B-VER:pper', 'B-VER:ppre',
100
+ 'B-VER:pres', 'B-VER:simp', 'B-VER:subi', 'B-VER:subp'
101
+ ],
102
  )],
103
  }
104
  )
105
+
106
  elif self.config.name.find("cls") != -1:
107
 
108
  features = datasets.Features(
 
110
  "id": datasets.Value("string"),
111
  "document_id": datasets.Value("string"),
112
  "tokens": [datasets.Value("string")],
 
113
  "label": datasets.features.ClassLabel(
114
+ names=['negation_speculation', 'negation', 'neutral', 'speculation'],
115
  ),
116
  }
117
  )
118
+
119
  elif self.config.name.find("ner") != -1:
120
 
121
  if self.config.name.find("_spec") != -1:
 
129
  "document_id": datasets.Value("string"),
130
  "tokens": [datasets.Value("string")],
131
  "lemmas": [datasets.Value("string")],
 
132
  "ner_tags": [datasets.features.ClassLabel(
133
+ names=names,
134
  )],
135
  }
136
  )
 
146
 
147
  def _split_generators(self, dl_manager):
148
 
149
+ data_dir = dl_manager.download_and_extract(_URL).rstrip("/")
150
+
 
 
 
 
151
  return [
152
  datasets.SplitGenerator(
153
  name=datasets.Split.TRAIN,
 
211
  id_doc, id_word, word, lemma, tag = splitted[0:5]
212
  if len(splitted) >= 8:
213
  tag = splitted[6]
214
+
215
  if tag == "@card@":
216
  print(splitted)
217
 
 
232
  id_words.append(id_word)
233
  words.append(word)
234
  lemmas.append(lemma)
235
+ POS_tags.append(f'B-{tag}')
236
 
237
  dic = {
238
+ "id_docs": np.array(list(map(int, id_docs))),
239
  "id_words": id_words,
240
+ "words": words,
241
+ "lemmas": lemmas,
242
  "POS_tags": POS_tags,
243
  }
244
 
 
278
  continue
279
 
280
  id_doc, id_word, word, lemma, _ = line.split("\t")[0:5]
281
+ tag = line.replace("\n", "").split("\t")[-1]
282
 
283
  if tag == "***" or tag == "_":
284
  tag = "O"
 
296
  ner_tags.append(tag)
297
 
298
  dic = {
299
+ "id_docs": np.array(list(map(int, id_docs))),
300
  "id_words": id_words,
301
+ "words": words,
302
+ "lemmas": lemmas,
303
  "ner_tags": ner_tags,
304
  }
305
 
 
392
 
393
  for r in all_res:
394
  if r["id"] in allowed_ids:
395
+ yield r["id"], r
data.zip CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:058040313dc91432d8d9b6985cdf112d466634e3956b5f7f1326e83eb0ee11a9
3
- size 1185006
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbf074c8034354f28d9909ac6f6339acb5d4e556dadca5cbe6e19fb0a85e7696
3
+ size 1183523
data/CAS_neg.txt DELETED
The diff for this file is too large to render. See raw diff
 
data/CAS_spec.txt DELETED
The diff for this file is too large to render. See raw diff